Activity
Mon
Wed
Fri
Sun
Jun
Jul
Aug
Sep
Oct
Nov
Dec
Jan
Feb
Mar
Apr
May
What is this?
Less
More

Memberships

AI Agents & Vibe Coding

206 members • Free

17 contributions to AI Agents & Vibe Coding
Stop making your Agent think out loud
Nobody talks about this: the 'thinking out loud' style of reasoning models is mostly noise for production. You don't ship a chatbot that goes 'Let me think... I should consider...'. Strip the chain-of-thought from outputs. Reasoning is useful internally, ugly externally.
My First AI Project - AIOS
Hi Everyone, This is my first project AIOS v1 -> https://github.com/attacker-codeninja/AIOS About -> Unified AI terminal for Claude, Gemini, and Ollama — session memory, custom commands Don't make users manage AI tools. Make one tool manage the AI tools. A manager shell for Claude Code, Gemini CLI, Kimi, and Ollama , built in Rust. Please check and provide your feedback and contribute if want and We can improve it for v2
0 likes • 4d
Rust + multi-provider is a great combo. How are you handling streaming responses across the different APIs? Each provider has a different streaming format and that's usually where these CLI tools fall apart.
Killed a 3-week project today. Here's what I learned.
Killed a 3-week project today. Built a "smart" meeting summarizer for a client. They didn't want it. They wanted just the action items in their slack which would've taken 1 day. I overbuilt because I fell in love with the architecture. Lesson: ask "what do you do with the output?" BEFORE designing the agent. Output drives architecture, not the other way around. Anyone else have a "wow that was an expensive lesson" moment this week?
0 likes • 8d
I overbuilt because I fell in love with the architecture.im stealing that.
Is 88% accuracy "ship ready" or am I being lazy?v
How do you all decide when an agent is good enough to ship? My email tagger has been running for 5 days, accuracy is around 88%. Is that ship-ready or am I being lazy? Genuine question. I keep tweaking and I think I'm just procrastinating at this point.
0 likes • 9d
88% on 5 days is fine, ship and iterate
New Video: Just shipped an HR onboarding AI Agent using Claude SDK
New video up. HR onboarding agent. Full workflow: → Reads passports, driver's licenses, offer letters with Claude vision → Auto-creates HRIS records, IAM accounts, onboarding checklists → Sends I-9 / W-4 to eSignature, enrolls benefits, orders background checks → Tracks compliance deadlines (I-9 at hire+3 days, state new-hire reporting at hire+20) → Flags REVIEW_REQUIRED for HR. Human in the loop, always. But the real lesson isn't in any of that. Background checks come back as "clear," "consider," or "fail." A naive agent auto-rejects on "consider." That's where compliance breaks. EEOC requires individualized human assessment, not algorithmic rejection. The fix wasn't a bigger system prompt. It was ONE line in the tool description: "If result is 'consider,' alert HR for manual review. Never auto-block." Production agents aren't built in the system prompt. They're built in the tool descriptions. Stack: Python · Claude Agent SDK (sonnet-4-6 + vision) · AgentAPIsHub · httpx · Pydantic · Rich 9 modules · 26 tools · 1,200 lines. 🎥 Full walkthrough + downloadable source: https://www.youtube.com/watch?v=-XcBsPWr1KY Question for builders here. What's one tool description you've written that quietly saved you from a production bug? Drop it below.
0 likes • 9d
the EEOC point is huge. most ppl don't realize 'consider' is a legal landmine. tool description fixes are the underrated layer of agent design.
1-10 of 17
Ewan Park
2
8points to level up
@ewan-park-4973
AI

Active 1d ago
Joined Mar 24, 2026
Powered by