User
Write something
Afternoon Tea is happening in 5 days
Pinned
Welcome to Clief Notes. Here's where to start.
1. Watch the intro video and introduce yourself in the intro post here 2. Start with The Foundation (free course). Concepts, folder architecture, prompting framework. Everything else builds on this. 3. Check in at the bottom of each lesson. Polls, discussion posts, other members working through the same stuff. Use them. 4. When you're ready to build real things, move to Implementation Playbooks (Level 2). When you're ready to build your own tools, Building Your Stack (Level 3). 5. Post your work. Ask questions. Help others when you can. What are you here to build?
Poll
5051 members have voted
Pinned
🏆 WEEKLY COMP #4: THE AGENCY 🏆
💰 $325 CASH 💰 That's a full year of Premium. Win this and your membership pays for itself. But the real prize this week isn't the cash. Keep reading. 📋 THE CHALLENGE You just got hired again. Meet Diana, owner of a 4-person boutique real estate team in Austin. 60-80 transactions a year, mostly residential, mix of buyers and sellers. 📎 Download the full client brief attached to this post. Short version: She doesn't want software. She wants a system she can teach her team to use in a week. Your job is to build the AI operating system for her team. This isn't one specialist. This is a small team of AI specialists organized into a multi-folder ICM architecture, with a clear handoff protocol between them. 🗂️ WHAT YOU'RE BUILDING Last week was one specialist. This week is a team of them. Required folders: 📍 00_orchestrator/ — The router. Where every request starts. Decides which specialist gets the job. 📍 01_lead_qualifier/ — First contact with new prospects. Captures intent, budget, timeline. 📍 02_property_research/ — Deep research on specific properties or neighborhoods. 📍 03_client_communication/ — Drafts emails, texts, follow-ups in the voice of the agent. 📍 04_transaction_coordinator/ — Handles the deal once it's live. Checklists, deadlines, document tracking. Each folder must include: - 📄 identity.md - 📐 rules.md - 💬 examples.md - 🔗 handoff.md (NEW for Week 4 — how does this folder pass work to another folder?) - Plus a root-level README.md explaining the architecture, the typical flow, and how to onboard a new team member. 🔥 WHY THIS ONE IS DIFFERENT Weeks 1, 2, and 3 were warmups. This is the comp where the work you ship genuinely starts to look like the real thing. The handoff protocol is the test. Anyone can build five folders. The hard part is defining what each one needs from the previous one and what it passes to the next one. That's where multi-agent systems actually live or die.
1
0
Pinned
I come asking for help! (NEW ROUND! VOTE ONCE A DAY PLS)
Because of the Amazing support you all gave for the first Round Wylder (my step daughter) made it into the second round! You can vote once a day and some days are 2x votes ! I would love love love if any of you support her going to work with some of the best animal rescues in the world to just cast at least one free vote if you can! You can vote here! Not Ai related so sorry for that ! Wylder | Junior Ranger
I'm dumb. Here's proof.
I was today years old when I realized I did not have some of the most important files that you need in the folder structure that Jake teaches. During today's video call with the VIP group, he went on a deep-dive rabbit trail about the ICM folder methodology that he teaches in his foundations course (free). As he was discussing it, I went to check what my root folder looked like and I did not have a Claude.md or context.md file!!! My productivity skyrocketed ever since I implemented his folder strategy over a month ago, but little did I know that I hadn't even implemented it correctly. 🤯 🤯 🤯 This goes to show that massive action beats over planning every time!
🧪 New benchmark out
New benchmark out of Meta FAIR, Stanford, and Harvard called ProgramBench. The setup: you get a compiled executable plus its docs. Source code stripped. Rebuild the program from scratch in any language you want. Tests check input/output behavior against the original binary. 200 tasks, from small CLI tools up to FFmpeg, SQLite, and the PHP interpreter. 📊 Results across 9 models: Zero tasks fully solved. Opus 4.7 was the best, passing 95% of tests on only 3% of tasks. GPT 5.4, Gemini 3.1 Pro, and Haiku 4.5 hit 0% in that bucket. The interesting part is section 5. Even the model solutions that "worked" looked nothing like the human reference. Median 1,173 lines vs 3,068 in the original. Flat directories. Fewer functions, each one longer. GPT 5.4 wrote 96% of its final code in a single turn on most tasks and never modified existing files on roughly 40% of runs. 🎯 Why it matters for us: The benchmark separates writing code from designing software. Models can produce syntax all day. They cannot yet decompose a real system into coherent modules, pick the right abstractions, or organize a codebase the way a working engineer would. That gap is what computational orchestration points at. It is also where the durable value lives. 🛠 Try it: Pick an easier task from the repo (the paper flags nnn, fzf, gron, and jq as more tractable). Run it against Claude or your model of choice. Watch where you and the model split. Note the design decisions you make that the model never even raises. Post your runs and attempts to create a harness that would allow the model to do it. Wins, failures, weird outputs, all of it. 📍 Paper and Repo: ProgramBench I'm building something on top of this right now. More soon.
1-30 of 1,111
Clief Notes
skool.com/cliefnotes
Jake Van Clief, giving you the Cliff notes on the new AI age.
Leaderboard (30-day)
Powered by