I was solely using a Telegram bot so the end-user could test a set of mind clones I'd done for a client-friend.
The clone felt dumb. Actually smart, just not smart enough. Context was lost between conversations, some chats started from zero, and I was burning a load of tokens.
Everything lived in the system prompt — voice, preferences, past decisions, context, all stuffed in. Every call paid the full tax.
## The fix came from a friend on a call ##
Not a blog post, not a tutorial. One piece of advice during an AI session call. Someone who had knowledge to share about this threw one sentence, and everything clicked!
Externalize the persona. Vector store for the knowledge. Redis for session memory.
That was it.
## Where it is now
Context is kept with Redis — the bot remembers between conversations. Search runs on a large knowledge base with Upstash Vector. The system prompt shrank to the essentials: who the agent is, how to reach into its own memory, how to use what comes back.
Message comes in → embed the query → pull the handful of chunks that matter → reason with that.
Not the whole persona every time. Just the slice that matters now.
Tokens dropped. The clone stopped feeling dumb.
Still working on it. But the architecture is right now.
## The mind cloning bits
A mind clone isn't a prompt that says "act like me." That was my first mistake. It's a retrieval surface that has your decisions and can pull the right one when the situation rhymes with something you've already decided.
Prompts don't learn. Vector stores do. Every new decision, every corrected output, every preference — write it back. The store gets denser. The clone gets sharper.
## The big takeaway for me
A single advice from a friend did all this.
Not a course. Not a framework. Not a week of YouTube. One sentence from someone in the squad who'd already solved it.
And that's community power.