The Model Isn't the Product. The Harness Is.
After auditing a publicly posted Claude Code source snapshot and reading OpenAI's harness essays, I came away with one conclusion: the model is only the brain. The product is the runtime around it.
6 parts · Mar 4, 2026 – Apr 6, 2026
After auditing a publicly posted Claude Code source snapshot and reading OpenAI's harness essays, I came away with one conclusion: the model is only the brain. The product is the runtime around it.
The most important file in the Claude Code snapshot isn't a prompt file or a tool definition. It's a state machine. Once you see that, a lot of agent design decisions start making sense.
The most novel thing in the Claude Code snapshot isn't a tool or a prompt. It's how aggressively the runtime is built around preserving prompt-cache identity across forks, compaction, and long-running sessions.
LLMs love to talk. For a companion app, that's a problem — real people don't write essays when you text them. Here's how I built a hybrid system to control response length without killing personality.
An AI companion that forgets your name after 200 messages isn't a companion — it's a chatbot with amnesia. Here's how I'm rebuilding Mio's memory from a dumb key-value store into something that feels like actual recall.
I built a personal knowledge compiler that treats the LLM as a programmer and your wiki as a codebase. 15 source files, 5 dependencies, zero databases.
© Xingfan Xia 2024 - 2026 · CC BY-NC 4.0