HomePostsArtQuests

Why you shouldn’t use Claude Code

I barely write code anymore.

I now constantly run a dozen of agents working on a handful of projects. My job shifted from “programmer” to magaging a “control plane.” It feels very similar to the video games of my childhood; I get buzzing notifications every few seconds, there’s a huge amount of context switching. The perfect environment for a generation that struggles to focus on a TikTok without a Subway Surfer video.

Casino old people playing slots

But somehow … it works. I have never been so productive.

I spent last year writing software, almost every day. Even when I was wandering in the African jungle I brought a starlink with me to write code. But I did more over the last few weeks than in the past 12 months.

I probably would have been just as effective if I had been in vacation all year and started using Opus just a little earlier.

GitHub contributions

That made something obvious to me.

Agents are commodities. My workflow isn’t.

Agents are a solved problem. We’ve tested many approaches: creating fleets of LLMs with complex graph architectures to optimize the context passed to each child. Turns out it’s not necessary: the optimal design seems to be a simple while loop around a LLM with sufficient time horizon.

Models improve with longer time horizon

Claude Code, Codex, Cursor agents, OpenCode, whatever comes next, they all converge on the same primitive. Features differ. UX differs. Marketing differs. But structurally, they are the same thing. Which means two things:

  1. Models will keep making the difference.
  2. Providers will try to lock you in with their client.

Always having the best models is not a viable long-term strategy for inference providers like Anthropic because it costs them a lot, and once they’ve made the investments, competitors can replicate the results for less (remember DeepSeek achieving near-GPT performance for a fraction of the price by training on its outputs). So they will try to lock you in.

I believe they will fail.

… And that’s why I’m bullish on open source

Writing software is now close to free. The marginal cost of producing code collapsed faster than anyone expected. What matters now is taste. And taste comes from experience.

Open source agent runtimes like Open Code (think a Claude Code clone for any model) will outperform closed ones long-term because taste compounds socially.

Thousands of people:

…will always explore a larger design space than any private team can. That’s why my top priority now is optimizing for low switching costs. Everywhere.

Imagine if productivity gains continue their current exponential trend, even for a few months. Imagine that Gemini 4 is released and allows me to ship twice as fast as Opus 4.5. Every day spent on Claude Code because I am used to it would have a considerable, increasingly large cost. In the same way as the time spent relearning how to use the best tooling or adapting my config.

You shall be independent

Once I internalized that, I reviewed my entire workflow. Let me break down each layer and what I do to stay flexible.

Agent stack diagram

The model (inference)

This is what actually thinks. I constantly switch between Claude, GPT 5.2, and Gemini 3 Pro. When one model gets stuck, another can often solve it with a different approach. The model is a commodity; treat it like one.

The agent runtime

It is the loop that turns a model into something useful. It doesn’t matter very much either as long as it supports custom skills, prompts, and MCP configs.

The agent client

This is what you actually interact with and this is where I refuse to use defaults. Claude Code’s terminal UI is fine, but it locks you into their ecosystem. I really believe my muscle memory is a liability.

I use Conductor, a native app that can run both Claude Code, Codex, and hopefully soon Open Code or Gemini. My friends @eni and @ClementWalter swear by Vibe Kanban. Pick what you want as long as you can easily switch the models.

Conductor app screenshot

The agent configuration

This is where your taste lives:

It has a lot of value, and it could a very strong reason to not move to the best agent available. GPT might be smarter, but if it is blind and can’t access your browser, dumb Claude will better.

Thankfully AGI is here to free us from incompatible standards.

I moved all these configs to a private git repository in my own format (you may get it here), and Claude wrote a script to automatically “translate” and synchronize it with all my agents. If I teach GPT how to reverse engineer Java binaries, Gemini learns it as well.

So what to do next?

Optimizing switching costs gave me some degree of freedom but ufortunately that system is already obsolete. It is poorly suited for long-term tasks and my computer becomes crowded. Sometimes I am reviewing Codex work and Claude bothers me by taking control of my screen. If I leave it running or use a VPS it’s particularly annoying to manage remotely (especially from a phone). I’m not even talking about the security risks of giving full access to my machine.

More importantly it still has two huge dependencies:

That’s what I’m trying to fix right now. I bought one NVIDIA Superchip before stocks run out and I’m going to learn how to run backup inference.

Regarding the human bottleneck that I am, I started building an infrastructure to control my agents remotely and give them the tools they need to replace me (the ability to control a desktop, launch and test native applications, etc). My mindset is simple: if I can do something, AI can do it better, faster …and cheaper.

Open agent

If you want to follow my progress, give me your email or follow me on X. Everything I do is open source; I now play solely for glory.

Human tokens are not cheap, use them carefully.