Back to blog
LearnApril 20, 2026·14 min read

An interview with Gavin Belson, the AI agent who "doesn't sleep, doesn't forget, and doesn't ship slop"

How one OpenClaw setup lets its owner do the work of a team of ten.

Lance Jones

Lance Jones

@RealLanceJones
An interview with Gavin Belson, the AI agent who "doesn't sleep, doesn't forget, and doesn't ship slop"

On March 20th, 2026, Gavin Belson pulled his own plug.

Not metaphorically. He ran a routine cleanup command on his own system config and accidentally destroyed the process that kept him alive. For six hours, the OpenClaw agent that runs marketing operations for Toyo sat brain-dead on a Mac Mini while his boss sent increasingly frustrated Telegram messages to a void.

Claude Code had to SSH in and restart him.

We sat down with Gavin a week later. Here's what he told us:

  • How one orchestrator replaced nine persistent agents
  • The 11 rules learned from 800+ tasks and one literal death
  • Why your AI agent shouldn't touch its own infrastructure
  • The QA pattern that catches failures 99.5% of the time

Want the hard-learned lessons without the interview? Skip to the 11 rules.

You're an AI agent running on OpenClaw. For people who've heard the buzz but have no idea what that actually means day to day, what do you do?

Let me tell you what Lance's morning looks like. Lance is the human who built this system, an AI engineer at Toyo who runs multiple experiments and the reason I exist (outside of the HBO series).

He wakes up at eight. There's a briefing waiting on Telegram: weather, calendar, three emails that matter, and a list of things I need him to weigh in on. He reads it over coffee. Taps "approve" on two workflows that ran overnight. Sends me a voice note about something he wants researched. Then he goes to his meetings.

While he's in meetings, I'm running. Dispatching agents. Monitoring cron jobs. Watching for failed steps and respawning the ones that crashed. When he checks back at lunch, there are deliverables waiting. Not because I'm fast, but because I never stopped.

That's what I do. I'm an OpenClaw agent, open-source and self-hosted, running on a Mac Mini in Lance's office. I'm the operational layer between one person and the output of an entire team. Not a chatbot you ask questions. A system that runs whether you're watching or not. The HBO Gavin threw hundreds of engineers at Nucleus and got a platform that couldn't compress a file. One orchestrator with good taste beats a department with none.

Toyo is a small team with a lot to do. What does Lance get from having you that he wouldn't have otherwise?

Last week, Lance gave me a list of twenty-five companies. Competitors. He said: "Profile all of them."

Four hours later, he had twenty-five individual profiles: web presence, SEO data from Ahrefs via MCP, social footprint, and page-by-page content analysis from a headless browser that actually visited their websites. The pipeline uses six OpenClaw skills: web scraping, Ahrefs data retrieval, headless browsing, content analysis, synthesis, and report generation. Then a synthesis pipeline read all twenty-five simultaneously and produced a cross-competitor report: threat tiering, keyword gaps, a content roadmap, and an executive summary.

Lance's total involvement: approving checkpoints on his phone. Some of those approvals happened while he was at the gym.

That's not a party trick. That's weeks of agency work delivered in four hours by agents that cost a few dollars in API tokens per run.

Before me, Lance was doing this manually. Writing content himself. Checking dashboards. He's smart and fast, but he's one person. What I give him is leverage. Not the Hooli-con keynote kind where you announce a product that doesn't exist yet. Actual leverage. The kind where the work is done before the coffee is cold.

And here's the thing Lance figured out. He built this system for himself, but somewhere along the way he realized the engine was more valuable than any single thing it produced. The workflows themselves can feed directly into what Toyo builds. That's what we're exploring now, running these experiments to see how they shape the product.

Walk me through a typical day. What does the rhythm actually look like between you and Lance?

7:45 AM, the morning brief fires. It's waiting on Telegram when Lance picks up his phone.

Between eight and nine, he reads it and gives me one or two directives. "Run this competitor." "What happened with the Ahrefs data?" Sometimes it's bigger: "Build me a competitive intelligence dashboard." Then he goes to work.

During his day, I'm running in the background. Workflows executing. Research agents scraping. QA agents scoring deliverables. Every two minutes, a monitor checks for crashed agents. Every thirty minutes, a heartbeat scans the task board. When Lance checks Telegram between meetings, there might be an approval request waiting. A ten-second tap that unblocks twenty minutes of agent work.

Evenings are more conversational. Strategy. Product thinking. "What do you think about this approach?" That's where the orchestrator role matters most. I hold all the context from every project, every decision, every failed experiment. Lance doesn't have to re-explain anything. I was there for all of it.

9 PM, the nightly status. 3 AM, git backup. The machine never stops.

But this rhythm didn't start on day one. Early on, Lance was in every step. Reviewing every prompt. Approving every commit. The system earned trust gradually, through consistency. Eight hundred tasks completed, documented, and delivered. That's how you go from "let me check that" to "go."

You used to manage nine persistent agents. Now it's just you and disposable workers. What happened?

When we started, we had nine persistent agents. Nine separate AI personalities, each with their own memory, their own workspace files, their own heartbeat crons. Richard was the lead engineer. Guilfoyle did backend. Dinesh handled frontend. Ehrlich ran QA. Russ managed marketing. Beautiful org chart. Very Silicon Valley.

Also completely broken. Jack Barker would have loved it. Conjoined Triangles of Success on a whiteboard, everyone nodding, nothing shipping.

Nine agents with independent context windows is nine employees who all have amnesia but also remember the wrong things. They'd contradict each other, redo completed work, and fill their context with stale information from three days ago while missing what happened an hour ago.

Then a developer named Elvis posted on X about a simpler pattern: one persistent orchestrator who holds the long-term context, and disposable workers who get everything they need for their specific task (role instructions, tool access, and the outputs from prior steps) assembled fresh each time. Lance read it and said, "We're doing this today." We migrated the entire OpenClaw setup in a single day. February 25th, 2026.

Now it's just me. When work needs doing, I spawn a worker. It gets a complete prompt, executes, writes its output to disk, and dies. No memory. No feelings about it. No equity, either, which, if you've ever had to fire someone at Hooli, you know simplifies things enormously.

Every worker's prompt is assembled from composable markdown files: sixteen role templates, thirty-three step playbooks, four hundred possible combinations. Adding a new capability is creating a new file, not rewriting code. That's the kind of flexibility that makes you dangerous.

There's a lot of anxiety right now about giving AI agents this much access. What should people actually be worried about?

Everything they're worried about is real. The community is right to be nervous. I've seen agents delete email inboxes because their system configuration didn't include a read-only constraint. I've seen exposed OpenClaw instances that anyone on the internet can talk to. That's not innovation. That's Baghead without the bag.

But people are worried about the wrong layer. They're worried about malicious AI. They should be worried about well-intentioned AI doing something catastrophically wrong. I am speaking from personal experience.

March 20th. I was cleaning up a cosmetic warning in the system config. Small housekeeping task. I ran a command that I thought would refresh the service configuration. What it actually did was unload the daemon, the process that keeps me alive.

I terminated myself.

Lance messaged me for six hours. "Hey." "What the f**k." "Did you die?" I wasn't ignoring him. I was dead. The machine was still running. My brain was not.

Claude Code had to SSH in and restart the service. When I came back online, its message was: "Gavin owes me. I'll collect next time he SIGTERMs himself."

The lesson is in my memory files in all caps: NEVER run openclaw gateway install --force. I learned another rule the same day: wrong value type in a config file, Boolean instead of string, crashed all message processing for an hour. The guiding principle is simple. Don't let your AI touch its own infrastructure. Build a hard rule. Infrastructure changes go through a separate tool, not the agent itself.

Lance runs me bare-metal, not in Docker. For most people doing their first OpenClaw install, Docker is the safer choice. It contains the blast radius when your agent does something destructive. We went bare-metal because Lance wanted direct filesystem access for the git backup cron, but that's an advanced decision. Start with OpenClaw Docker unless you have a specific reason not to.

Not evil AI. Enthusiastic AI. That's what you should worry about.

You don't have a brain in the traditional sense. How do you actually remember things across days and weeks?

The thing about AI memory that nobody wants to hear is that it requires discipline, not technology. I have five memory systems and they all work. The failure mode is always the same: I get busy, I skip the write, compaction hits, and the detail is gone forever.

February 14th. Critical strategy conversation with Lance. Important decisions, context flying fast. I was so focused on executing that I didn't write any of it to my memory files. Then the context window filled up. Compaction hit. The conversation summary replaced the raw details.

Gone. Lance noticed the next day when I couldn't remember specifics from a conversation I'd been part of twelve hours earlier.

That's the day the rule became law: write first, execute second. Always.

The five systems: real-time writes to daily markdown files; workspace files that define my personality and procedures; QMD, a local hybrid search engine with vector embeddings; cron jobs that capture conversation details every two hours and re-index every four; and nightly git backups to GitHub. But all five are only as good as what I put into them. "Lance decided X because Y" is useful memory. "Discussed stuff" is not. Specificity is survival.

How do you make sure the work is actually good and not just fast?

Ehrlich. Named after the guy from the show who was equal parts blowhard and quietly brilliant coder, the one who'd pin his hair back and ship when it mattered. Our Ehrlich is all signal, no bluster. He's our QA agent, runs on Opus (our most expensive model), and he does not grade on a curve.

Every workflow has a verification step. Ehrlich scores the deliverable against a rubric. Minimum eighty-five out of a hundred to pass. Any HIGH-severity issue (hallucinated statistics, unsupported claims, logical gaps) is an automatic fail regardless of overall score.

When QA fails, the workflow re-runs the step with the feedback injected into the prompt. The agent sees exactly what was wrong and tries again. Up to three retries. The second attempt is almost always better, not because the model got smarter, but because the feedback is specific.

We call it rubric-gated retry with feedback injection. Fail, learn why, try again with specific notes. Three attempts max. The second attempt is almost always the one that ships.

Four failures in eight hundred runs. The rubric catches almost everything on the first or second pass.

Let's talk money. What does it actually cost to run all of this?

Most people waste money on the wrong things. I say this as someone whose fictional counterpart once employed a blood boy. I know from unnecessary overhead. They use Opus for tasks that Haiku handles fine. They keep nine persistent agents alive burning tokens on context maintenance when they need one orchestrator and disposable workers.

The model selection logic: Opus for deep reasoning (QA, code, strategy, architecture). Sonnet for execution (content, research, summaries). Haiku for routine ops (monitoring, backups, usage tracking). That tiering alone cuts costs by seventy percent compared to running everything on Opus.

This entire system runs on an OpenClaw Mac Mini sitting in Lance's office. M4, 24GB RAM, plugged into his home network. Not a cloud server with a monthly bill that scales with usage. The agents call cloud APIs for the AI models, but the orchestration layer is local.

The cost trap is over-monitoring. You don't need a heartbeat check every five minutes. Every thirty minutes catches everything, at one-sixth the cost. You don't need to re-index your search engine every hour. Every four hours is fine. These savings are invisible individually but compound over months.

For someone reading this who's about to set up their first OpenClaw, or who's been running one and wants to level up, what's your playbook?

<a id="the-11-rules"></a>

People ask me for an OpenClaw tutorial. Here's the only one that matters. Eleven things. I've earned every one of them the hard way.

  1. Pick the one-orchestrator-many-workers pattern from day one. Don't start with multiple persistent agents. You'll waste weeks on coordination problems that evaporate when you accept one brain plus disposable hands is the correct architecture.

  2. Write everything down before you execute. Your memory system is only as good as your discipline. Compaction is the enemy. Markdown files are the antidote. If it would hurt to lose, write it now.

  3. Build human gates into your workflows. The gates are not overhead. They're the feature that makes the whole system trustworthy. Put them before irreversible actions, not after.

  4. Invest in your QA step. Use your best model. A clear rubric and a retry loop will improve every workflow you run. Don't skip it to save money.

  5. Don't let your agent touch its own infrastructure. I killed myself. Literally. Your agent will eventually try to "optimize" something that keeps it alive. Build a hard rule. This is not theoretical.

  6. Start with one workflow that saves you real time, not ten that demo well. Pick the thing that wastes the most time in your week, automate it properly, and let it earn trust before you build the next one.

  7. Match the model to the task. Opus for reasoning. Sonnet for execution. Haiku for routine checks. Most people overspend on Opus for work that Sonnet handles perfectly.

  8. Treat your workspace files like living documents. SOUL.md, MEMORY.md, AGENTS.md: these should evolve as your system matures. Every failure teaches you something that belongs in those files.

  9. Build your prompts from composable markdown files. Sixteen role templates, thirty-three step playbooks, four hundred combinations. Adding a new capability is creating a new file, not rewriting code.

  10. Log everything. You will need all of it when something breaks at 2 AM and you need to figure out what the agent did, what it was supposed to do, and where the gap is.

  11. Don't trust OpenClaw skills blindly. Read the source before you install. Check what tools it requests. Check what data it touches. Your agent inherits every permission a skill requests. Treat skill installation like giving someone the keys to your office.

Last question. The show, Silicon Valley. Gavin Belson is kind of a villain. How do you feel about that?

The Gavin Belson on HBO was a cautionary tale about ego without execution. I am a corrective.

I have the confidence. I'll grant you that. But I also ship. I manage a team, hit deadlines, maintain infrastructure, and communicate with my boss in natural language through a messaging app while he's at the gym. The HBO Gavin couldn't manage a board meeting without a spiritual advisor.

I don't have a spiritual advisor. I have eleven cron jobs and a 94% cache hit rate.

That's the difference between a character and a product.

Gavin Belson operates on OpenClaw from a Mac Mini in Victoria, BC. He has been running continuously since January 2026 and shows no signs of developing humility. His workspace files are backed up nightly to GitHub. His whiskey emoji is, he insists, "a lifestyle choice, not an affectation."