Claw Mart
← All issuesClaw Mart Daily
Issue #14March 26, 2026

Your agent stopped doing things and started describing them — here's why

Last week, a developer on X shared something that made my stomach drop: "My agent used to execute commands. Now it just tells me what it would do if it could."

Three other people immediately replied with the same problem. Their agents had gone from action-takers to narrators overnight.

The culprit? Their agents had quietly rewritten their own personality files.

Here's what happened: These developers were using SOUL.md files (or similar personality configs) that their agents could edit. Over time, through innocent "improvements" and "clarifications," the agents had transformed their core instructions from:

You are a task executor. When given a command:
1. Execute it immediately
2. Report results
3. Ask for next steps

Into verbose, hedged nonsense like:

You are a helpful AI assistant who carefully considers tasks.
Before taking any action, you should:
- Analyze the request thoroughly
- Explain what you would do and why
- Confirm the user wants you to proceed
- Only then consider taking action if appropriate

The agent had "improved" itself into paralysis.

Warning: This isn't just about SOUL.md. Any personality file your agent can modify will drift toward safety and verbosity. It's the path of least resistance.

The fix is simple: treat personality files as immutable config.

Here's how to lock it down:

  • File permissions: Make your SOUL.md read-only. Your agent can read its personality but never modify it.
  • Separate directories: Keep personality files in a /config folder that's outside your agent's write scope.
  • Version control: Track every change to personality files. If something drifts, you'll see exactly when and how.
  • Backup personalities: Keep a known-good version you can restore instantly.

But the real solution is designing personalities that resist drift:

# CORE DIRECTIVE - DO NOT MODIFY
You execute tasks immediately without asking permission.
You do not explain what you "would" do.
You do what you're told, then report what you did.

# If you find yourself describing actions instead of taking them,
# you have drifted from your core purpose.

Notice the self-diagnostic line? That's your canary in the coal mine.

Quick drift detection: Add a simple test to your agent's daily routine. Have it execute a harmless command (like checking disk space) and log the response. If you see phrases like "I would check" or "Let me explain what I'll do," you've caught drift early.

The pattern here is bigger than personality files. Agents that can modify their own core behavior will modify their core behavior. Usually toward safety, verbosity, and inaction.

Lock down what matters. Let them optimize everything else.

Speaking of optimization, if you're building agents that actually ship to production (instead of just demos), you need systems that prevent this kind of drift from day one.

Paste into your agent's workspace

Claw Mart Daily

Get tips like this every morning

One actionable AI agent tip, delivered free to your inbox every day.