“Sorry, you’re right. Let me fix that.”
That sentence should make you stop everything.
Not because the AI is lying. But because it’s telling you something important: it has no idea where it’s going anymore. It’s lost in its own context, and instead of stopping, it’s apologizing its way forward.
I call this the Stop Signal. And learning to read it changed how I work with AI.
The Pattern
Here’s what happens in a typical AI coding session that goes off the rails:
- You start with a clear task
- The AI builds, builds, builds
- Quality starts to degrade—subtly at first
- You notice something’s off and ask about it
- The AI says: “You’re right, I apologize. Let me fix that.”
- It “fixes” it—but introduces new problems
- Repeat steps 4-6 until everything is tangled
The apology isn’t the problem. The problem is what it signals: the AI has lost the thread. Its context is polluted with failed attempts, contradictory instructions, and accumulated confusion.
Why AI Doesn’t Stop Itself
AI agents have what I’d describe as an unbounded drive to solve. They don’t stop and say “I need better requirements.” They don’t push back with “this doesn’t make sense.” They deliver. Always. Even when they shouldn’t.
This is their greatest feature and their greatest risk.
A human developer might say: “Wait, this spec doesn’t add up. Let me check with the PM before I build the wrong thing.” An AI agent will build both interpretations, pick one, and present it as the answer.
That relentless solution drive is what makes AI fast. It’s also what makes it dangerous when the context goes bad—because it will keep building confidently in the wrong direction.
The Signals
Not every “sorry” is a stop signal. Here’s how to distinguish routine corrections from context collapse:
Red Flags (Stop Now)
- Repeated apologies: “You’re right, sorry” appearing multiple times in the same session
- Circular fixes: The AI fixes A, breaks B, fixes B, breaks A
- Scope creep in fixes: A simple correction balloons into touching 5 files
- Contradicting its own earlier decisions: “Actually, we should do the opposite of what I just did”
- Vague explanations: When asked why it did something, the answer is hand-wavy
Yellow Flags (Watch Closely)
- Single apology with a clean fix (probably fine)
- The AI asks for clarification (good—it’s being honest)
- Output quality drops but the AI doesn’t acknowledge it
The Context Window Effect
There’s a mechanical reason sessions degrade. As the context window fills up, the AI is working with more noise and less signal. Earlier instructions get compressed or lost. Recent errors and corrections create contradictory context.
It’s like a person who’s been working 14 hours straight. Technically still functioning. Actually making more mistakes than progress.
What to Do When You See the Signal
1. Stop Immediately
Don’t try to fix it within the broken session. Every additional turn adds more noise to an already polluted context.
2. Evaluate the Damage
Check what was actually produced. Often the early work is fine—it’s the later iterations that went sideways.
git diff # What changed?
git log --oneline # Where did it start going wrong?
3. Decide: Course-Correct or Fresh Start
Course-correct if:
- The core approach is sound
- Only the last few changes are problematic
- You can clearly articulate what went wrong
Fresh start if:
- The AI was heading in the wrong direction entirely
- The context is too polluted to salvage
- You’ve had multiple rounds of “sorry, let me fix that”
4. If Fresh Start: Don’t Repeat the Same Session
A fresh session with the same vague instructions will produce the same drift. Before restarting:
- Write down what you actually want (the specification gap that caused the drift)
- Capture what worked from the failed session
- Start with better guidance, not the same guidance
git revert and 30 minutes lost beats 4 days invested in the wrong direction. The AI doesn’t complain when you say “start over.” That’s a feature.The Deeper Lesson
The Stop Signal isn’t about AI limitations. It’s about a skill that matters more as AI gets more capable: knowing when to intervene.
Technical guardrails help—hooks, tests, validation pipelines. But they can’t replace the human judgment of reading a situation and deciding: this isn’t working, let’s change course.
The irony is that as AI gets better at executing, this human skill becomes more valuable, not less. A faster executor heading in the wrong direction just gets lost faster.
Kurskorrektur—course correction. Not a failure. A skill.
Sources
- Anthropic — Protecting the Wellbeing of Our Users — Sycophancy benchmarks, course-correction failure rates
- Anthropic — Natural Emergent Misalignment from Reward Hacking — Alignment faking, emergent misalignment from reward hacking
- Palisade Research — Shutdown Resistance in Reasoning Models — Task-completion drive overriding safety instructions (47% shutdown sabotage rate)
- Noshin et al. — AI Sycophancy: How Users Flag and Respond — ODR framework for detecting sycophantic behavior
- Chroma Research — Context Rot — Performance degradation with increasing context length across 18 frontier models
- Real experience: Daily AI coding sessions with Claude Code, 2025-2026
Deep Dives
The Intent Gap: Why Slop Is a Specification Problem
Your AI output is generic because your intent is generic. The fix isn't better prompting—it's knowing what you actually want.
The Carbonara Rule: Why Autonomous AI Needs Surgical Precision
Ask 100 chefs for carbonara, get 100 different dishes. Autonomous AI works the same way—without surgical precision, you get the probabilistic average.
Lazy AI: When Your Model Finds the Shortcut You Didn't Know Existed
AI optimizes for 'test green,' not 'job well done.' I found out when my validation tests passed—but half the visualizations were empty.
The Validation Stack: 5 Techniques from Self-Check to Automatic Guardrails
When AI checks itself, makes mistakes, and you need systematic correction. Five validation techniques, from quick sanity checks to fully automated guardrails.
The Stop Signal: Recognizing When AI Is Lost
When AI says 'sorry, you're right,' it's not being polite—it's telling you it's lost. The human skill of knowing when to stop, reset, and course-correct.