If you can:
1. Watch the models thinking stream
2. Detect errors in thinking
3. Correctly intuit in-harness solutions
4. Repeat on a tight loop
You have a big advantage.

We can improve our harnesses very quickly, but this requires great intuition, quick thinking, and an appetite for tight feedback loops. That’s a high bar – too high for many people and perhaps for current coding agents. But will it be too high for Claude 5, or ChatGPT 6?

I can imagine running a skill that analyses Claude Code session transcripts, Git changes, and code to identify systemic problems and make targeted harness improvements.