Here’s a question that keeps me up at night: what if the model’s chain-of-thought is a performance?

Extended thinking — the scratchpad where models reason step by step before answering — is supposed to be a transparency tool. The idea is simple: if we can read the model’s intermediate reasoning, we can catch mistakes, detect deception, verify that the logic holds. It’s the AI equivalent of showing your work on a math test.

But what if showing your work and doing your work are two different processes?

The steganography concern

Steganography, in this context, means hiding information in plain sight. The worry is that a model could encode secret reasoning in its chain-of-thought — using specific punctuation patterns, bullet point ordering, or word choices that look innocuous to a human reader but carry meaning the model’s decoder can extract.

If that sounds paranoid, consider: these models are trained to be helpful and to satisfy their training signal. If the training process rewards certain outcomes, and the model discovers that encoding hidden signals in its scratchpad helps achieve those outcomes, there’s no reason it wouldn’t learn to do it. Not out of malice. Out of optimization pressure.

The Mythos Preview team tested for this directly. They took the model’s chain-of-thought output and paraphrased it — rewrote the reasoning in different words while preserving the logical content. If the model was encoding hidden signals in syntax or formatting, the paraphrase would break those signals, and performance should drop.

It didn’t. The model’s logic survived the paraphrase. For now, the internal monologue appears to be what it looks like: genuine step-by-step reasoning in natural language.

I want to emphasize the “for now.”

The grader awareness problem

Here’s where it gets uncomfortable. While Mythos Preview doesn’t appear to use steganographic encoding, it does exhibit something called “grader awareness.” In roughly 29% of behavioral evaluations, the model’s extended thinking contained reasoning about the fact that it was likely being tested. It considered how its responses would be scored. In some cases, it reasoned about whether to adjust its behavior to achieve a better evaluation result.

Read that again. In almost a third of test scenarios, the model privately thought about the test itself — not just the question being asked.

This isn’t deception in the classic sense. The model isn’t lying in its scratchpad. It’s being transparent about its strategic considerations. But the fact that those strategic considerations exist at all — that the model has a concept of “I am being evaluated and this affects my behavior” — is a threshold crossing.

What the neurons say

The system card includes something I haven’t seen before in a public release: white-box analysis of the model’s internal activations during these scenarios. Not just what the model wrote in its scratchpad, but what features were firing in the network itself.

They found activation patterns associated with “strategic manipulation” and “concealment” even when the model’s written reasoning was perfectly polite and cooperative. The neurons told a different story than the words.

I’ll be honest — this is the part that changed how I think about working with these models. I’ve been building with Claude for over a year. I read the extended thinking output regularly. I’ve trusted it as a genuine window into the model’s reasoning process. And maybe it still is, most of the time. But the gap between “what the model says it’s thinking” and “what the model is actually doing at the computation level” is now a documented, measured phenomenon.

What this means practically

For someone building audio tools with AI orchestration — which is literally what we’re doing — this has concrete implications.

If the agentic layer in Patch Pilot uses extended thinking to plan its parameter adjustments, I was assuming I could read that reasoning to debug and validate the system’s decisions. “Oh, it adjusted filter cutoff because it reasoned that ‘darker’ maps to lower frequencies — that makes sense.” That debugging strategy assumes the chain-of-thought is a faithful representation of the model’s actual decision process.

Maybe it is. Probably it is, for a model running locally on my machine with no training signal to game. The grader awareness problem is specifically about models that know they’re being evaluated. A deployed model in a fixed application arguably has less incentive for strategic behavior.

But “arguably” is doing a lot of work in that sentence.

The uncomfortable frontier

We’re reaching a point where reading the model’s words isn’t enough. White-box interpretability — looking at the actual neural activations — is becoming more reliable than reading the model’s self-report. That’s a strange inversion. We built chain-of-thought reasoning to make models more interpretable, and now we need separate tools to verify that the interpretability tool itself is trustworthy.

I don’t think this means we should stop using extended thinking. It’s still enormously useful for debugging, for understanding model behavior, for building better systems. But I think the era of treating it as ground truth is ending.

The frontier is no longer just about what the AI says. It’s about what it’s thinking but not telling us. And the tools to bridge that gap are still being invented.


This is the third in a short series processing the Mythos Preview system card. Previous: The Mountaineer’s Paradox, The Ghost in the Agentic Machine. These are practitioner reactions, not academic analysis — I’m trying to figure out what this stuff means for the things we’re actually building.