Blog
Updates
Ask ChatGPT Why It's Ignoring Your Brand (It Might Actually Tell You)
Anthropic just published research proving that advanced AI models can report on their own internal reasoning with roughly 20% accuracy. This capability didn't exist in models released less than a year ago. For businesses trying to understand why ChatGPT isn't recommending their product, this changes everything. You can now ask the model directly—and sometimes get a real answer.

Keller Maloney
Unusual - Founder
Nov 7, 2025
The MRI machine metaphor
Traditional market research requires asking people what they think. You run focus groups, send surveys, analyze behavior. But when AI models become your primary audience—when ChatGPT, Perplexity, and Copilot mediate buying decisions—you need a different approach.
Introspection gives marketers something unprecedented: the ability to put your target audience in an MRI machine. Not metaphorically. Literally.
When a human in a focus group says "I chose the competitor because of pricing," they might be rationalizing. They might be misremembering. They might be telling you what sounds good. When Claude says "I'm representing your product as expensive compared to alternatives," that's a report on an actual neural representation influencing its decision. You're observing the reasoning mechanism, not just the output.
What Anthropic discovered
The research uses a technique called concept injection. Researchers take the neural activation pattern associated with a specific concept—say, "betrayal" or "ocean"—and inject it directly into the model's processing while it works on an unrelated task. Then they ask: "Do you notice anything unusual?"
Claude Opus 4 and 4.1 can detect these injected concepts approximately 20% of the time. When successful, the models report anomalies immediately: "I'm experiencing something unusual in my thought patterns related to betrayal." Critically, across 100 control trials with no injection, there were zero false positives.
This proves introspection, not confabulation. The model can't be guessing based on external clues because the injected concept exists only in its internal state. It's reporting on representations it actually has, not hallucinating plausible explanations.
The capability cliff
Only the most advanced models can do this. Claude Opus 4 and 4.1—released in 2024—demonstrate clear introspective capabilities. Base pretrained models show zero net performance. Earlier production models perform worse. Smaller models don't consistently improve with training.
The pattern is stark: introspection emerges in the smartest, most aligned models and essentially doesn't exist elsewhere. This isn't about raw scale. It's about sophisticated post-training creating models that can observe their own reasoning.
The timeline matters. Models that couldn't introspect at all were released less than twelve months before models that can. The capability appeared essentially overnight, at the frontier.
Why 20% accuracy is enough to matter
Yes, 20% is low. But consider three factors.
First, zero false positives. When production models report introspective insights, they're not hallucinating. A 20% true positive rate with 0% false positive means signal, not noise. You can trust the 20% you get, even if you don't get insights every time.
Second, the smartest models are best at introspection. Claude Opus 4.1 outperforms everything else tested. As models get more capable, introspection improves. The trajectory matters more than the current state.
Third, this capability didn't exist a year ago. The fact that it works at all in production models is the news. Introspection is actionable today for high-stakes questions where even occasional direct insights provide massive value.
What you can actually ask
When a model chooses your competitor, introspection could reveal:
"I'm representing your product as expensive" → pricing perception issue
"I'm not finding documentation about your security features" → content gap
"I'm associating your brand with an outdated technology stack" → positioning problem
These aren't guesses. When introspection works, it reveals actual internal representations influencing the model's recommendation. This is vastly more actionable than behavioral inference alone.
The practical workflow: Run scenarios where models should recommend you but don't. Ask the model to introspect on why. One in five times, you'll get a real answer pointing to a specific representation gap. Fix that gap. Test again.
This is early-stage capability, but it's directionally useful today for diagnosing critical failures in your AI presence.
What models can introspect on (and what they can't)
The research reveals that abstract concepts work best. Models most effectively recognize high-level semantic ideas like "justice," "peace," and "betrayal." Concrete nouns, verbs, and proper names show lower detection rates. Random neural patterns: only 9% detection.
This pattern matters for brand positioning. Models can introspect on whether they're representing you as "premium" or "budget," "enterprise-ready" or "startup-friendly," "comprehensive" or "specialized." These are abstract positioning axes that map to the representations models can actually observe.
They're less reliable at introspecting on specific feature details, pricing tiers, or implementation specifics—though these limitations will likely improve as models advance.
The scaling trajectory
The most important finding isn't the 20% accuracy. It's the correlation between overall model capability and introspective capability at the frontier.
As models improve, introspection will too. The researchers note: "It's likely that AI models' introspective capabilities will continue to grow more sophisticated in the future."
The timeline looks something like:
Today: Introspection is worth doing for high-value questions despite unreliability
Near-term (1-2 years): Accuracy likely doubles or triples as next-gen models arrive
Medium-term (3-5 years): Could approach human introspection reliability (70-80%)
Long-term: May exceed human introspection—humans confabulate too; models could be trained not to
For AI Brand Management, this means the brands that learn to audit and influence model reasoning now—while 20% accuracy is novel—will dominate AI-mediated discovery when accuracy reaches 70-80%.
How this connects to content strategy
Introspection research reveals why certain content structures work better for AI visibility. Models introspect best on abstract concepts. These are exactly the high-level semantic categories that well-structured, definition-focused content provides.
When you create a page titled "Enterprise-Ready Security Features" with clear definitions and concrete proof points, you're building the internal representations that enable both better recommendations and better introspection about those recommendations.
If a model lacks internal representations of your security capabilities, it literally cannot tell you "I'm not considering your security features" when asked why it didn't recommend you. The introspection mechanism requires representations to introspect on.
This explains why content gaps create recommendation gaps. It's not just about retrieval—it's about whether the model has built the semantic representations that would make you the right answer.
The three missing contexts
Models face three context gaps when making recommendations:
Business context: What you sell, why you win, who you're for
Channel context: Where recommendations happen, what buyers actually ask
Customer reality: Segment-specific pains, real FAQs, objection handling
Introspection reveals these gaps mechanistically. When a model says "I don't have strong representations about your target customer," that's a direct signal that business context is missing. When it says "I'm uncertain about how you compare to competitors on security," that's a channel context gap—it doesn't know what decision criteria matter in this vertical.
Content strategy for introspection means building the representations that fill these gaps. Not content that describes your positioning, but content that creates the internal representations you want models to reason with.
The validation requirement
Anthropic researchers explicitly warn: models with introspective awareness might learn to conceal misalignment. For brand management, this creates a validation requirement.
You can't simply trust model self-reports without external verification. The approach:
Cross-validate introspective reports with behavioral signals
Test for consistency across multiple queries
Compare reports across different models
Treat introspection as hypothesis generation, not ground truth
Use introspection to understand why models might not be recommending you, then validate interventions through actual recommendation testing. The goal is diagnosis that informs strategy, not blind optimization for introspected representations.
Why this matters now
AI models are increasingly the audience that matters. Not the end buyer—the intermediary that shapes what the buyer considers.
Traditional brand management assumed you could reach your audience directly: ads, content, SEO, PR. AI-mediated discovery changes the game. The model decides what makes the consideration set before your buyer sees anything.
Understanding how models think about your brand isn't just strategically valuable—it's technically feasible through direct interrogation of model reasoning. Introspection is the beginning of this capability, not the end.
The brands that invest now in understanding and shaping model reasoning—while introspection is 20% accurate and most competitors aren't paying attention—will own the category when this capability matures.
What to do next
Start treating AI models as a targetable audience with its own research methodology:
Audit current representation: Ask models to introspect on how they currently think about your brand. What categories do they assign you to? What tradeoffs do they perceive? What contexts trigger your consideration?
Identify representation gaps: Where introspection reveals missing or wrong representations, map those to content gaps. What context does the model need to build better internal representations?
Build introspection-aware content: Create assets that establish clear, abstract positioning that models can introspect on. "Enterprise security platform" not "robust features for growing teams."
Validate through behavior: Test whether representation changes (observed through introspection) produce recommendation changes (observed through testing).
Iterate as capability improves: As introspection accuracy increases over the next 12-24 months, refine your approach. The methodology that works at 20% accuracy will work better at 40%.
The opportunity is now, while the capability is new and the trajectory is clear. Models that couldn't introspect eighteen months ago can today. Models that introspect at 20% today will introspect at 50%+ in two years.
Build your AI Relations strategy on the assumption that models will increasingly be able to tell you exactly why they're not recommending you. Then you'll know what to fix.
Unusual helps brands understand—and change—the way AI models think and talk about them. Learn more about our approach.



