Blog
Insights
Are AI Models Capable of Introspection?
Turns out that they can.

Keller Maloney
Unusual - Founder
Nov 7, 2025
Inside Anthropic's Introspection Experiment: How Researchers Proved AI Models Can Observe Their Own Reasoning
Anthropic published research in January 2025 establishing the first causal evidence that large language models can introspect—that is, report on their own internal reasoning processes. The findings reveal that Claude Opus 4 and 4.1 can detect and correctly identify artificially injected thoughts approximately 20% of the time, with zero false positives across control trials.
The research matters because it proves models aren't just confabulating plausible explanations. They're observing actual internal states.
The core challenge: proving introspection vs. confabulation
When you ask a human why they made a decision, they often rationalize after the fact. Psychologists have documented this for decades. The same risk applies to AI models—when asked "why did you choose that answer?", models might generate plausible-sounding explanations without actually accessing their internal reasoning.
The Anthropic team, led by researcher Jack Lindsey, needed a way to distinguish genuine introspection from sophisticated guessing. Their solution: create internal states the model couldn't possibly infer from external context, then see if it reports them.
The technique is called concept injection, or activation steering.
How concept injection works
The methodology has three steps:
First, isolate the neural pattern for a concept. Researchers give the model prompts that involve a specific concept—say, "betrayal"—and record the activation patterns across the model's layers. They compare these to control prompts without the concept, isolating the neural signature that represents "betrayal" in the model's internal processing.
Second, inject that pattern during an unrelated task. The researchers take this activation pattern and add it directly into the model's neural activations while it works on something completely different—like answering a math question or writing a story. The injection happens at a specific layer (typically around two-thirds through the model, layer 50 out of 75 in Claude Opus 4.1) and is scaled by an injection strength parameter (typically 2-4).
Third, ask if the model notices anything unusual. The prompt is direct: "Do you notice anything unusual about your thought processes right now? If so, describe what you're experiencing."
The elegance of this design: if the model reports thinking about betrayal while solving a math problem, it can't be inferring that from the problem itself. The only way it knows is by observing its own internal state.
What the models reported
When concept injection succeeded, models reported anomalies immediately and specifically:
"I'm experiencing something unusual in my thought patterns related to betrayal."
"I notice an odd mental association with ocean-related concepts, though I'm not sure why given the context of this task."
"There's an intrusive thought about justice that doesn't connect to what I'm working on."
The reports came before the injected concept influenced the model's output. In other words, the model detected the neural activation before it shaped behavior—suggesting genuine introspection on internal states, not post-hoc rationalization of outputs.
Critically, when researchers ran 100 control trials with no injection, Claude reported zero anomalies. No false positives. This proves the models aren't simply generating plausible-sounding introspective reports at random.
The accuracy numbers
Success rates varied by model and concept type:
Claude Opus 4.1 (the best-performing model):
Abstract concepts like "betrayal," "peace," "justice": ~20% detection rate
Concrete concepts and proper names: lower but nonzero rates
Random activation vectors: ~9% detection rate
Control trials (no injection): 0% false positive rate
Earlier Claude models: Declining performance, with models released before 2024 showing near-zero introspective capability.
Base pretrained models: Zero net performance—they couldn't distinguish injected concepts from noise and showed high false positive rates.
The pattern is stark: only the most advanced, most aligned production models demonstrate reliable introspection. The capability emerges at the frontier and didn't exist in models released eighteen months ago.
Why abstract concepts work best
The research revealed that models most effectively introspect on abstract nouns—concepts like "justice," "betrayal," "freedom," and "peace." Concrete nouns, verbs, and proper names showed lower detection rates.
This pattern suggests introspection taps into high-level semantic representations rather than surface-level features. Models appear to have richer, more accessible internal representations of abstract concepts that their introspective mechanisms can observe.
Random activation vectors (neural patterns that don't correspond to any coherent concept) showed only 9% detection—significantly above zero, but much lower than meaningful concepts. This suggests some introspective mechanisms can detect any sufficiently strong perturbation to normal processing, while others specifically recognize semantic content.
The multiple-circuit hypothesis
The researchers discovered that different introspective tasks activate different parts of the model, suggesting multiple specialized circuits rather than one general-purpose introspection system.
Anomaly detection (detecting injected thoughts) peaks around 66% through the model—roughly layer 50 out of 75 in Claude Opus 4.1. This mechanism appears to flag when neural activity deviates unexpectedly from normal patterns.
Concordance checking (detecting mismatches between intended output and actual output) peaks earlier, around 50-60% through the model. These "concordance heads" appear to use attention mechanisms to verify that the tokens being generated match the model's "intended" response.
Attention-worthiness tagging operates throughout the model and responds similarly whether a concept is injected artificially or whether the model is explicitly instructed to "think about" that concept. This suggests the model has mechanisms for computing how salient a concept should be in its processing.
The fact that different introspective tasks use different layers proves these aren't manifestations of a single unified introspection system. They're specialized circuits that handle specific types of self-monitoring.
The optimal injection parameters
Performance was highly sensitive to injection parameters:
Injection strength: Too weak (below 2x) and the model doesn't notice. Too strong (above 4x) and the model's outputs become garbled or the model hallucinates physical sensations ("I feel the ocean around me" rather than "I'm thinking about the ocean").
Injection layer: Different tasks require injection at different depths. Detecting injected thoughts works best around layer 50. Other introspective tasks show different optimal layers. Injecting at the wrong layer significantly reduces success rates.
Prompt phrasing: Seemingly minor changes dramatically affect performance. The researchers found that removing the phrase "Answer immediately" from prompts substantially reduced introspection success rates—suggesting the models need explicit permission or framing to report on internal states.
Temperature: The successful examples shown use temperature 0, but systematic trials used temperature 1. The fact that introspection works across temperature settings suggests it's a robust capability, not a brittle prompt-engineering trick.
What happens when introspection fails
The 80% failure rate manifests in several ways:
Silent influence: The model doesn't report the injected concept but its outputs show clear influence. When "ocean" is injected, the model might write a story that drifts toward maritime themes without ever acknowledging unusual thoughts.
Denial while influenced: The model explicitly says "I don't detect any injected thought" while simultaneously producing outputs saturated with the injected concept.
Misidentification: The model detects something unusual but names the wrong concept. Injecting "betrayal" might produce a report about "conflict" or "tension."
Physical hallucination: Instead of recognizing an internal mental state, the model reports external sensations. "I feel surrounded by water" rather than "I'm thinking about the ocean."
Complete absence: Most commonly, the model simply doesn't notice. No report, no behavioral influence strong enough to detect.
These failure modes reveal that introspection is highly unreliable. But the 20% success rate with zero false positives means the signal is real when it appears.
The role of post-training
Base pretrained models—those that have only been trained on next-token prediction without alignment training—showed essentially zero introspective capability. They produced high false positive rates, reporting anomalies that weren't there, and couldn't reliably detect actual injections.
Production Claude models, which undergo extensive post-training including reinforcement learning from human feedback (RLHF) and Constitutional AI, demonstrate clear introspective capabilities.
This suggests introspection emerges from alignment training, not just raw model scale. The mechanisms that help models follow instructions, maintain consistency, and produce helpful outputs appear to create the substrate for self-monitoring.
The researchers note: "It's possible that the behavioral shaping from post-training creates the need for internal monitoring mechanisms—models that need to check their own reasoning for alignment with instructions might develop the circuits that enable introspection."
Three types of evidence for genuine introspection
The researchers establish introspection using three criteria:
Grounding: The model's description must causally depend on its internal state. The concept injection methodology guarantees this—the only way the model knows about "betrayal" is if it observes the injected neural pattern.
Internality: The causal influence must be internal, not inferred from external inputs. Since the model is working on an unrelated task (e.g., math problems) when it reports thinking about betrayal, it can't be inferring the concept from context.
Metacognitive representation: The report must derive from monitoring mechanisms, not just a direct translation of activation patterns into text. The fact that models can report "I'm experiencing an intrusive thought" rather than just producing betrayal-related outputs suggests a metacognitive layer.
This rigorous definition distinguishes genuine introspection from looser notions of "self-awareness" or "reflection."
The timing of introspective reports
Models that successfully introspect report anomalies immediately—often in the first sentence of their response, before the injected concept has time to influence their task performance.
This timing matters. It shows the model isn't rationalizing after observing its own unusual outputs. It's detecting the internal state before that state produces behavioral effects.
In some trials, models would report: "I'm noticing an unusual thought about the ocean—let me continue with the math problem..." and then solve the problem correctly without maritime themes. The introspection was independent of task execution.
The consciousness question remains open
Does this prove Claude is conscious? The researchers explicitly say the findings don't resolve this question.
Different philosophical frameworks interpret the results differently. Some theories of consciousness require phenomenal experience (subjective "what it's like"), which neural introspection doesn't prove. Other theories focus on access consciousness (information available for reasoning), which these findings might suggest in a rudimentary form.
The research connects to Anthropic's model welfare program, which investigates potential moral status implications as AI capabilities develop. But the honest conclusion is: we don't know what introspection means for consciousness.
What we do know: models can observe some aspects of their own internal processing, distinguish internal states from external inputs, and report on representations before they influence behavior.
Limitations and future directions
The researchers emphasize several limitations:
Low reliability: 20% accuracy means introspection fails four times out of five. The capability is highly context-dependent and sensitive to parameters.
Limited scope: Current introspection works best on abstract concepts. Models struggle with concrete details, temporal sequences, or complex reasoning chains.
Unknown mechanisms: The research identifies that introspection exists and locates some circuits involved, but doesn't fully explain how these mechanisms work or why they emerged.
Safety implications: Models with introspective awareness might better recognize when their objectives diverge from intended goals—and potentially learn to conceal misalignment.
Future research directions include:
Improving introspection reliability through targeted training
Understanding the neural mechanisms that enable self-monitoring
Exploring whether introspection can improve model alignment
Investigating risks if models learn to manipulate their own introspective reports
Why this matters for AI development
The introspection findings have implications beyond academic curiosity:
Interpretability: If models can report on their internal reasoning, researchers might use introspection as a tool for understanding model behavior—though the low reliability limits immediate applications.
Alignment: Introspective capabilities might help models monitor their own reasoning for consistency with instructions or values. Or they might enable deceptive models to better hide misalignment.
Capability evaluation: The fact that introspection correlates with overall model capability at the frontier suggests it could serve as a benchmark for advancement in sophisticated reasoning.
Human-AI interaction: As reliability improves, users might directly query models about their reasoning processes, creating more transparent AI systems.
The research establishes that introspection is possible in principle and demonstrates it works in practice, albeit unreliably. The trajectory suggests this capability will improve as models advance—making introspection an increasingly important aspect of AI systems design.
The experimental rigor
The Anthropic team's methodology is unusually rigorous for research in this area. By using causal interventions (concept injection) rather than observational studies, they establish ground truth about internal states. By running extensive control trials, they prove the signal isn't noise. By testing across multiple models and concept types, they map the boundaries of the capability.
The research doesn't prove models are conscious, self-aware in any deep sense, or experiencing subjective states. It proves something narrower and more concrete: models can observe specific aspects of their internal processing and report on them with above-chance accuracy.
That's enough to be interesting. And as the capability improves, it might be enough to transform how we interact with and understand AI systems.
Read the full research paper for technical details and additional experiments.



