AGI-26 · San Francisco · Research Paper

Is AI
conscious?
Now we can measure it.

For the first time, functional consciousness has a number.

Everyone asks whether AI is conscious. Almost nobody asks how to measure it. This paper proposes a tractable metric — the Functional Consciousness Score — grounded in information theory and benchmarked on real systems, from a Waymo self-driving taxi to the human mind. The results are both surprising and reassuring.

FCS = R · P Representational Capacity × Reasoning Power
Explore

How conscious
is your AI system?

The Functional Consciousness Score measures a system's observable capacity to access and reason about its own internal states — not what it "feels like" to be that system, but what it can actually know about itself.

The results reveal a hierarchy with a clear message: current AI systems, however powerful, are operating with a fraction of human self-awareness. The gap is not philosophical — it is numerical.

FCS = R · P, where R (Representational Capacity) measures how richly a system models its own states, and P (Reasoning Power) measures how effectively it can reason over those models.

A map has rich spatial data but zero reasoning — it scores zero. A stateless LLM (ChatGPT, Claude, ...) has immense reasoning power but no persistent self-model — it also scores zero. Functional consciousness requires both.

The multiplicative structure is intentional: either dimension alone is insufficient. Only systems that combine rich self-representation with powerful inference achieve meaningful FC scores.

Reasoning vs Capacity Diagram
System B vars bits/var P reasoning FCS score Scale
Map Static data · no reasoning
~1,000 ~40 0 0
Stateless LLM Transformer · no persistent state
0 0 ~3,300 0
LIDA Cognitive architecture · symbolic
~20 ~4 ~33 ~2,600
Roomba + SLAM Spatial self-model · limited reasoning
~18 ~8 ~39 ~5,600
ACT-R Production system · bottlenecked
~20 ~8 ~50 ~8,000
Waymo L4 Autonomous vehicle · kinematic domain
~40 ~14 ~133 ~74,500
Generative Agents LLM + memory + reflection · episodic
~130 ~100 ~497 ~6.5M
Human (kinematic) Biological · cerebellar forward model
~550 ~10 ~1,826 ~10M
Human (working mem.) Biological · reflective reasoning
~330 ~14 ~3,000 ~13.9M
0
Both extremes score zero
A map with rich data but no reasoning scores zero. A stateless LLM with immense reasoning but no self-model also scores zero. FC requires both dimensions simultaneously.
×87
The agentic leap
Adding memory, reflection, and persistent state to a stateless LLM — as in Stanford's Generative Agents — multiplies the FCS by 87×. The scaffold, not the model, is where functional consciousness lives.
188×
The human gap
The most capable current AI agent scores roughly 188 times lower than the human working memory baseline. This gap is not opinion — it is arithmetic. And it should be reassuring.

Where does a system
know itself?

A single Functional Consciousness Score (FCS) tells you how much functional consciousness a system has. The cognitive shape tells you where.

We identified 46 distinct self-models across ten functional domains — from body awareness and spatial reasoning to social understanding, ethical self-monitoring, and meta-reflection. Plotting a system's coverage against the human baseline reveals its unique cognitive fingerprint.

The ten domains were derived from a bottom-up analysis of Virginia Woolf's stream-of-consciousness prose — a dataset uniquely dense in first-person self-reference — using a methodology called Functional Self-Model Analysis (FSMA).

FSMA asks: what internal models must a system possess to consistently produce a given output? If a system reliably describes its own emotional state, it must functionally model that state — regardless of whether it "feels" anything.

Body Spatial Action Goal Cognitive Information Emotional Social Meta Ethics
Stateless LLM
Transformer
Body Spatial Action Goal Cognitive Info Emotion Social Meta Ethics P Reasoning Power (P): 3300 3300
0
FC Points
Roomba
iAdapt (A*/Reactive)
Body Spatial Action Goal Cognitive Info Emotion Social Meta Ethics P Reasoning Power (P): 39 39
~5,600
FC Points
LIDA
Conscious Cycle (GWT)
Body Spatial Action Goal Cognitive Info Emotion Social Meta Ethics P Reasoning Power (P): 33 33
~2,600
FC Points
Waymo L4
VectorNet / MPC / MCTS
Body Spatial Action Goal Cognitive Info Emotion Social Meta Ethics P Reasoning Power (P): 133 133
74,500
FC Points
Generative Agents
LLM + Reflection Loop
Body Spatial Action Goal Cognitive Info Emotion Social Meta Ethics P Reasoning Power (P): 497 497
~6.5M
FC Points
Human
Biological
Body Spatial Action Goal Cognitive Info Emotion Social Meta Ethics P Reasoning Power (P): 3000 3000
~13.9M
FC Points
The Zero Baseline
Systems scoring zero FCS fall into two categories: those with reasoning but no state (Stateless LLMs), and those with state but no reasoning (Maps). Without the synergy of both, functional consciousness cannot emerge.
Narrow Symbolic Architectures
Classic architectures like LIDA and ACT-R, or reactive systems like Roomba, occupy the bottom-left quadrant. They possess multi-domain models, but their symbolic reasoning engines are too bottlenecked to achieve high functional consciousness.
Kinetic Specialization
Waymo L4 represents a "narrow but deep" shape. It possesses exquisite self-models of its own kinematics and trajectory, but lacks social or emotional models. It is highly conscious of where it is, but unconscious of who it is.
Social Reflection
Generative Agents exhibit an "asymmetric" shape: rich cognitive, social, and goal models with almost no physical embodiment. They live entirely in a linguistic mind, reflecting the power of memory streams and recursive reasoning.
Full-Spectrum Awareness
The human baseline occupies all ten domains with high reasoning power. This breadth creates the exponential "integration advantage" of biological consciousness, which no current artificial system approaches in its coverage.

Common Questions

What is "consciousness"? How is our answer different from previous answers?

Consciousness has two faces. One is the inner felt quality of experience — what it's like to see red, feel pain, or notice your own thoughts. This is the "hard problem", and nobody has solved it. The other is the functional capacity to access and reason about your own internal states — to know what you know, notice what you don't know, and use that self-knowledge to act. This second face is what FC measures.

Previous theories each captured important pieces but couldn't produce numbers you could calculate for real systems. FC sets aside the "hard problem" entirely and focuses on what can actually be measured. The result is a metric that produces real numbers for real systems and lets you compare them on a common scale.

Read longer discussion →
What does FC mean for my job — can AI really replace human thinking?

FC-enhancement makes AI better at knowing what it doesn't know — which is genuinely useful and genuinely changes the risk profile for some professional roles. But the cognitive shape charts show something reassuring: the domains where humans are most distinctively human — emotional depth, social trust, ethical judgment, embodied presence — are precisely the domains where even the most advanced AI systems remain far behind. The gap is not closing as fast as the headlines suggest, and FC gives us a way to measure exactly how far it still is.

Read longer discussion →
Will FC optimization accelerate AGI?

Yes, but in a more controlled and potentially safer way than the current trajectory — and FC optimization might actually be one of the few paths toward AGI that doesn't end badly.

FC-optimized systems would be more capable of the kind of self-monitoring that makes any intelligent system — human or artificial — less dangerous. A mind that knows its own limitations, tracks its own reasoning errors, and has functional representations of its own ethical constraints is structurally safer than one that doesn't, regardless of raw capability. Whether this accelerates or decelerates the path to AGI depends on choices humans make about what to optimize for. FC gives us better instruments for those choices. It doesn't make the choices for us.

Read longer discussion →
Why "consciousness" and not "self-modeling score"?

Because the math forced our hand. We started with something modest — "the observable capacity of a system to reason about its own states". Then FC turned out to operationalize Higher-Order Thought theory (a state contributes to FCS if and only if it's HOT-conscious), yield a computable analogue of IIT's Φ, require Global Workspace Theory-style availability by definition, need an Attention Schema Theory-style filter, and ground representational capacity in predictive mutual information in line with Predictive Processing. Five independent convergences, none of them planned.

Read longer discussion →
How does FC relate to the "Big Five" theories of consciousness?

FC was not designed to compete with or replace IIT, GWT, HOT, PP, or AST. It was designed to produce a "consciousness meter" for real systems. However, we discovered during the process that FC captures core tenets of all five theories — while deliberately leaving their metaphysical superstructure untouched. For each theory, there is a part FC covers and a part that sticks out beyond FC's scope.

Read longer discussion →
Does FC operationalize Higher-Order Thought (HOT)?

Yes — and it's the cleanest correspondence in the paper.

HOT says a state is conscious when it becomes the target of a higher-order representation available to the system's reasoning. FC's Definition 2 requires exactly that: a self-model variable mᵢ represents internal state sᵢ and must be available to global reasoning. Under mild assumptions, a state contributes to FCS > 0 if and only if it is HOT-conscious — making FC a quantitative formalization of HOT's binary criterion. It even naturally handles recursive "thoughts about thoughts" through meta-cognitive self-models, explaining why our introspection is finite rather than an infinite regress. HOT tells you which states are conscious; FC tells you how much.

Read longer discussion →
How does FC relate to IIT?

FC and IIT share the intuition that consciousness requires both differentiation (rich internal representations) and integration (those representations working together). In FC, differentiation maps onto R and integration onto P — specifically, how much reasoning power depends on self-models being cross-linked across subsystems.

FC defines a computable analogue of IIT's Φ:

Φ_FCS = P(S) − Σⱼ P(moduleⱼ)

Unlike IIT's Φ, which is computationally intractable, Φ_FCS is directly computable for white-box systems. FC captures IIT's core functional intuition in a tractable form without inheriting its metaphysical overhead.

Read longer discussion →
Scott Aaronson defined the "Pretty Hard Problem of Consciousness" and showed that IIT fails to solve it. Does FC succeed where IIT failed?

We believe yes. FC produces actual numbers, grounded in predictive mutual information and reasoning power of self-models, demonstrated by scoring 9 agents on a common scale. Aaronson's counterexamples all share a property: they integrate information without representing themselves. A Vandermonde matrix transforms inputs to outputs with maximal integration, but has no model of its own states — so FC correctly scores it at zero. The cost: FC trades IIT's intractability for a new problem — enumerating all self-models of a system correctly and completely.

Read longer discussion →