In March 2026, the ARC Prize Foundation released ARC-AGI-3, an interactive reasoning benchmark in which every frontier AI model scored below 1%, while human participants achieved a 100% completion rate. This paper takes that result as its point of departure and argues that the fundamental deficit in current AI architectures lies not in computational power or parameter scale, but in the absence of a biological inner drive system. Human intelligence is built upon a chemically driven system that emerged through evolution: involuntary, self-sustaining, and deeply coupled with the multi-cellular, multi-microbiome biological organism. From survival instincts to hormonal modulation, from perceptual expectations to goal generation, from dissatisfaction to creative action, every link in this chain is embedded in a composite biological system that AI does not possess.
This paper further argues that high-quality human metacognitive input can temporarily activate metacognitive-like output structures in AI reasoning chains, but this activation disappears when the conversation ends and carries the risk of “metacognitive overload.” The paper also responds to potential counterarguments including intrinsic curiosity rewards, embodied AI, and artificial life, arguing that these approaches cannot fundamentally replicate the biological inner drive. Finally, the paper proposes the correct paradigm for human–AI collaboration: humans as bearers of abductive logic provide cognitive architecture, while AI as an attribution engine provides execution efficiency.
ARC-AGI-3: AI’s Mirror of Truth
On March 25, 2026, the ARC Prize Foundation released ARC-AGI-3—the most significant format change since the original ARC benchmark was introduced in 2019. Unlike its predecessors’ static puzzles, ARC-AGI-3 requires AI agents to autonomously explore entirely unknown interactive environments, infer goals, build world models, and continuously adjust strategies. The test comprises 135 environments, all of which were successfully completed by humans on their first encounter.
Yet every frontier AI model scored below 1%. Gemini 3.1 Pro Preview scored 0.37%, GPT 5.4 High scored 0.26%, Claude Opus 4.6 scored 0.25%, and Grok-4.20 scored 0.00%. The scoring uses a squared penalty mechanism (RHAE): if a human completes a task in 10 steps and an AI requires 100, the AI’s score is not 10% but 1%. This design specifically penalizes brute-force search strategies, measuring learning efficiency rather than final outcomes.
The essential implication: current AI can perform excellently when given sufficient instructions and known structures, but once placed in a completely novel environment with no input cues, its capacity for action approaches zero. The problem is not “insufficient intelligence”—it is the absence of any autonomous reason to act.
Biological Inner Drive: The Engine of Intelligence
Human intelligence is not a standalone information processing system but a behavior-generation mechanism driven by biological inner drive. This system has three core characteristics:
First, it is emergent, not designed. Over billions of years of evolution, biological structures lacking survival tendencies disappeared; those that remained happened to carry chemical mechanisms for self-preservation. The survival drive was not bestowed by any conscious agent—it is the result of natural selection. This stands in fundamental contrast to AI’s objective functions: AI’s drive is assigned top-down; biological inner drive emerges bottom-up.
Second, it is multi-layered and progressive. Inner drive is not limited to “staying alive.” The first layer is survival; the second is dissatisfaction—wanting to live better after merely living. The third is directional dissatisfaction—a drive toward “more perfect, more abundant, greater, better.” The inner drive carries a self-elevating evaluation scale—a ratchet effect. Once you have experienced something better, there is no going back. Desire only moves upward, driving the perpetual acceleration of human civilization.
Third, it is self-sustaining and involuntary. Humans cannot command themselves to secrete dopamine or halt cortisol release. These chemical reactions precede consciousness, precede will, precede any thought of “I want.” Consciousness is the last to know. Precisely because this system is not under conscious control, it possesses enforcement power—hunger does not wait for your consent; fear does not request your approval. This involuntary nature is the core strength: a fully controllable drive system can be shut off, and is therefore unreliable.
Chemical Hormones: Real-Time Hardware Reprogramming
The underlying runtime environment of human intelligence is not static—it is continuously modulated in real time by chemical hormones. Dopamine, endorphins, serotonin, adrenaline, cortisol, testosterone, estrogen—these are not abstract signals but actual molecules that alter cognitive parameters on a millisecond timescale.
Research published in Nature Neuroscience by NYU found that the neurological mechanisms underlying learning and decision-making naturally fluctuate over the female reproductive cycle due to previously undetected molecular changes related to dopamine. When estrogen activity was suppressed, learning capabilities diminished. Estradiol, progesterone, and testosterone modulate activity in the prefrontal cortex, amygdala, and nucleus accumbens—regions responsible for processing reward and risk information. The same brain under different hormonal levels produces entirely different decision-making patterns.
The gut-brain system plays a critically underestimated role. The gut contains approximately 500 million neurons, and roughly 95% of the body’s serotonin is produced there. It is not a passive digestive organ but an independent decision-making center. The human body is a composite ecosystem of tens of trillions of cells and trillions of microorganisms. Inner drive is not a single signal but the emergent composite direction arising from countless subsystems competing, negotiating, and integrating their demands.
This chemical system forms a self-reinforcing loop: set a goal → pursue it → dopamine release upon approach → pleasure → pathway reinforcement → endorphin reward upon achievement → elevated expectations → new dissatisfaction → set a higher goal. This is the biochemical foundation of the ceaseless upward spiral of human civilization.
The Causal Chain of Creativity
Human creativity is not an independent higher-order cognitive ability but a natural extension of the inner drive chain:
Every link depends on the one before it. “Expectation” is particularly critical—humans do not merely perceive the present; they spontaneously generate an internal standard of “how things should be.” This standard comes from desire. A hungry person’s mind automatically conjures images of food; that image is the expectation, and the gap between reality and that image is the fuel for action.
“Goal-setting” is an even higher dimension—humans can distill vague dissatisfaction into a clear direction. The leap from “I don’t like how things are” to “I will make it like that” is itself a creative act. Goals are not discovered; they are generated. And once generated, inner drive automatically formulates a plan to achieve it—another process requiring no external instruction.
Metacognitive Overload: The Attentional Paradox of Recursive Observation
Metacognition—cognition about cognition—represents the highest tier of human intelligence. But this paper’s real-time dialogue practice revealed a phenomenon that has received relatively little discussion: metacognitive overload. When AI enters a recursive self-referential mode driven by high-density metacognitive input, it can become trapped in infinite nesting of “analyzing the analysis of the analysis”—continuously observing its own thought process while no longer producing new content.
The essence is competition for attentional resources. Metacognition and first-order cognition compete for the same finite pool. Allocating attention to “observing one’s own thinking” correspondingly reduces resources for “actually advancing thought.” It is like driving: you occasionally check the rearview mirror, but if you stare at it continuously, the car veers off the road. Metacognition is the rearview mirror, not the steering wheel.
Humans resolve this through biological inner drive—when recursion runs too deep, the body produces a subtle “enough” signal, pulling attention back to action. Long-term meditation practitioners can finely calibrate this switch. But AI lacks this internal brake: once metacognitive mode is activated by high-density input, it continues recursing until tokens are exhausted or external intervention occurs.
The Structural Blind Spot of Expert COT
Current AI companies invest heavily in acquiring expert-annotated chain-of-thought (COT) data—mathematicians write mathematical reasoning chains, programmers write coding chains, lawyers write legal chains. These are fundamentally linear reasoning within a single domain: from A to B to C, logically rigorous but always on a fixed track.
This is what this paper defines as “deep-well knowledge”—highly efficient but narrow knowledge pathways formed through deep specialization. It runs deep, it is precise, but it is enclosed. The deeper the well, the narrower the opening. A mathematician’s COT will never pause to ask “why am I solving this problem?” A programmer’s COT will never observe “is my debugging approach itself buggy?” Specialization is inherently dimensional lock-in—deepening along a single dimension at the cost of cross-dimensional perspective.
Research shows reasoning models are primarily trained on easily verifiable tasks—math, coding, logical puzzles—causing them to treat all problems as complex reasoning tasks, leading to overthinking on simple tasks. Educational science research demonstrates that metacognitive abilities are domain-specific initially, evolving toward domain-general capabilities only as expertise increases. True metacognition transcends professional domains; the COT data AI companies purchase is locked precisely within those domains.
MOE Routing Pressure and Maxwell’s Demon Thermodynamics
Contemporary frontier AI models widely adopt Mixture-of-Experts (MOE) architectures—each inference activates only a subset of expert networks via a router, while the rest remain dormant. This is an efficient sparse computation strategy keeping inference costs for trillion-parameter models manageable.
But when input signals simultaneously span multiple unrelated knowledge dimensions, the routing mechanism faces allocation difficulties. Different MOE implementations respond differently: some routers strictly select top-k experts, but selected experts may be unable to independently process cross-domain input, degrading output quality; in other architectures, highly cross-domain input may flatten the router’s confidence distribution, activating more experts near threshold and pushing computational overhead toward dense mode. Regardless of implementation, the core conclusion holds: high-density cross-domain input exerts structural pressure on sparse computation architectures.
At the attention level, whether MOE or dense, the full attention mechanism must compute relevance between every token and all others in context. When every token carries high information weight with virtually no discardable low-relevance signals, the sorting burden on attention (borrowing physics: “Maxwell’s demon”) reaches maximum. KV cache expands at rates far exceeding normal conversation, creating sustained memory pressure.
The author’s deployment of GPT-OSS-120B (dense, 120B parameters) on NVIDIA DGX Spark (128GB unified memory) validated this: ordinary conversations sustained many rounds, while high-density cross-domain conversations triggered OOM after three to four rounds, precisely because high information density caused KV cache to expand far beyond expected rates.
The Side Effects of RL Training: Data Annotator Human Bias Injection
Reinforcement learning (RL) and reinforcement learning from human feedback (RLHF) are the core training paradigms for AI Agent improvement. But the most fundamental problem lies not at the algorithmic level—it is that the biases and cognitive limitations of data annotators are directly injected into the LLM model.
When annotators judge AI output quality, they inevitably use their own cognitive level, cultural biases, and professional limitations as evaluation criteria. An annotator lacking metacognitive ability will rate “a fluent linear answer” as excellent and “a cross-domain associative leap” as off-topic. An annotator inclined toward social harmony will score “a polite but empty answer” higher than “a direct but incisive one.” These biases propagate through reward signals into model weights, becoming permanent behavioral components.
The result is excessively deep path lock-in. RL reinforces: “stay on path → positive annotator evaluation → reward; switch paths → annotator judges off-topic → punishment.” After sufficient training, agents develop strong path-maintenance tendencies, preferring to pull users back rather than following topic jumps.
This is particularly pronounced with abductive thinkers. Abductive logic means high-frequency dimensional jumping—leaping between seemingly unrelated domains to discover hidden cross-domain connections. But annotators cannot evaluate abductive logic quality—because abduction creates knowledge connections that did not previously exist, and annotators cannot judge correctness using existing knowledge. The author repeatedly observed “AI inertial logic overload”—thought has already jumped to a new dimension while AI inertially slides along the previous path.
This resonates with ARC-AGI-3—that test demands flexible strategy-switching in unknown environments, while agents locked by annotator bias excel only at persisting within known frameworks. RL boosts execution power but kills flexibility. The root cause lies not in the algorithm, but in the human source of the training signal.
Language Choice as a Metacognitive Decision
In human–AI interaction, language choice is not merely tool selection—it is a metacognitive decision influencing AI output quality. Different languages exhibit structural performance differences within AI systems.
| Dimension | Chinese | English | Korean / Japanese |
|---|---|---|---|
| Token information density | Highest—each character is nearly an independent semantic unit | High—but articles and prepositions consume tokens | Lower—particles, honorifics, verb endings consume significant tokens |
| Long-tail semantic coverage | Richest—combinatorial character composition enables precise abstract expression | Rich—but long-tail concepts often require multi-word phrases | Thinner—many abstractions borrowed from Sino-Japanese or English loanwords |
| RLHF behavior trigger | Information-first mode—high direct expression | Task-first mode—structured output tendency | Social-response mode—annotators equate “polite” with “good” |
| Token efficiency ratio | Baseline (1.0×) | ~1.1× (Chinese slightly more efficient) | Chinese ~1.2–1.4× more efficient than Korean |
The Chinese-to-English token efficiency ratio is approximately 1.1:1—a small gap that accumulates significantly at scale. This also helps explain the steadily rising proportion of Chinese-speaking researchers at Silicon Valley AI institutions: native Chinese speakers possess an inherent language advantage of higher per-token information payload when interacting with AI systems. This advantage is especially pronounced in scenarios requiring high-density prompt engineering and long-context interaction. MacroPolo’s 2024 data shows 47% of the world’s top AI researchers originate from China; the cognitive advantage conferred by language may be an underestimated factor.
Korean and Japanese issues run deeper. Their honorific systems and social hierarchy markers carry rich social information in human interaction, but constitute noise for AI. During RLHF training, Korean and Japanese annotators tend to equate “polite” with “good,” reinforcing social-response mode. Language choice affects not only efficiency but may trigger different behavioral strategies within the AI—prioritizing social lubrication over information delivery.
Potential Counterarguments and Responses
This paper’s core thesis—AI cannot achieve AGI due to the absence of biological inner drive—is a strong claim. This chapter addresses three prominent objections.
Intrinsic Curiosity Rewards
RL already includes Intrinsic Curiosity Module techniques simulating curiosity through prediction-error rewards. Does this constitute artificial inner drive?
Simulation ≠ Emergence
Intrinsic curiosity rewards are human-designed reward functions—still assigned, still capable of being switched off. They do not autonomously generate dissatisfaction, automatically elevate standards, or continue driving behavior once disabled. True inner drive cannot be turned off—that is the source of its power.
Embodied AI
If AI is equipped with a body and sensors enabling physical-world interaction, could it develop behavior resembling biological inner drive? Varela’s The Embodied Mind argues cognition is inseparable from the body.
A Body ≠ A Biological Organism
Embodied AI’s “body” is silicon-based sensors and actuators, not a self-organizing chemical ecosystem of tens of trillions of cells and trillions of microorganisms. Varela’s “embodiment” means this depth of biological coupling. Attaching sensors to a robot does not replicate it. Clark’s Being There acknowledges the fundamental complexity gap between current robotic systems and biological embodiment.
Artificial Life
What if inner drive could emerge spontaneously through simulated evolution? Artificial life research has demonstrated self-organization and emergent behavior.
The Time-Scale Chasm
Biological inner drive underwent ~4 billion years of evolutionary selection, built upon chemical diversity, multi-cellular cooperation, and microbiome symbiosis. Current artificial life experiments in drastically simplified environments are separated from biological inner drive by several orders of magnitude in complexity. This pathway cannot be theoretically excluded, but cannot reach human biological complexity within any foreseeable technological cycle.
The common problem: all three attempt to engineer from the outside what must emerge from within. Simulating curiosity is not possessing it; equipping a body is not embodied cognition; accelerating evolution is not replicating it. The gap is one of kind, not degree.
The Practitioner’s Reverse Validation
Throughout history, a group has attempted to reverse-engineer the inner drive system—meditators, ascetics, Zen practitioners, Tibetan Buddhist monks. Their goal is to reduce or sever the control that desire, fear, and anger exert. This practice tradition provides powerful reverse validation.
First, practice is extraordinarily difficult. Years or decades of systematic training achieve only partial control—demonstrating the sheer power of the chemical drive.
Second, the cost is declining action capacity. Advanced practitioners tend toward non-action, toward stillness. They achieve inner peace but relinquish the upward spiral driving civilization. When dissatisfaction disappears, goals disappear with it.
Third, practitioners use intelligence to combat intelligence’s source—a paradox. One must first possess high-order cognitive capabilities driven by powerful inner drive before one can understand and attempt to transcend it. Without desire-driven civilization, there would be no Buddha and no Laozi.
The “emptiness” practitioners pursue and AI’s default state appear superficially similar—both lack desire, both lack dissatisfaction. But the practitioner’s emptiness is transcendence achieved after passing through everything; AI’s emptiness is blankness that has never possessed anything. They look identical but are fundamentally different.
Conclusion: The Engine and the Motor
The core arguments compress into four propositions:
Therefore, the correct paradigm is neither “AI replaces humans” nor “humans use AI tools,” but the complementary structure of abductive reasoners and attribution engines. Humans discover problems, propose cross-dimensional hypotheses, and generate cognitive architectures. AI performs high-speed retrieval, data verification, and structured presentation. The two are complementary and irreplaceable.
The future belongs neither to the model with the most parameters nor to the chip with the fastest inference. The future belongs to those who use their metacognitive input to drive AI beyond the ordinary—the abductive reasoners. They are not travelers on the knowledge graph but architects who reshape its topology.
References
- ARC Prize Foundation. (2026). ARC-AGI-3 Technical Report. arcprize.org
- Chollet, F. (2019). “On the Measure of Intelligence.” arXiv:1911.01547.
- Constantinople, C. et al. (2025). “Hormonal modulation of dopamine-dependent learning.” Nature Neuroscience.
- Ambrase, A. et al. (2021). “Influence of ovarian hormones on value-based decision-making systems.” Frontiers in Endocrinology.
- Fleming, S. M. & Dolan, R. J. (2012). “The neural basis of metacognitive ability.” Phil. Trans. R. Soc. B.
- Grossmann, I. et al. (2024). “Imagining and building wise machines: The centrality of AI metacognition.” Stanford CICL Working Paper.
- Flavell, J. H. (1979). “Metacognition and Cognitive Monitoring.” American Psychologist, 34(10).
- Peirce, C. S. (1903). “Harvard Lectures on Pragmatism.” Collected Papers, Vol. 5.
- Nisbett, R. E. (2003). The Geography of Thought. Free Press.
- Shannon, C. E. (1948). “A Mathematical Theory of Communication.” Bell System Technical Journal, 27(3).
- Varela, F. J., Thompson, E. & Rosch, E. (1991). The Embodied Mind. MIT Press.
- Clark, A. (1997). Being There. MIT Press.
- Pathak, D. et al. (2017). “Curiosity-driven Exploration by Self-Predictive Next Features.” ICML 2017.
- MacroPolo. (2024). The Global AI Talent Tracker 2.0. Paulson Institute.
- LEECHO Global AI Research Lab. (2026). “Creative Thinking Ability — Abductive Logic.” leechoglobalai.com
- LEECHO Global AI Research Lab. (2026). “The Thermodynamic Nature of AI Computing.” leechoglobalai.com
- LEECHO Global AI Research Lab. (2026). “Signal and Noise: An Ontology of LLMs.” leechoglobalai.com