[Paper] The brain-AI convergence: Predictive and generative world models for general-purpose computation

Published: (December 2, 2025 at 12:03 AM EST)
4 min read
Source: arXiv

Source: arXiv - 2512.02419v1

Overview

A new perspective paper draws a direct line between the way modern transformer‑based AI systems learn and the fundamental computations performed by the brain’s neocortex and cerebellum. By framing both biological and artificial intelligence around predictive, generative world models, the authors argue that a shared “predict‑and‑learn‑from‑error” loop underlies everything from visual perception to motor control—and that this loop is the key to building truly general‑purpose, adaptable AI.

Key Contributions

  • Cross‑domain computational analogy – Shows that attention‑driven neocortical circuits and non‑attentional cerebellar circuits converge on the same predictive‑error learning principle.
  • World‑model framing – Recasts diverse brain functions (sensory understanding, motor generation) as different uses of a single internal model of the world.
  • Parallel with modern AI – Demonstrates that large‑scale transformer training (self‑supervised next‑token prediction) mirrors the brain’s prediction‑error learning, suggesting a convergent solution to general‑purpose computation.
  • Unified theory of intelligence – Proposes that a uniform circuit architecture can support a wide range of high‑level tasks when coupled with predictive world modeling.
  • Roadmap for interdisciplinary research – Highlights concrete points where neuroscience insights can inform AI architecture design and vice‑versa.

Methodology

The authors adopt a theoretical‑comparative approach rather than new experiments. Their workflow consists of:

  1. Literature synthesis – Pull together recent findings on neocortical attention mechanisms, cerebellar predictive coding, and transformer training dynamics.
  2. Conceptual mapping – Identify analogous computational primitives (e.g., attention as a dynamic weighting of past context, cerebellar forward models as error‑correcting predictors).
  3. World‑model abstraction – Formalize a “predictive generative model” that takes past sensory/motor streams, predicts future states, and updates itself via prediction error.
  4. Cross‑validation – Show how this abstraction explains known phenomena in both biology (e.g., sensory inference, motor adaptation) and AI (e.g., language modeling, image generation).

The paper stays high‑level, using diagrams and equations only where they clarify the shared algorithmic loop.

Results & Findings

  • Predictive coding is universal – Both neocortex and cerebellum continuously generate predictions and adjust internal weights when reality deviates, matching the loss‑gradient updates in transformers.
  • Attention ≠ the only route – Even non‑attentional cerebellar circuits achieve similar outcomes by weighting temporal context through recurrent dynamics, suggesting attention is a implementation rather than a necessity.
  • World models enable repurposing – A single predictive model can be read out for perception (interpretation) or for generation (action planning), mirroring how a language model can be used for text completion or for code synthesis.
  • Uniform circuitry, diverse behavior – The brain’s relatively homogeneous microcircuitry can support a spectrum of cognitive abilities when driven by a shared predictive loop, offering a plausible explanation for the emergence of high‑level intelligence.

Practical Implications

DomainTakeaway for Developers / Engineers
Model ArchitectureConsider prediction‑error loops as a design principle: combine a forward predictor with a fast error‑feedback pathway (e.g., residual connections, auxiliary loss heads).
Multi‑Task LearningA single transformer trained on next‑token prediction can be re‑used for generation, classification, or control tasks without task‑specific heads—mirroring the brain’s reuse of world models.
Robotics & ControlImplement cerebellum‑inspired forward models that predict motor outcomes and correct actions on the fly; this can improve sample efficiency and safety in real‑time control.
Continual / Adaptive AILeverage prediction‑error signals to trigger rapid online updates, enabling systems that adapt to distribution shift as naturally as the brain does.
ExplainabilityThe shared predictive framework offers a common language for interpreting model failures (high prediction error) across vision, language, and reinforcement‑learning domains.
Hardware DesignUniform micro‑architectures (e.g., systolic arrays) that support both attention and recurrent predictive updates could be more power‑efficient, echoing the brain’s hardware simplicity.

In short, the paper suggests that building AI around a single, continuously updated world model—instead of stacking many specialized modules—could yield more flexible, human‑like intelligence.

Limitations & Future Work

  • Empirical validation needed – The arguments are largely conceptual; concrete experiments that directly map transformer internals to cerebellar/neocortical activity are still missing.
  • Scope of biological fidelity – Real neurons exhibit dynamics (spiking, neuromodulation) that aren’t captured by current transformer math; bridging that gap remains a challenge.
  • Scalability of cerebellar‑style predictors – It’s unclear how well non‑attention recurrent predictors scale to the massive context windows used in today’s LLMs.
  • Future directions – The authors call for joint neuro‑AI projects that (1) embed explicit prediction‑error pathways in deep nets, (2) test cerebellar‑inspired forward models in robotics, and (3) develop neuroscientifically grounded benchmarks for world‑model competence.

Bottom line: By spotlighting the brain’s predictive, generative world model as the common denominator of intelligence, this paper gives developers a fresh lens to rethink AI architecture—favoring unified, error‑driven learning loops over a patchwork of task‑specific components. The next wave of “brain‑inspired” systems may look less like a collection of attention heads and more like a single, continuously self‑correcting model of the world.

Authors

  • Shogo Ohmae
  • Keiko Ohmae

Paper Information

  • arXiv ID: 2512.02419v1
  • Categories: q-bio.NC, cs.AI, cs.CL, cs.NE
  • Published: December 2, 2025
  • PDF: Download PDF
Back to Blog

Related posts

Read more »