AI Theory: Fragmented but Expansive

TL;DR — Don’t wait for a “theory of everything.” AI will mature the way physics, biology, and engineering did: via a network of local theories that are right at their scale. That’s good news for progress—and for researchers who can bridge ideas across domains.


1) A Personal Prediction

I don’t expect a single grand unifying theory of AI any time soon. The field looks set to evolve like post-Newtonian physics or modern biology: patchworks of locally powerful laws, models, and intuitions that each explain part of intelligence under specific assumptions and scales.


2) Why Unification Is Unlikely (for Now)

Three structural forces keep AI theory plural:

  1. System complexity and heterogeneity
    LLMs, diffusion models, tool-using agents, and multi-modal stacks behave differently because they are different systems with different bottlenecks.

  2. Scale vs. abstraction mismatch
    Techniques that characterize small models (e.g., VC bounds) often lose bite at trillion-parameter scales; statistical-mechanics-style arguments capture trends but miss mechanism-level precision.

  3. Incompatible objectives and constraints
    Efficiency, alignment, interpretability, reasoning, privacy, and safety often pull in orthogonal directions, requiring different abstractions and evaluation regimes.

Outcome: a pluralistic ecosystem of theories, each valid in its operating region.


3) Analogies That Actually Map


4) What “Fragmented Progress” Looks Like

Each speaks a completely different language; together they form a growing atlas of intelligence.


5) Counterpoints—and What Would Falsify This

Could we still arrive at a unifying theory of AI? Possibly—but I suspect it would be as difficult as reconciling the four fundamental forces in physics or proving $\mathsf{P} \neq \mathsf{NP}$:

Absent breakthroughs of this magnitude, local theories will continue to dominate—just as in physics, where unification remains elusive, or in complexity theory, where $\mathsf{P} \stackrel{?}{=} \mathsf{NP}$ persists as one of the deepest open problems: beautiful, profound, but far beyond immediate reach.


6) Why Local, Messy Formulas Still Matter

And even if a grand unifying theory of AI were discovered, local, messy formulas would remain indispensable. Civil engineering makes this obvious. Earthquake design, for instance, often relies not on elegant PDEs but on blunt empirical rules. A typical “base shear–type” formula looks something like:

\[V = C_s \cdot W, \quad C_s = \frac{0.44\, S}{R/I + 0.5\,(T/6.0)^{0.8}}\]

Here $S, R, I, T$ are just code-defined factors—seismic intensity, response modification, importance level, and structural period—stitched together with constants and exponents. Plugging in some typical numbers, you might get a coefficient around $0.06$. Multiply by the building’s weight $W$ and—voilà—the design shear force.

To a physicist, this looks like an arbitrary patchwork of constants and powers. Yet to an engineer, it is nothing more than \(+, -, \times, \div, \sqrt{}\)—simple operators wrapped in messy-looking fractions—that have been validated through decades of practice. Sure, refinements exist, and in aerospace or nuclear engineering they matter. But for ordinary civil structures, these empirical rules are more than enough: practical, robust, and trustworthy.

Likewise in AI, unification—if it ever comes—will not displace practical “local laws.” Scaling curves, regret bounds, or optimization heuristics endure not because they are elegant, but because they are usable at the right scale. They may look crude compared to a hypothetical “grand theory,” yet they remain the workhorses precisely because they solve the problems we actually face.


7) Why Fragmentation Is Good (Now)

Think “federated science”: many lenses, frequent cross-checks, rapid iteration.


8) Where This Leaves AGI

Advances toward AGI will likely accelerate this pluralism before they reduce it. As capabilities expand, new regimes (long-horizon planning, tool economies, social learning) will demand new local theories.

History gives a guide:


9) Closing Thought

AI’s scientific richness won’t come from compressing everything into one equation. It will come from cultivating a network of partial, overlapping theories—each precise at its scale, each falsifiable in its domain, and each useful for building and understanding intelligent systems.

Fragmented, but expansive—and that’s exactly how science often wins.


Yufa Zhou — August 18, 2025