← Hub

The Bridge

From Replication to Reasoning

TL;DR

Replication is necessary but not sufficient for intelligence. The gap between self-copying programs and abstraction-capable systems is the central mystery. Three approaches might bridge it: environmental coupling, meta-learning, and predictive coding.

Sections
The Gap · Hypotheses · Evo-Learning Theory · Approaches · Evidence · Critique · Research Program

The Gap

BFF shows that self-replicators emerge spontaneously from random programs. ARC shows that intelligence requires abstraction from few examples. What's between them?

BFF: What Emerges
  • Self-copying programs
  • Autocatalytic networks
  • Phase transitions
  • Entropy dynamics
ARC: What's Required
  • Transfer learning
  • Novel problem-solving
  • Core knowledge priors
  • Abstraction from examples
BFF: What's Missing
  • No within-lifetime learning
  • Systems plateau
  • No transfer to novel tasks
  • No world models
ARC: What's Needed
  • Test-time adaptation
  • Program synthesis
  • Refinement loops
  • Compositional generalization
The Core Question

Can evolutionary dynamics produce systems that learn within their lifetime, not just across generations? Can replicators develop internal models that transfer to novel situations?

Three Bridging Hypotheses

H1: Environmental Pressure

Abstraction emerges when environments demand it. If survival requires predicting patterns, organisms that develop internal models outcompete those that don't.

The Argument

In a static environment, hardcoded responses win. But in environments with:

...organisms that model their environment gain advantage over those that merely react.

Evidence for: Coevolution in Avida dramatically increases complexity. Host-parasite arms races force organisms to develop more sophisticated behaviors.

Evidence against: Increased complexity ≠ abstraction. Avida organisms develop better heuristics, not transferable reasoning.

H2: Meta-Learning is Algorithmic Replication

Learning is replication at the algorithm level. Just as genes replicate through organisms, learning algorithms replicate through problem-solving episodes.

The Argument

Consider what a "learning algorithm" does:

This is autocatalysis at the algorithmic level. The algorithm catalyzes its own improvement. If evolution can produce molecular autocatalysis, perhaps it can produce algorithmic autocatalysis.

Evidence for: SOAR (Self-improving Language Models for Evolutionary Program Synthesis) achieves 52% on ARC by combining evolutionary search with self-improvement.

Evidence against: SOAR requires explicit meta-learning objectives. The learning loop is engineered, not emergent.

H3: Free Energy Minimization

Friston's Free Energy Principle proposes that all adaptive systems minimize variational free energy—a measure of surprise. Both perception and action serve this goal.

The Equation
\[ F = D_{KL}\left[ q(s) \,\|\, p(s|o) \right] - \log p(o) \]

Agents minimize F by either:

If replicators are implicitly minimizing free energy (staying alive = minimizing surprise), perhaps explicit learning is just efficient free energy minimization.

Evidence for: In vitro neural networks experimentally validated the free energy principle in 2023.

Evidence against: Free energy minimization is a framework, not a mechanism. It describes what systems do, not how to build them.

Evolutionary Learning Theory

Deep Research Available

A comprehensive literature review on evolutionary learning theory is available: evolutionary-learning.md (15+ sources with full citations).

The question "Can evolution produce learning?" has a rich theoretical foundation. Key findings:

Evolution IS Learning (Formally)

Livnat & Papadimitriou (2014) proved that sexual evolution under weak selection is mathematically equivalent to the Multiplicative Weights Update Algorithm—a powerful online learning algorithm. Evolution doesn't merely resemble learning; the equations are identical.

The Baldwin Effect

Hinton & Nowlan (1987) demonstrated computationally that learning "smooths" the fitness landscape. Organisms that learn can find solutions that pure evolution cannot reach, then genetic assimilation can make those learned behaviors innate. Learning guides evolution.

Phenotype Leads Genotype

West-Eberhard (2003) argues that "genes are followers, not leaders." Phenotypic plasticity—organisms adapting to their environment—generates the variation that evolution acts upon. What organisms learn to do, their descendants may be born knowing.

Evolution of Learning Algorithms

AutoML-Zero (2020) evolved complete machine learning algorithms from basic operations, discovering backpropagation and dropout without human guidance. Clune's AI-GAs framework (2019) proposes this as the path to general AI: evolve the architectures, the learning algorithms, and the learning environments.

What This Means for the Bridge

The literature suggests three key insights:

  1. Evolution and learning are not separate — they're different expressions of the same adaptive process operating on different timescales.
  2. The transition exists — genetic assimilation provides a mechanism for learned behaviors to become innate, bridging within-lifetime and across-generation adaptation.
  3. Engineering confirms biology — artificial evolution can produce learning algorithms (NEAT, AutoML-Zero), suggesting the biological transition is not mysterious but mechanistic.

The remaining question is whether this transition can be observed spontaneously in minimal systems, without explicitly engineering meta-learning objectives.

Research Approaches

Computational: ARC Tasks in BFF Soups

Embed simple ARC-like tasks into BFF simulations. Reward programs that solve transformation puzzles. Measure whether evolved programs generalize to unseen variants.

Research Speculative

Theoretical: Abstraction as Compression

Model abstraction as lossy compression of environmental regularities. Define metrics connecting Kolmogorov complexity, logical depth, and ARC performance.

Research

Empirical: Developmental Studies

Study how human children develop abstraction. What's the minimal environmental structure? What core knowledge priors are truly innate vs learned?

Pilot Existing Literature

Hybrid: Evolutionary Program Synthesis

Combine genetic algorithms with program synthesis. Evolve populations of programs that solve ARC tasks. Use novelty search to escape local optima.

Active Research

What Would Confirm/Refute

Confirmation Criteria
Refutation Criteria

Honest Critique

This Is Well-Trodden Ground

This research question has been asked for 30+ years. Tierra (1991), Avida (1998), NEAT (2002), POET (2019)—all explored whether evolution produces learning. None demonstrated spontaneous emergence of within-lifetime abstraction.

Steel-Man Critiques

"This is just genetic algorithms with extra steps"

The "soup" framing is aesthetic, not functional. Population → variation → selection → repeat. Whether you call it soup, ecosystem, or GA, the dynamics are identical. No primordial soup has produced qualitatively different behavior than standard evolutionary algorithms.

"Learning requires credit assignment"

Backprop, TD-learning, evolutionary strategies—all need fine-grained credit assignment. Evolution provides only coarse-grained credit (offspring success). No theoretical framework predicts spontaneous fine-grained credit assignment emergence.

"Universality ≠ capability"

A Turing-complete system can compute learning algorithms, but won't spontaneously produce them. The Kolmogorov complexity of gradient descent is hundreds of bits. Probability of random emergence: negligible.

What Would Actually Be New

For this research to contribute novelty, it must demonstrate one of:

  1. Within-lifetime adaptation clearly distinguished from evolution
  2. Meta-learning emergence without meta-learning objectives
  3. Unbounded complexity growth over evolutionary time
  4. Predictive models emerging from soup dynamics, verifiable through interpretability

Tractable Research Program

The Program

Phase 1: Accelerate BFF complexity through environmental structure and coevolution.
Phase 2: Introduce external resources; measure within-lifetime adaptation.
Phase 3: Embed ARC-like tasks; test for transfer learning.

Phase 1: Accelerate Complexity

Spatial Heterogeneity

Create resource gradients in the BFF soup. Different regions provide different "nutrients" (tape values that aid replication). Measure: do organisms specialize? Do niches form?

Host-Parasite Coevolution

Introduce parasites that exploit replicators. Measure: does an arms race develop? Does complexity increase sustainably?

Island Models

Spatially isolated populations with periodic migration. Maintains diversity, enables parallel exploration.

Phase 2: Environmental Coupling

External Resources

Map tape regions to environmental "food." Programs that write to certain addresses gain replication advantage. Environment varies in learnable patterns.

Metric: Within-Lifetime Change

Does the same organism improve performance during execution? Track: instructions executed before successful replication. Does this decrease over organism's "lifetime"?

Phase 3: Abstraction Tasks

Embedded ARC-like Tasks

Programs must solve simple transformation puzzles to access resources. Puzzles vary but share underlying rule. Measure: do programs generalize to unseen variants?

Novelty Metric

\[ \rho(x) = \frac{1}{k} \sum_{i=1}^{k} \text{dist}(x, \mu_i) \]

Use novelty search instead of fitness. Reward behavioral diversity. Escape local optima.

Next Steps

🔬 Run Experiments Concrete code and metrics 💻 BFF Implementation TypeScript source code 📚 Reading List Primary sources 🧩 ARC Deep Dive Abstraction benchmark