Can evolution produce abstraction? A research hub.
Two research frontiers are converging: self-replicating programs emerge from chaos (BFF), and measuring intelligence requires abstraction (ARC). Can evolution produce abstraction? This hub explores that question.
In 2024, Google researchers showed that random programs placed in a "primordial soup" spontaneously evolve self-replicators about 40% of the time. No fitness function. No design. Just interaction and time.
Meanwhile, FranΓ§ois Chollet's ARC benchmark reveals that current AIβdespite trillions of parametersβfails at the kind of abstract reasoning a child performs effortlessly. The benchmark tests fluid intelligenceFluid IntelligenceThe ability to solve novel problems without prior training. Contrast with crystallized intelligence (accumulated knowledge).: can you learn new abstractions from just a few examples?
This creates a fascinating gap:
Self-copying programs. Autocatalysis. Replication without design. But systems plateauβcomplexity stops growing. No abstraction, no learning, just copying.
Transfer learning. Novel problem-solving. Core knowledge priors. Abstraction from examples. Current deep learning fails; program synthesis helps but isn't enough.
Replication is necessary but not sufficient for intelligence. The gap between self-copying programs and abstraction-capable systems is the central mystery. Can evolutionary dynamics bridge it?
BFF runs showing emergence
Pure LLM on ARC-AGI-2
TRM params beating 671B
The ARC Prize 2025 winner was a 7-million parameter recursive model that outperformed models with 100,000Γ more parameters. Size isn't the answer. Architecture matters. Perhaps how you learn matters more than how much you've seen.
EMERGENCE INTELLIGENCE
β β
βΌ βΌ
βββββββββββ ββββββββββββ ββββββββββββββ βββββββββββ
β Random βββββΆβ Self- βββββΆβ Complexity βββββΆβ Abstrac-β
β Programsβ β Replica- β β Growth β β tion β
β (BFF) β β tors β β (gap here) β β (ARC) β
βββββββββββ ββββββββββββ ββββββββββββββ βββββββββββ
β β β β
chaos autocatalysis plateau? transfer
learning
The dashed arrow is where research is needed. What happens between replication and reasoning? Three hypotheses:
Random BFF programs in a primordial soup spontaneously evolve self-replicators. The foundation of emergence research.
Intelligence is skill-acquisition efficiency. Introduces ARC benchmark. The foundation of abstraction research.
Test-time training, refinement loops, and program synthesis emerge as key techniques. The current frontier.
Understanding how intelligence works requires understanding where it fails. Esoteric programming languages expose the limits of current AI through minimal, Turing-complete systems that challenge different cognitive capabilities.
MFF establishes computational life with eight commands. Whitespace makes syntax invisible. Unlambda removes variables entirely. Befunge makes execution spatial. Each language isolates a different dimension of computationβand reveals where statistical learning cannot substitute for genuine understanding.
This research question has been asked for 30+ years. Most ALife systems plateau. Replication β learning. The burden of novelty is high. See the honest critique.