SYDE 537: Artificial Life: Biology and Computation

Chrystopher L. Nehaniv

Estimated study time: 1 hr 20 min

Table of contents

Sources and References

Primary sources — Langton, C.G., “Artificial Life” in Artificial Life (1989); Langton et al. (eds.), Artificial Life II, Santa Fe Institute Studies in the Sciences of Complexity (1991); Adami, C., Introduction to Artificial Life (1998), Springer; Holland, J.H., Adaptation in Natural and Artificial Systems (1975/1992), MIT Press; Koza, J.R., Genetic Programming (1992), MIT Press; Lindenmayer, A. & Prusinkiewicz, P., The Algorithmic Beauty of Plants (1990), Springer; Turing, A.M., “The Chemical Basis of Morphogenesis,” Philosophical Transactions of the Royal Society B 237 (1952); von Neumann, J. & Burks, A.W., Theory of Self-Reproducing Automata (1966), University of Illinois Press; Wolfram, S., A New Kind of Science (2002), Wolfram Media; Kauffman, S.A., The Origins of Order (1993), Oxford University Press; Maynard Smith, J. & Szathmáry, E., The Major Transitions in Evolution (1995), W.H. Freeman; Dorigo, M. & Stützle, T., Ant Colony Optimization (2004), MIT Press; Kennedy, J. & Eberhart, R., Swarm Intelligence (2001), Morgan Kaufmann; Mitchell, M., An Introduction to Genetic Algorithms (1996), MIT Press; Mitchell, M., Complexity: A Guided Tour (2009), Oxford University Press; Bonabeau, E., Dorigo, M. & Theraulaz, G., Swarm Intelligence: From Natural to Artificial Systems (1999), Oxford University Press; Eigen, M. & Schuster, P., The Hypercycle (1979), Springer; Ray, T.S., “An Approach to the Synthesis of Life,” in Artificial Life II (1991); Wiener, N., Cybernetics (1948), MIT Press; McCulloch, W.S. & Pitts, W., “A Logical Calculus of the Ideas Immanent in Nervous Activity,” Bulletin of Mathematical Biophysics 5 (1943); Maturana, H.R. & Varela, F.J., Autopoiesis and Cognition (1980), Reidel.

Online resources — Santa Fe Institute working papers and lecture notes (santafe.edu); ALife Conference proceedings (alife.org); MIT Press Artificial Life journal archives; NECSI (necsi.edu) complexity resources; NASA Astrobiology Institute resources; Avida platform documentation (avida.devosoft.org); Conway’s Game of Life community resources (conwaylife.com).


Chapter 1: What Is Artificial Life?

1.1 Origins and Intellectual Context

Artificial Life (ALife) as a formal research programme was inaugurated by Christopher Langton at the first workshop on Artificial Life, held at Los Alamos National Laboratory in 1987. The proceedings, published as Artificial Life (1989) and the subsequent Artificial Life II (1991) through the Santa Fe Institute, established the intellectual scope of the field. Langton’s foundational definition positioned ALife as the study of “life-as-it-could-be” in contrast to biology’s traditional study of “life-as-we-know-it.” The ambition was to take biology beyond the single example of terrestrial carbon-based life and to use synthetic and computational methods to discover what properties were essential to life in general versus what properties were merely contingent features of the organisms we happen to know.

This ambition did not arise in a vacuum. ALife crystallised several decades of preceding work that had grappled with the logic of biological organisation from a computational and systems perspective. Understanding ALife requires placing it within three intellectual traditions: cybernetics, artificial intelligence, and theoretical biology.

1.2 Cybernetics and Its Legacy

Norbert Wiener published Cybernetics: Or Control and Communication in the Animal and the Machine in 1948, founding the science of feedback-regulated systems. Wiener drew a deep analogy between biological regulation—homeostasis, goal-directed behaviour, nervous system function—and the engineering of automatic control systems. The central concept was the negative feedback loop: a system’s output is monitored, compared to a target, and the difference (error signal) is fed back to correct the system’s behaviour. This abstracted the logic of purposive behaviour away from its physical substrate, enabling a common mathematical vocabulary for engineers and biologists.

The Macy Conferences (1946–1953) brought together Wiener, John von Neumann, Warren McCulloch, Claude Shannon, Gregory Bateson, and Margaret Mead, among others, to develop this cross-disciplinary programme. Several threads emerged that were directly ancestral to ALife:

Cybernetics — The scientific study of regulatory systems, their structures, constraints, and possibilities, with emphasis on feedback, information flow, and goal-directed behaviour in both biological and artificial systems.

Warren McCulloch and Walter Pitts published “A Logical Calculus of the Ideas Immanent in Nervous Activity” in 1943, showing that networks of idealised neurons could in principle compute any logical function expressible in propositional calculus. Their work demonstrated that biological neural computation could be formulated mathematically, anticipating both connectionist AI and the study of complex networks.

Von Neumann contributed to cybernetics his theory of reliable computation from unreliable components and, most significantly for ALife, his detailed analysis of self-reproducing automata. The question that occupied him late in his career — how can a physical system produce an exact copy of itself, including a copy of the instructions for doing so? — became one of the defining problems of ALife.

Shannon’s information theory provided a rigorous measure of information content independent of meaning, which proved applicable both to the analysis of genetic information and to the study of complexity in evolving systems.

1.3 Turing’s Contributions

Alan Turing made two contributions decisive for ALife. The first is universally known: the theoretical Turing machine established what computation is, independent of any physical implementation. The Universal Turing Machine — a machine that can simulate any other Turing machine when given that machine’s description as input — is a formal model of the universality of computation. Von Neumann’s self-reproducing automaton construction was directly inspired by Turing’s insight that a universal computer could treat its own description as data.

The second contribution, less famous but equally important for ALife, is Turing’s 1952 paper “The Chemical Basis of Morphogenesis.” Turing asked how a spatially uniform collection of cells could spontaneously break symmetry and develop into a spatially patterned organism. His answer — reaction-diffusion systems — showed that two interacting chemical species (morphogens), one activating and one inhibiting, with different diffusion rates, could spontaneously generate stable spatial patterns from a nearly uniform initial state. This was among the first demonstrations that complex biological order could emerge from simple local rules through self-organisation, a paradigm that permeates modern ALife.

1.4 Properties of Life

A central theoretical concern in ALife is defining what life is — not as a list of features contingent on terrestrial biochemistry, but as a set of functional or organisational properties that any living system must exhibit. Several influential frameworks have been proposed.

The NASA working definition (Horowitz/Joyce formulation): life is a self-sustaining chemical system capable of Darwinian evolution. This is deliberately minimal and substrate-neutral.

The Maturana–Varela autopoiesis concept: Humberto Maturana and Francisco Varela introduced the concept of autopoiesis (literally “self-making”) to characterise the fundamental organisation of living systems. An autopoietic system is a network of processes that (1) produces the components of the network, (2) maintains the boundary that distinguishes the system from its environment, and (3) is constituted by the very components it produces. A living cell is the paradigm case: its metabolic network produces the lipids and proteins that form the membrane, which in turn contains the metabolic network.

Autopoiesis — An organisational property of systems that produce and maintain themselves through a network of processes. An autopoietic system generates its own components, which in turn regenerate the processes. The system is self-bounding and self-maintaining. Distinguished from allopoietic systems, which produce something other than themselves.

Langton’s framework: Langton characterised living systems as possessing: metabolism (energy transduction and storage), self-replication (producing offspring from available materials), evolvability (heritable variation subject to selection), and autonomy (maintaining internal organisation against environmental fluctuation). He emphasised that life is fundamentally about the organisation of matter and energy flows — a pattern that matter flows through rather than a particular arrangement of matter.

The traditional list: A widely used pedagogical enumeration includes: (1) organisation — structurally bounded; (2) metabolism — energy transformation; (3) growth and reproduction — increasing in number or size; (4) adaptation — responding to environment; (5) response to stimuli; (6) homeostasis — maintaining internal stability; (7) heredity — passing traits to offspring; (8) evolution — heritable variation subject to selection.

1.5 Simulation vs. Synthesis

Langton articulated a crucial distinction between two methodological approaches to studying life computationally:

Simulation (top-down): Model a known biological system in a computer. Identify the mechanisms, encode them mathematically, run the model, verify against known data. This is the approach of computational biology — simulation is a tool for understanding systems we already know about.

Synthesis (bottom-up): Specify simple rules for simple components, allow them to interact, and observe what complex organisation emerges without prescribing it. This is the ALife approach — synthesis is a tool for discovering what kinds of organisation are possible and what rules give rise to them.

The synthetic approach is philosophically more radical. If a sufficiently life-like system is synthesised from non-living components according to simple rules, what follows? This question intersects with philosophy of mind (strong vs. weak AI debates), philosophy of biology (what is life?), and ethics (what moral status does artificial life have?).

1.6 The “Life-as-it-Could-Be” Programme

The deepest motivation for ALife is Langton’s insistence that we have only one example of life — terrestrial, carbon-based, descended from a common ancestor. This is a sample size of one from what may be a vast space of possible life forms. Any generalisation from this one example risks mistaking contingent features of Earth life for necessary features of life in general.

ALife aims to explore this space systematically. By implementing alternative life-forms in computers or in chemistry, researchers can study which properties are truly universal — necessary for life of any kind — and which are merely parochial features of the one example we happen to have. This includes asking whether life necessarily involves carbon chemistry, aqueous solvent, double-stranded DNA, the specific genetic code, sexual reproduction, or any other feature of terrestrial biology that we might naively treat as a defining characteristic.


Chapter 2: Biological Foundations

2.1 Evolution of Life on Earth: A Timeline

Life on Earth began approximately 3.8–4.0 billion years ago, most probably in aquatic environments. The geological and molecular fossil record provides a broad outline:

  • ~4.0 Ga: Chemical evolution; formation of organic molecules (amino acids, nucleotides) in prebiotic environments
  • ~3.8 Ga: First evidence of cellular life (microfossils, isotopic signatures)
  • ~3.5 Ga: Stromatolites — laminated structures produced by photosynthetic microbial mats
  • ~2.7 Ga: Oxygenic photosynthesis by cyanobacteria
  • ~2.1 Ga: Great Oxidation Event — atmospheric oxygen rises dramatically
  • ~1.8 Ga: Eukaryotic cells (cells with nuclei), almost certainly through endosymbiosis
  • ~1.2 Ga: First multicellular eukaryotes (algae)
  • ~0.6 Ga: Ediacaran fauna — soft-bodied multicellular animals
  • ~0.54 Ga: Cambrian explosion — rapid diversification of animal body plans
  • ~0.5 Ga–present: Colonisation of land, radiation of plants, insects, vertebrates, mammals; first primates ~65 Ma, Homo sapiens ~0.3 Ma

This timeline is critical for ALife because it reveals that the most significant innovations in the history of life — the origin of the cell, the origin of the nucleus, the origin of sex, the origin of multicellularity — were not gradual modifications of a pre-existing system but qualitative transitions in the mode of organisation. Maynard Smith and Szathmáry called these the “major transitions in evolution.”

2.2 Molecular Genetics: DNA, RNA, and Protein

The molecular foundation of terrestrial life consists of three types of macromolecules in mutual dependence:

DNA (deoxyribonucleic acid) is the hereditary material. It consists of two antiparallel polynucleotide strands wound into a double helix, held together by hydrogen bonds between complementary base pairs: adenine with thymine (A–T), guanine with cytosine (G–C). The sequence of bases along one strand constitutes a linear digital message. The complementarity of base pairing provides the mechanism for accurate copying: each strand serves as a template for the synthesis of a new complementary strand.

RNA (ribonucleic acid) is structurally similar to DNA but uses ribose rather than deoxyribose as the sugar, and uracil (U) instead of thymine. RNA serves as (1) messenger RNA (mRNA), carrying sequence information from DNA to ribosomes; (2) transfer RNA (tRNA), the adapter molecules that match codons to amino acids; (3) ribosomal RNA (rRNA), a catalytic and structural component of the ribosome; and (4) various regulatory RNAs.

Proteins are linear polymers of amino acids, of which there are twenty standard types. The sequence of amino acids determines the three-dimensional folded structure (via thermodynamic minimisation of free energy), and the structure determines function. Proteins serve as enzymes (catalysts), structural materials, signalling molecules, transport vehicles, and much else.

2.3 The Central Dogma and Its Digital Character

Francis Crick’s “central dogma” of molecular biology (1958, elaborated 1970) states that sequence information flows from DNA to RNA to protein, but not in the reverse direction. More precisely:

\[ \text{DNA} \xrightarrow{\text{replication}} \text{DNA} \xrightarrow{\text{transcription}} \text{RNA} \xrightarrow{\text{translation}} \text{Protein} \]

The digital character of this system is remarkable and has been a key inspiration for ALife. The genetic message is a linear sequence of discrete symbols drawn from an alphabet of four bases (A, T/U, G, C). This is structurally identical to a digital string. Three consecutive bases form a codon. There are \(4^3 = 64\) possible codons, which specify 20 amino acids plus stop signals — the genetic code is the mapping from codons to amino acids.

The genetic code is degenerate (redundant): most amino acids are encoded by multiple codons. This degeneracy is non-random — synonymous codons (those encoding the same amino acid) tend to differ only in the third position of the codon (the "wobble position"), providing partial error buffering. The code is also nearly universal across all life — strong evidence for a common ancestor and for the code's early fixation.

Protein biosynthesis proceeds as follows:

  1. Transcription: RNA polymerase binds to a promoter sequence on DNA and reads the template strand in the 3’→5’ direction, synthesising mRNA in the 5’→3’ direction. The mRNA sequence is complementary to the template strand and identical to the coding strand (with U replacing T).

  2. Translation: The ribosome (a complex of rRNAs and proteins) binds the mRNA at the start codon (AUG, encoding methionine). tRNA molecules, each charged with a specific amino acid, recognise the mRNA codon by their complementary anticodon. The ribosome catalyses peptide bond formation between successive amino acids and moves along the mRNA, extending the polypeptide chain until a stop codon is reached.

  3. Protein folding: The completed polypeptide chain spontaneously folds into its native three-dimensional structure in milliseconds to seconds, guided by interactions among amino acid side chains and with the aqueous environment. Chaperone proteins assist in preventing misfolding.

This entire process is an instance of algorithmic information processing: a digital program (the DNA/mRNA sequence) is interpreted by a computing machinery (the ribosome) to produce functional outputs (proteins). The analogy to computer programs is not merely metaphorical — von Neumann identified it as structurally precise decades before molecular genetics was fully worked out.

2.4 Darwinian Evolution: Core Framework

Darwin’s theory of evolution by natural selection requires three conditions:

  1. Heritable variation: Individuals in a population differ in heritable traits.
  2. Differential fitness: Different variants have different average reproductive success (fitness) in a given environment.
  3. Heredity: Offspring resemble parents more than they resemble random members of the population.

When these conditions hold, allele frequencies in the population will shift over generations in the direction of higher average fitness — this is natural selection. The modern synthesis integrates Darwinian selection with Mendelian genetics and population genetics, adding:

  • Mutation: Random changes to DNA sequence, providing the raw material of variation
  • Recombination: Shuffling of genetic material during sexual reproduction
  • Genetic drift: Random fluctuations in allele frequency, most significant in small populations
  • Gene flow: Exchange of alleles between populations

Fitness landscapes, introduced by Sewall Wright (1932), provide a geometric visualisation of evolution. The “landscape” maps genotype space (or phenotype space) to reproductive fitness. Peaks represent locally optimal genotypes; valleys represent low-fitness combinations. Natural selection drives populations “uphill” in this landscape. The ruggedness of the landscape — how many peaks it has, how separated they are — determines the difficulty of evolutionary search.

Fitness landscape — A mapping from the space of possible genotypes (or phenotypes) to fitness values, forming a high-dimensional surface. Local maxima are locally optimal genotypes; global maxima are the most fit. The geometry of the landscape constrains evolutionary dynamics: smooth (single-peaked) landscapes permit efficient hill-climbing; rugged (multi-peaked) landscapes trap populations at local optima.

2.5 Digital Organisms: Tierra and Avida

The concept of digital organisms — self-replicating computer programs evolving in a simulated computational environment — is a central experimental platform of ALife.

Tierra (Thomas Ray, 1991): Ray created a virtual computer with a special assembly-language instruction set, and introduced into it a single self-replicating “ancestor” program of 80 instructions. The program allocates memory, copies itself instruction by instruction, and divides to produce an offspring. Mutations (random instruction changes) occur during copying. An operating-system layer called the “Reaper” kills the oldest or most error-prone organisms when memory fills up. Ray observed spontaneous evolution: parasites emerged (organisms that used the ancestor’s copy loop without having their own), then hyper-parasites (which exploited the parasites), then social organisms (which could only replicate in clusters). This demonstrated that digital media can support open-ended Darwinian evolution, including ecological dynamics.

Avida (Adami, Ofria & Collier, 1994; extensively developed by Ofria and colleagues): A more controlled and scientifically rigorous platform. Avida organisms are programs in a virtual CPU with a register-based instruction set. Each organism occupies a cell in a two-dimensional grid. Fitness is determined by metabolic rate, which increases when the organism executes logic operations (NAND, AND, OR, etc.) on numbers injected into its environment. Avida has been used to demonstrate: the evolution of complex logic functions from simpler ones, the evolution of sex, Lenski-style long-term evolution experiments, the importance of environmental complexity for organismal complexity, and direct tests of theoretical predictions from population genetics.

Lenski et al. (2003) Avida experiment: Researchers asked whether complex features could evolve by natural selection. They examined the evolution of the ability to perform EQU (logical equality) — a complex operation requiring a sequence of simpler operations as intermediate steps. They found that when simpler operations were rewarded, populations evolved EQU far more readily than when only EQU was rewarded. This supported the claim that evolution climbs fitness ladders step by step, with earlier steps providing scaffolding for later ones. Deleting the intermediate rewards prevented the evolution of EQU in most runs.

Chapter 3: Cellular Automata and Automata Networks

3.1 Cellular Automata: Definition and Motivation

A cellular automaton (CA) is a discrete dynamical system consisting of a regular grid of cells, each in one of a finite number of states, updated synchronously according to a local rule that depends only on the cell’s own state and the states of a defined neighbourhood of cells. The update rule is applied simultaneously to every cell at each time step.

CAs were introduced by von Neumann and Stanislaw Ulam in the late 1940s. Von Neumann was seeking a model of self-reproduction that would demonstrate that a sufficiently complex automaton could build copies of itself. Ulam suggested that two-dimensional automata provided a good medium.

The formal specification of a CA involves:

  • A grid geometry (1D line, 2D rectangular grid, triangular, hexagonal, etc.)
  • A state set \(S\) of finite cardinality \(k\)
  • A neighbourhood template \(N\) specifying which cells influence each cell
  • A local transition rule \(f: S^{|N|} \to S\)
  • Boundary conditions (periodic, fixed, absorbing)

3.2 Wolfram’s Classification of 1D CAs

Stephen Wolfram undertook a systematic study of one-dimensional, two-state, three-cell-neighbourhood CAs in the 1980s (see also A New Kind of Science, 2002). There are \(2^{2^3} = 256\) such rules, each identified by a “Wolfram code.” Wolfram observed that their long-term behaviour fell into four qualitative classes:

Wolfram's CA classes:
  • Class I — Evolution leads rapidly to a fixed, spatially uniform state (all cells same state). Examples: Rules 0, 255.
  • Class II — Evolution leads to simple periodic structures — repetitive patterns of fixed period. Examples: Rules 4, 108.
  • Class III — Evolution produces apparently random (aperiodic, chaotic) behaviour. Examples: Rules 30, 45.
  • Class IV — Evolution generates complex, localised structures with long transients — neither fully regular nor fully random. Examples: Rules 54, 110.

Wolfram proposed — controversially — that Class IV rules correspond to computation at the “edge of chaos,” a region of maximal computational capacity. Rule 110 was later proved by Matthew Cook (2004) to be Turing-complete: it can simulate a universal Turing machine.

Langton formalised this intuition with the lambda parameter (\(\lambda\)): the fraction of transitions in the rule table that map to a non-quiescent (non-zero) state. As \(\lambda\) increases from 0 to 1, CA behaviour passes through Classes I, II, IV, III in order. Class IV behaviour tends to occur near a phase transition between ordered and chaotic phases, a region Langton identified as the “edge of chaos” where computation is maximised.

3.3 Conway’s Game of Life

The most famous two-dimensional CA is John Horton Conway’s Game of Life (1970), described by Martin Gardner in Scientific American. The rules are:

  • Grid: infinite 2D square lattice
  • States: each cell is alive (1) or dead (0)
  • Neighbourhood: Moore neighbourhood (8 surrounding cells)
  • Rules:
    • A live cell with 2 or 3 live neighbours survives
    • A dead cell with exactly 3 live neighbours becomes alive
    • All other live cells die; all other dead cells remain dead

Despite — and because of — their simplicity, these rules support extraordinary complexity. Conway’s Game of Life contains:

  • Still lifes: stable patterns (block, beehive, loaf)
  • Oscillators: periodic patterns (blinker, pulsar, pentadecathlon)
  • Spaceships: patterns that translate through the grid while maintaining shape (glider, lightweight spaceship)
  • Guns: patterns that periodically emit spaceships (Gosper Glider Gun)
  • Self-replicating patterns: patterns that create copies of themselves
  • Universal computation: the Game of Life is Turing-complete; logic gates and memory can be constructed from glider streams

The Game of Life was a proof of concept that extreme complexity — including computation — could emerge from the simplest possible local rules on a regular grid.

3.4 Von Neumann’s Self-Reproducing Automaton

Von Neumann’s original CA for self-reproduction used 29 cell states and a neighbourhood of 5 cells (von Neumann neighbourhood: cell plus 4 orthogonal neighbours). The self-reproducing configuration contains approximately 200,000 cells and operates as follows:

  1. The description (tape): a stored representation of the constructor, encoded as a sequence of cell states
  2. The universal constructor: machinery that reads the description and builds the described structure
  3. The copy operation: after building, the constructor also copies the description tape and inserts it into the offspring
  4. Activation: the offspring is then switched on

This design solves the bootstrapping problem of self-reproduction: the constructor needs to be described in order to be copied, but the description also needs a constructor to execute it. Von Neumann’s solution — using the description in two modes, as data (for copying) and as instructions (for building) — is structurally identical to the solution used by biological cells, where DNA is copied by replication machinery that is itself encoded in the DNA.

Von Neumann's self-reproduction theorem: A sufficiently complex automaton (one that is computationally universal) can, in principle, construct any automaton whose description it is given, including a complete copy of itself. Self-reproduction in complex systems necessarily involves a stored description (genotype) that is both executed and copied, just as in biological organisms.

3.5 Synchronous vs. Asynchronous Automata Networks

Standard CAs update all cells simultaneously (synchronously). However, biological systems rarely operate this way — neurons fire asynchronously, gene expression is stochastic, cell division is not synchronised across a tissue. Asynchronous automata networks update one or a subset of cells at each step.

Several updating schemes exist:

  • Fully asynchronous: one cell chosen uniformly at random per step
  • Sequential: cells updated in a fixed sequential order
  • Stochastic: each cell updated independently with probability \(p\) per step
  • Event-driven: cells updated only when their neighbourhood changes

The dynamics of asynchronous updates differ qualitatively from synchronous updates. Many patterns that are stable or oscillatory under synchronous updates become unstable or show different period under asynchronous updates. For systems modelling biological processes, asynchronous models often capture biological behaviour more faithfully.

3.6 Boolean Networks and Kauffman’s NK Model

Stuart Kauffman introduced random Boolean networks (RBNs) as a model of genetic regulatory networks (GRNs). A Boolean network consists of \(N\) binary variables (genes, each “on” or “off”), each regulated by \(K\) other variables according to a Boolean function chosen randomly. The network is updated (typically synchronously): each node’s new state is determined by its Boolean function applied to the current states of its \(K\) inputs.

Random Boolean Network (RBN / NK model) — A network of N binary nodes, each with K inputs chosen randomly from other nodes. The transition function for each node is a Boolean function of its K inputs, also chosen randomly. The network iterates deterministically from any initial state; since the state space is finite (\(2^N\) states), the network must eventually enter a cycle (attractor).

Key results from Kauffman’s analysis:

Phase transitions: As \(K\) varies, the network passes through three phases:

  • \(K < 2\): Ordered phase — most initial conditions lead to very few attractors with short periods; most of the state space freezes
  • \(K = 2\): Critical phase — the network sits at the boundary between order and chaos; number of attractors scales as \(\sqrt{N}\), period scales as \(\sqrt{N}\)
  • \(K > 2\): Chaotic phase — exponentially many attractors with exponentially long periods; tiny perturbations propagate through the network

Kauffman proposed that cells sit in the critical or ordered regime (\(K \approx 2\)), which explains why typical mammalian cells have \(\sim 300\) cell types with \(\sim 25,000\) genes — comparable to \(\sqrt{N} \approx \sqrt{25000} \approx 158\).

Attractors as cell types: Each attractor (cycle of states) in an RBN represents a stable pattern of gene activity, analogous to a cell type. The number and structure of attractors predicts the number of cell types. This is a purely dynamical, substrate-independent theory of cell-type differentiation.

3.7 Genetic Regulatory Networks

Real genetic regulatory networks share important structural features with RBN models but differ in details:

  • Connectivity is not random but shaped by evolution to achieve specific functions
  • Regulation involves continuous variables (mRNA and protein concentrations)
  • Regulatory interactions involve transcription factors binding to promoter regions
  • Network dynamics show both robust (canalised) and sensitive (plastic) responses

Waddington’s epigenetic landscape visualises cellular differentiation: a ball (cell) rolls down a landscape of valleys, settling in stable attractors (cell fates). Gene regulatory networks shape this landscape; development corresponds to trajectories through it.

Key structural motifs in GRNs include:

  • Negative feedback loops: maintain oscillatory or homeostatic gene expression (e.g., circadian clocks)
  • Positive feedback / bistable switches: allow cells to “commit” irreversibly to a fate
  • Feed-forward loops: provide signal filtering, delay, or pulse generation
  • Coherent vs. incoherent feed-forward: produce different input-output responses

Chapter 4: Swarm Intelligence and Stigmergy

4.1 Collective Behaviour and Emergence

Swarm intelligence studies how complex, adaptive collective behaviour arises from the local interactions of many simple agents — biological or artificial — without centralised control. The defining characteristic is that no individual agent has a global view of the system state or explicit knowledge of the collective goal being pursued. The collective intelligence is an emergent property of local interactions.

Biological examples include: ant and termite colonies, honeybee colonies, bird flocking (murmuration), fish schooling, bacterial quorum sensing, immune system response, and social foraging in many species. The adaptive coordination seen in these systems often rivals or surpasses what a single, highly intelligent agent could achieve, despite the relative simplicity of individual agents.

Stigmergy — Indirect coordination of behaviour through modifications to a shared environment. An agent's action modifies the environment; this modification stimulates and guides subsequent actions by the same or different agents. Coined by Pierre-Paul Grassé (1959) to describe termite nest construction, where a termite deposits material, creating a stimulus that attracts other termites to deposit more material in the same location. No individual termite has a plan of the nest; the plan is encoded in the emerging structure itself.

Stigmergy is a powerful mechanism because it decouples the producer of a signal from its consumer in both space and time. Unlike direct communication (where signal lifetime is limited), stigmergic signals can persist and accumulate in the environment, enabling coordination over extended timescales without the agents themselves needing persistent memory.

4.2 Ant Colony Optimization

The foraging behaviour of real ants provides one of the most studied instances of stigmergic collective intelligence. When ant colonies forage, they initially explore their environment via random walks. Upon discovering food, a returning ant deposits a chemical signal — a pheromone trail — along its return path. Other ants, encountering pheromone, are biased to follow it. If the path to food is short, ants on that path complete their round trips faster and deposit pheromone more frequently, so the trail becomes stronger. Pheromone also evaporates at a fixed rate. The result is a positive-feedback / negative-feedback balance that causes the colony to converge on shorter paths — a distributed shortest-path algorithm.

Marco Dorigo formalised this insight as the Ant Colony Optimization (ACO) metaheuristic, beginning with his doctoral thesis (1992) and consolidated in Ant Colony Optimization (Dorigo & Stützle, 2004).

The ACO framework:

Let \(G = (V, E)\) be a graph whose edges represent problem components. Each edge \((i,j)\) has a pheromone level \(\tau_{ij}\) and a heuristic value \(\eta_{ij}\) (problem-specific desirability). Each artificial ant \(k\) builds a solution by probabilistically moving from node to node:

\[ p_{ij}^k = \frac{\left[\tau_{ij}\right]^\alpha \left[\eta_{ij}\right]^\beta}{\sum_{l \in \mathcal{N}_k} \left[\tau_{il}\right]^\alpha \left[\eta_{il}\right]^\beta} \]

where \(\mathcal{N}_k\) is the set of nodes not yet visited by ant \(k\), and \(\alpha, \beta\) are parameters controlling the relative influence of pheromone and heuristic information.

After all ants complete their tours, pheromone is updated:

\[ \tau_{ij} \leftarrow (1 - \rho)\tau_{ij} + \sum_{k=1}^{m} \Delta\tau_{ij}^k \]

where \(\rho \in (0,1)\) is the evaporation rate and \(\Delta\tau_{ij}^k\) is the pheromone deposited by ant \(k\) on edge \((i,j)\), typically \(1/L^k\) if the ant used that edge (where \(L^k\) is its tour length), and 0 otherwise.

ACO has been successfully applied to: the Travelling Salesman Problem (competitive with other metaheuristics), vehicle routing, job-shop scheduling, network routing, and protein folding.

4.3 Particle Swarm Optimization

Particle Swarm Optimization (PSO) was introduced by James Kennedy and Russell Eberhart (1995), inspired by the flocking of birds and schooling of fish. PSO operates in continuous parameter spaces and has been widely applied to numerical optimisation.

Each particle \(i\) in the swarm has position \(\mathbf{x}_i\) and velocity \(\mathbf{v}_i\) in the search space. Each particle remembers its personal best position \(\mathbf{p}_i\) and shares information about the global best position \(\mathbf{g}\) found by any particle in the neighbourhood.

Update equations:

\[ \mathbf{v}_i \leftarrow w\mathbf{v}_i + c_1 r_1 (\mathbf{p}_i - \mathbf{x}_i) + c_2 r_2 (\mathbf{g} - \mathbf{x}_i) \]\[ \mathbf{x}_i \leftarrow \mathbf{x}_i + \mathbf{v}_i \]

Here \(w\) is the inertia weight (controlling exploration vs. exploitation), \(c_1\) and \(c_2\) are cognitive and social acceleration coefficients, and \(r_1, r_2\) are uniform random values in \([0,1]\).

The cognitive component (\(c_1 r_1 (\mathbf{p}_i - \mathbf{x}_i)\)) attracts the particle toward its own best-known position. The social component (\(c_2 r_2 (\mathbf{g} - \mathbf{x}_i)\)) attracts it toward the globally best-known position. The balance determines the exploration-exploitation tradeoff.

PSO is parameter-free in the sense that it requires no gradient information, making it suitable for non-differentiable, noisy, and multimodal objective functions. It has been applied to neural network training, feature selection, game theory, and many engineering design problems.

4.4 Bee Algorithms and Other Swarm Models

Honeybee foraging provides another celebrated model of swarm intelligence. Scout bees explore the environment and report rich food sources by performing the waggle dance, whose duration encodes distance and whose orientation encodes direction relative to the sun. The probability that a scout will advertise a source is proportional to its quality; the number of followers recruited depends on dance intensity; and bees may abandon sources. This results in collective decision-making that identifies and exploits the best available food sources.

The Bee Algorithm and Artificial Bee Colony (ABC) algorithm of Karaboga (2005) translate this into an optimisation metaheuristic with employed bees (exploit known sources), onlooker bees (observe dances and select sources probabilistically), and scout bees (random exploration).

Flocking models (Boids): Craig Reynolds (1987) demonstrated that realistic bird flocking could be simulated with three simple rules per agent: (1) separation — avoid collisions with nearby flock mates; (2) alignment — steer toward the average heading of local flock mates; (3) cohesion — steer toward the average position of local flock mates. No individual agent tries to form or maintain a flock — the flock emerges from purely local interactions.

4.5 Stigmergy in Engineering and Robotics

Stigmergic principles have been applied extensively in robotics and multi-agent systems:

  • Task allocation: In ant-inspired multi-robot task allocation, robots deposit virtual pheromone signals (in a shared data structure) when completing tasks, guiding other robots toward high-priority work
  • Collective construction: robots building structures by depositing and responding to blocks follow stigmergic protocols analogous to termite nest building
  • Distributed sensing: sensor networks that amplify signals by passing them between nodes with local thresholds implement stigmergic information aggregation
  • Network routing: the first practical application of ACO, AntNet (Di Caro & Dorigo, 1998), demonstrated competitive routing in packet-switched networks

Chapter 5: Self-Reproducing Systems

5.1 The Logic of Self-Reproduction

Self-reproduction is one of the most fundamental properties attributed to living systems. Von Neumann analysed the logical structure of self-reproduction and showed that it requires more than naive copying — it requires a mechanism that uses its own description in two distinct ways. This insight anticipated molecular biology by several years.

Von Neumann identified a logical paradox: a machine that simply copies what it contains would contain only an infinite regress. The solution is to introduce a stored description (the genotype) that is:

  1. Interpreted by the constructor to build the structure described
  2. Copied by a separate copying mechanism to produce the offspring’s description

This is exactly how biological cells reproduce: the DNA is transcribed and translated (interpreted) to produce proteins, and also replicated (copied) by DNA polymerase to produce offspring DNA.

5.2 Langton’s Self-Reproducing Loops

In 1984, Christopher Langton simplified von Neumann’s 200,000-cell construction to a compact self-reproducing loop operating in a CA with 8 states and the von Neumann neighbourhood. The loop consists of a rectangular sheath of “wall” cells enclosing a “core” of cells that circulate around the loop. The core contains a sequence of instructions that, when the loop is executed, extends an arm, forms a new loop at the arm’s end, transfers a copy of the instruction sequence into the new loop, and detaches it.

Langton’s loops demonstrated that self-reproduction did not require the full complexity of von Neumann’s construction — simpler self-reproducing structures were possible. Subsequent researchers (Byl, Reggia, Perrier, Tempesti) produced even smaller self-reproducing CAs, exploring the minimum complexity required.

However, Langton’s loops replicate but do not evolve — mutation destroys the loop structure rather than producing viable variants. This distinction between replication and evolvability is important: not all self-replicating systems support open-ended evolution.

5.3 Tierra and the Ecology of Digital Evolution

Thomas Ray’s Tierra (1991) went further than CAs by creating a digital environment in which self-replicating programs could mutate and evolve. Key features:

  • Instruction set: designed to be evolvable — instructions are forgiving (some errors are tolerable) and there is no distinction between code and data
  • Mutations: point mutations (random instruction substitution) during copying, and flaws (random execution of wrong instruction) during computation
  • Fitness: implicit — organisms that replicate faster and consume less memory survive longer before being killed by the Reaper
  • No explicit fitness function: organisms are not optimised toward a predetermined goal; they evolve toward efficient self-replication in the face of competition

Tierra demonstrated: spontaneous evolution of parasites, host-parasite coevolution, evolution of immunity, evolution of sociality, and ecological succession — all without being programmed. These are open-ended evolutionary dynamics driven entirely by competition and selection in an artificial medium.

5.4 RNA World and Minimal Life

The RNA world hypothesis proposes that the earliest life on Earth used RNA both as information carrier (like DNA) and as catalyst (like proteins), before the specialisation of DNA for storage and proteins for catalysis that characterises all modern life. Evidence includes:

  • Ribozymes: naturally occurring RNA molecules with catalytic activity (ribonuclease P, self-splicing introns, hepatitis delta virus ribozyme)
  • The ribosome: the catalytic centre of translation is RNA (ribosomal RNA), not protein
  • RNA can fold into complex three-dimensional structures supporting catalysis
  • In vitro evolution (SELEX) can generate RNA molecules with novel binding and catalytic activities

Thomas Cech and Sidney Altman won the 1989 Nobel Prize in Chemistry for discovering ribozymes. The RNA world posits a pre-biotic period in which:

  1. RNA molecules could replicate (using other RNA as templates or autocatalytically)
  2. Selection favoured RNA sequences with better catalytic properties
  3. RNA eventually “invented” protein synthesis, and proteins gradually took over most catalytic roles due to greater chemical versatility
  4. DNA replaced RNA as the hereditary material due to its greater chemical stability

Minimal life research asks: what is the smallest set of components sufficient for a self-replicating, evolving system? Experiments with self-replicating peptides, minimal RNA replicases, and fatty acid vesicles that spontaneously form, grow, and divide point toward protocells as the pre-cellular stage.

5.5 Artificial Cells

A convergent research programme aims to build artificial cells — minimal systems that exhibit key properties of living cells, assembled from non-living components. Approaches include:

  • Lipid vesicles encapsulating RNA and ribosomes: Jack Szostak’s group has demonstrated protocells in which vesicle growth and division can be coupled to RNA replication
  • JCVI minimal genome project (Venter et al.): reducing the Mycoplasma mycoides genome to the minimum essential gene set; JCVI-syn3.0 (2016) has 473 genes, ~57% of the original genome, and is the smallest self-replicating synthetic organism created
  • Cell-free transcription-translation (TXTL) systems: all components of protein synthesis assembled in a test tube, capable of executing genetic programs in vitro
  • DNA origami and synthetic compartments: engineered nucleic acid structures that mimic cellular organisation

Chapter 6: Evolutionary Computation

6.1 Holland’s Genetic Algorithm

John Holland developed the genetic algorithm (GA) in the 1960s–70s as a formal theory of adaptive systems, published in Adaptation in Natural and Artificial Systems (1975). Holland was interested in designing computer systems that could adapt as well as natural systems — and in understanding why natural selection worked as well as it did.

The canonical genetic algorithm operates as follows:

  1. Initialise a population of \(N\) candidate solutions (chromosomes), typically binary strings of fixed length \(L\)
  2. Evaluate the fitness \(f(x)\) of each chromosome
  3. Select parents for reproduction: selection probability proportional to fitness (roulette-wheel selection)
  4. Crossover: with probability \(p_c\), exchange segments between two parents at a random crossover point
  5. Mutation: with probability \(p_m\) per bit, flip each bit
  6. Replace the old population with the new one
  7. Repeat from step 2 until termination criterion
Single-point crossover: Given chromosomes A = 11001010 and B = 01110001, with crossover point at position 5: offspring are A' = 11001001 and B' = 01110010. The first offspring takes the left portion of A and the right portion of B; the second takes the complementary segments.

Holland’s Schema Theorem provides a theoretical justification for GA effectiveness. A schema is a template specifying fixed values at some positions and “don’t cares” (#) at others. A schema \(H\) is characterised by its order \(o(H)\) (number of fixed positions), its defining length \(\delta(H)\) (distance between the outermost fixed positions), and its fitness \(f(H)\) (average fitness of strings matching \(H\)).

The Schema Theorem states that short, low-order, above-average schemata receive exponentially increasing trials from one generation to the next. This “building block hypothesis” suggests that GAs work by combining short, high-fitness schema (building blocks) into increasingly fit solutions — a parallel search over an exponential number of schemata using a population of polynomial size.

Schema Theorem (Holland): The expected number of strings matching schema H in generation t+1 satisfies: \[ m(H, t+1) \geq m(H,t) \cdot \frac{f(H)}{\bar{f}} \cdot \left[1 - p_c \frac{\delta(H)}{L-1} - o(H) \cdot p_m\right] \]

where f(H) is the average fitness of strings matching H, \(\bar{f}\) is the population mean fitness, p_c is the crossover rate, p_m is the mutation rate per bit, and L is the chromosome length. Schemata with above-average fitness, short defining length, and low order receive exponentially increasing representation.

6.2 Genetic Programming

John Koza’s Genetic Programming (1992) extends genetic algorithms from fixed-length bit strings to variable-length, tree-structured programs. The candidates are programs (parse trees) in some language (typically Lisp s-expressions), and fitness is evaluated by running the program.

Crossover in GP: select a random node in each parent tree and swap the subtrees rooted at those nodes. The offspring are valid programs of varying size.

Mutation in GP: replace a randomly selected subtree with a randomly generated subtree.

GP has been applied to: symbolic regression (finding mathematical expressions that fit data), classification, program induction, electronic circuit design, game playing, and robotics controller evolution. Koza coined the term “human-competitive results” to describe cases where GP has produced solutions comparable to or better than those designed by human experts.

6.3 Evolution Strategies and Differential Evolution

Evolution Strategies (ES) were developed independently by Rechenberg and Schwefel in Germany in the 1960s–70s, applied primarily to continuous parameter optimisation. Key features:

  • Chromosomes are real-valued vectors
  • Mutation by adding Gaussian noise: \(\mathbf{x}' = \mathbf{x} + \sigma \mathcal{N}(0, \mathbf{I})\)
  • The mutation step size \(\sigma\) is adapted during search — a meta-parameter that evolves alongside the solution
  • Selection is typically truncation selection (the top \(\mu\) of \(\lambda\) offspring survive)
  • The CMA-ES (Covariance Matrix Adaptation Evolution Strategy) adapts the full covariance matrix of the mutation distribution, making it highly effective on non-separable problems

Differential Evolution (Storn & Price, 1997) generates new candidates by adding scaled differences between pairs of population members to a base vector — a form of directional mutation that automatically adapts to the problem structure.

6.4 Fitness Landscapes and the NK Model

Kauffman’s NK model provides a formal model of fitness landscape ruggedness. Each organism is characterised by a genome of \(N\) binary loci; each locus contributes a fitness component that depends on its own value and the values of \(K\) other “epistatic” loci. The overall fitness is the average of all \(N\) contributions:

\[ f = \frac{1}{N} \sum_{i=1}^{N} f_i\left(a_i, a_{i_1}, \ldots, a_{i_K}\right) \]

where \(f_i\) values are assigned randomly from \([0,1]\) for each possible combination of the \(K+1\) loci values.

  • \(K = 0\): each locus contributes independently; single-peaked, smooth landscape; easy for evolution to find the optimum
  • \(K = N-1\): every locus depends on every other; maximally rugged landscape (equivalent to a random energy model); exponentially many local optima; evolution easily gets trapped
  • Intermediate \(K\): landscapes of intermediate ruggedness; tuning \(K\) tunes evolvability

The NK model predicts: as epistasis increases (K increases), the number of local optima grows rapidly, the correlation between neighbouring genotypes decreases, and the expected fitness of locally optimal genotypes decreases. This implies a fundamental tension between ruggedness (which hampers search) and the functional requirements that create ruggedness.

6.5 Sex and Recombination: Why Evolve Sexually?

The evolution of sexual reproduction is one of the central puzzles of evolutionary biology — because sex imposes a twofold fitness cost (only females produce offspring; males are “wasted”) that any asexual mutant would not pay. Sex must therefore confer substantial advantages to persist in the face of this cost.

Several hypotheses:

Red Queen hypothesis (Van Valen, 1973): Parasites and pathogens evolve rapidly to exploit the most common host genotypes. Sexual reproduction continuously generates novel genotype combinations, keeping hosts one step ahead of rapidly evolving pathogens. The “Red Queen” runs to stay in place — constant evolutionary change is required just to maintain relative fitness.

Mutational determinism hypothesis (Kondrashov): Sexual recombination purges deleterious mutations more efficiently than asexual reproduction, which causes mutational load to accumulate (Muller’s ratchet).

Fisher-Muller hypothesis: Recombination brings together favourable mutations from different lineages, accelerating the rate of adaptive evolution.

In the context of GAs, recombination (crossover) is justified by the building-block hypothesis: if high-fitness genotypes are combinations of compatible partial solutions (building blocks), recombination assembles these faster than mutation alone. However, the effectiveness of recombination depends strongly on linkage disequilibrium (whether high-fitness alleles are correlated in the population) and the ruggedness of the fitness landscape.

6.6 Co-evolution and Open-Ended Evolution

Co-evolution occurs when the fitness of organisms in one population depends on the composition of other populations (predators, prey, parasites, hosts, competitors, mutualists). This creates a perpetually shifting fitness landscape: optimising against current opponents may be suboptimal against future opponents.

Red Queen dynamics in artificial co-evolution: Populations caught in co-evolutionary cycles may make no persistent progress in absolute terms (each advance by one population is countered by the other), despite continuous evolutionary change. This produces complex, non-stationary dynamics without a fixed fitness landscape.

Open-ended evolution is the goal of generating artificial systems that continue to produce novelty indefinitely, as biological evolution apparently does. Current digital evolution systems (Tierra, Avida) tend toward extinction or stagnation rather than genuine open-ended exploration. Researchers have proposed that open-ended evolution requires: (1) unlimited complexity growth; (2) unbounded possibility space; (3) co-evolutionary dynamics maintaining perpetual novelty; (4) major transitions — qualitative changes in the unit of selection or mode of inheritance.


Chapter 7: Growth and Morphogenesis

7.1 The Problem of Morphogenesis

How does a single fertilised egg cell, following a genetic program, reliably develop into a complex multicellular organism with differentiated cell types, tissues, organs, and precise spatial arrangements? This is the problem of morphogenesis — the generation of biological form.

Morphogenesis involves: (1) differentiation — cells acquiring different identities from a shared genome; (2) spatial patterning — coordinating cell behaviour across spatial scales; (3) growth and proliferation — regulated increase in cell number and mass; (4) morphogenetic movements — cells migrating, folding sheets, extending protrusions.

7.2 Turing Reaction-Diffusion Systems

Alan Turing’s 1952 paper “The Chemical Basis of Morphogenesis” is one of the most important papers in theoretical biology. Turing proposed that two chemical species — an activator and an inhibitor — interacting through autocatalysis and mutual inhibition, with different diffusion rates, could spontaneously break the symmetry of an initially uniform chemical field and generate stable spatial patterns.

The mathematical form is a reaction-diffusion PDE system:

\[ \frac{\partial u}{\partial t} = D_u \nabla^2 u + f(u, v) \]\[ \frac{\partial v}{\partial t} = D_v \nabla^2 v + g(u, v) \]

where \(u\) is activator concentration, \(v\) is inhibitor concentration, \(D_u\) and \(D_v\) are diffusion coefficients, and \(f, g\) are reaction kinetics. For Turing instability to occur: the inhibitor must diffuse faster than the activator (\(D_v \gg D_u\)), and the activator must autocatalyse and stimulate the inhibitor, while the inhibitor suppresses the activator.

Turing instability condition: A homogeneous steady state that is stable to spatially uniform perturbations can become unstable to spatially non-uniform perturbations when activator and inhibitor diffuse at sufficiently different rates. This instability spontaneously amplifies spatial inhomogeneities of a characteristic wavelength, leading to stable spatial patterns (spots, stripes) whose scale depends on the ratio of diffusion coefficients and reaction rates.

Turing patterns have been proposed as the mechanism for: pigmentation patterns in animal coats (leopard spots, zebra stripes — Meinhardt, Murray), digit formation in vertebrate limbs, feather/hair follicle spacing, and many other biological spatial patterns. Direct molecular evidence has been found for Turing-type mechanisms in mouse digit patterning (Sheth et al., 2012) and in fish pigmentation (Kondo et al.).

7.3 Positional Information: Wolpert’s Framework

Lewis Wolpert (1969) proposed the concept of positional information: cells acquire a “positional value” from morphogen gradients (the French Flag Model), which they then interpret according to their genetic program to differentiate accordingly. In this view, patterning involves two separable processes:

  1. Setting up positional information: typically through graded morphogen concentrations established by local production, diffusion, and degradation
  2. Interpreting positional information: cells read the local morphogen concentration and activate different sets of target genes depending on whether the concentration exceeds threshold values

The Bicoid gradient in the Drosophila embryo is the paradigm: Bicoid protein is produced from maternally deposited mRNA at the anterior pole and diffuses along the embryo axis, forming a graded profile. Different Bicoid concentrations activate different target genes, defining the anterior-posterior axis.

7.4 L-systems: Lindenmayer’s Formal Grammars

Aristid Lindenmayer introduced L-systems (Lindenmayer systems) in 1968 as a mathematical model of plant development. An L-system is a formal string rewriting system:

L-system — A formal grammar consisting of:
  • An alphabet V: the set of symbols
  • An axiom (start string) \(\omega \in V^+\)
  • A set of production rules \(P\): each symbol a ∈ V is rewritten simultaneously by a production a → w where w ∈ V*
At each derivation step, every symbol in the current string is replaced in parallel by its production. The key feature distinguishing L-systems from Chomsky grammars is parallel rewriting — all symbols are rewritten simultaneously, modelling the simultaneous growth of all parts of an organism.

Example — algae model (Lindenmayer’s original):

  • Alphabet: {A, B}
  • Axiom: A
  • Rules: A → AB, B → A
  • Derivation: A → AB → ABA → ABAAB → ABAABABA → …
  • The lengths are Fibonacci numbers; the system models alternating growth states

Turtle geometry interpretation (Prusinkiewicz & Lindenmayer, The Algorithmic Beauty of Plants, 1990): L-system strings are interpreted as instructions for a turtle moving in 2D or 3D space:

  • F: move forward (draw segment)
  • +/−: turn left/right by angle \(\delta\)
  • &/^: pitch down/up
  • [/]: push/pop position-orientation stack (for branching)

This generates realistic branching structures (trees, ferns, flowers) from compact production rules.

Classes of L-systems:

  • Deterministic, context-free (D0L): simplest; each symbol has exactly one production
  • Stochastic L-systems (SL): productions chosen probabilistically; introduces natural variation
  • Context-sensitive L-systems (kL): productions depend on neighbouring symbols; model neighbour-mediated signalling
  • Parametric L-systems: symbols carry numerical parameters (age, size, nutrient level); model quantitative growth variables
  • Open L-systems: model interaction with environment (light, gravity, soil moisture)

7.5 Computational Development and 3D Morphogenesis

More detailed models of development include:

The virtual plant/virtual creature approach: Karl Sims (1994) evolved virtual creatures — 3D body plans described by a graph grammar, with neural controllers evolved simultaneously by a genetic algorithm. The morphology and physiology co-evolved to produce creatures capable of locomotion, swimming, following light, etc. This demonstrated morphology-controller co-evolution in a physically simulated 3D environment.

Agent-based developmental models: Cells are modelled as agents with internal state (gene expression levels), sensing capabilities, and behavioural rules (migration, division, apoptosis). Examples include: the CompuCell3D platform for multicellular modelling, and various models of Drosophila patterning, vertebrate somitogenesis, and vascular network formation.

Fractal geometry and branching: Many biological structures — bronchial trees, vascular networks, dendritic arbours, coastlines, river networks — exhibit self-similar fractal structure over multiple scales. L-systems naturally generate fractal branching; fractal dimension provides a measure of the space-filling efficiency of these structures.


Chapter 8: Advanced Topics — Major Transitions and Multicellularity

8.1 Major Evolutionary Transitions (Maynard Smith & Szathmáry)

John Maynard Smith and Eörs Szathmáry’s The Major Transitions in Evolution (1995) identified a series of qualitative transitions in the history of life, each involving a change in the way information is stored, transmitted, and expressed:

TransitionBeforeAfter
Origin of self-replicating moleculesChemical replicationTemplated replication
RNA → DNA + proteinRNA worldGenetic code, central dogma
ChromosomesIndependent replicatorsLinked chromosomes
Eukaryotic cellProkaryoteNucleus + organelles (endosymbiosis)
Sexual reproductionAsexual clonesMeiosis, syngamy
MulticellularitySingle cellsDifferentiated tissues
EusocialitySolitary organismsColonies with division of labour
LanguageSignalsCompositional syntax

Each transition involved the integration of previously independent entities into a new higher-level unit of selection: the individual replicator became a chromosome, the chromosome became a cell, the cell became a multicellular organism, the organism became a colony. At each transition, lower-level selection (competition among the integrated units) was suppressed, while higher-level selection (on the new unit) became dominant.

A recurring feature of major transitions is the emergence of division of labour among previously equivalent units. In the RNA world, a single molecule was required to do everything; specialisation allows higher efficiency. The transition to multicellularity involved the differentiation of cells into types specialised for different functions (nutrition, reproduction, defence). The evolutionary challenge is to explain why lower-level units "give up" independent reproduction in favour of cooperation.

8.2 Evolution of Individuality

What constitutes a biological individual? The question is harder than it appears. A beehive has many individual bees but also functions as a unit of selection. A grove of aspen trees may be genetically identical clones connected by a common root system. Physalia (the Portuguese man-o-war) is a colony of genetically identical polyps functionally differentiated into feeding, locomotion, defence, and reproduction.

Evolutionary individuality requires, at minimum, that the unit can be selected upon — that its reproductive success depends on its internal organisation. Buss (1987) and Michod (1999) analysed the transitions to multicellularity in terms of the evolution of individuality:

  • Conflicts between cellular lineages within an organism (e.g., cancer) are analogous to within-group selection
  • Mechanisms suppressing within-organism conflict (germline sequestration, policing mechanisms) are necessary for the emergence of robust individuality
  • The Volvocine algae (from unicellular Chlamydomonas through colonial Volvox) provide a model system for studying the transition to multicellularity and the evolution of individuality

8.3 Evolution of Evolvability

Evolvability — the capacity of a system to generate heritable variation that can be selected — is not fixed but is itself an evolved trait. Features that enhance evolvability include:

  • Modularity: decomposing phenotypes into semi-independent modules reduces the probability that a mutation affecting one function disrupts another. Modular genotype-phenotype maps have more evolvable regions.
  • Robustness/canalisation: Waddington’s canalisation is the property of developmental trajectories being insensitive to genetic and environmental perturbation. Counter-intuitively, robustness can enhance evolvability by allowing neutral variation to accumulate without phenotypic effect, creating reservoirs of cryptic variation that can be released when conditions change.
  • Recombination and sex: facilitates the combination of beneficial mutations from different lineages
  • Regulatory evolution: changes in regulatory networks (when and where genes are expressed) rather than protein-coding sequences (what the proteins do) can generate large phenotypic variation without disrupting conserved biochemical machinery — this is the central insight of evo-devo

8.4 Evo-Devo: Evolutionary Developmental Biology

Evo-devo (evolutionary developmental biology) investigates how changes in developmental processes generate evolutionary change in morphology. Key findings:

Hox genes: Highly conserved transcription factors controlling the anterior-posterior body plan of all bilateral animals. The same Hox genes are present (with minor modifications) in fruit flies, nematodes, mice, and humans; their expression patterns determine which body segments develop into which structures. The conservation of Hox genes across ~600 million years of divergent evolution indicates that evolution preferentially modifies regulatory elements rather than the Hox proteins themselves.

The toolkit: A small set of developmental regulatory genes and signalling pathways (Hox, Pax, Dll/Distal-less, Hedgehog, Wnt, Notch, etc.) is shared across most bilateral animals. The diversity of animal body plans arises largely from variations in the timing, location, and level of expression of these conserved toolkit genes — not from fundamentally different molecular mechanisms.

Regulatory evolution: Carroll (2005) estimated that the majority of evolutionary differences between closely related species arise from changes in non-coding regulatory sequences (enhancers, promoters) rather than coding sequences. A single enhancer of the yellow gene in Drosophila is responsible for wing spot patterns unique to particular species.

8.5 Symbiogenesis

Symbiogenesis — the formation of new organisms through the merger of two previously independent organisms — was proposed by Konstantin Mereschkowski (1905) and championed by Lynn Margulis from the 1960s onward. The endosymbiotic theory of mitochondria and chloroplasts is now firmly established:

  • Mitochondria are derived from alphaproteobacteria, engulfed by a host archaeon ~1.5 billion years ago
  • Chloroplasts are derived from cyanobacteria, engulfed by a eukaryotic cell ~1.2 billion years ago
  • Evidence: double membranes, circular DNA, 70S ribosomes, phylogenetic placement of rRNA sequences, susceptibility to antibiotics targeting bacterial ribosomes

Margulis argued that symbiogenesis is a major evolutionary mechanism, not merely a historical curiosity. The ciliates (which carry bacteria-derived endosymbionts), lichen (fungal-algal symbioses), mycorrhizal networks, and the human microbiome are all products of symbiotic integration at various stages. In the context of ALife and major transitions, symbiogenesis represents a horizontal route to new complexity — merging rather than splitting lineages.

8.6 Differentiated Multicellularity as Computation

A theme central to SYDE 537 is the interpretation of differentiated multicellularity as a computational paradigm. In this view:

  • Individual cells are processing units with internal state (gene expression profile)
  • Cell-cell signalling is inter-processor communication
  • The genome is a distributed program, with different subsets activated in different cells
  • Development is the execution of this distributed program, generating a spatially organised network of differentiated processors
  • The organism as a whole computes integrated responses to environmental inputs

This framework connects to theoretical work on: distributed computation (how do collections of locally communicating finite-state machines compute global functions?), reaction systems, and the computational complexity of developmental processes.

Nehaniv and colleagues have studied interaction machines — formal models of computation emphasising ongoing interaction with the environment, in contrast to classical machines that take an input and produce an output in isolation. Living systems are paradigmatic interaction machines: they do not compute a final answer but engage in continuous, open-ended interaction with their environment.

8.7 Assembly Theory

Assembly theory (Walker, Cronin and colleagues, 2021–2023) proposes a new framework for quantifying biological complexity and detecting life. The assembly index (or molecular assembly index, MA) of a molecule is the minimum number of joining operations required to construct it from monomeric units. The assembly space of a molecule traces all the distinct sub-structures that must be assembled on the pathway to the final molecule.

Key claims:

  • Complex molecules (high MA) are unlikely to arise by chance alone and are therefore indicators of selection or life
  • The joint probability of a molecule existing is \(\prod_i P(\text{step}_i)\), which decreases exponentially with assembly index — providing a quantitative signature of life
  • Molecules with high copy numbers in a sample that also have high MA are strong evidence for biological or evolutionary processes (LUCA-style copying)

Assembly theory has been proposed as a framework for astrobiology (detecting life on other worlds using mass spectrometry) and for understanding open-ended evolution (which generates molecules of continuously increasing assembly index).


Chapter 9: Origins of Life and Exobiology

9.1 The Prebiotic Chemistry Problem

Life arose on Earth from non-living chemistry. The origin-of-life problem asks: what sequence of chemical steps connects ordinary organic chemistry to the first self-replicating, evolving system? This is one of the most difficult open problems in science.

The Urey-Miller experiment (1953): Stanley Miller, following Harold Urey’s suggestion, passed electrical discharges through a mixture of gases thought to represent the early Earth atmosphere (H\(_2\), CH\(_4\), NH\(_3\), H\(_2\)O). The result was a mixture of amino acids and other organic molecules — demonstrating that the building blocks of life could form spontaneously from simple inorganic precursors under abiotically plausible conditions.

Possible environments for life’s origin:

  • Tide pools / warm ponds (Darwin’s “warm little pond”): concentrated by evaporation, heated by UV and thermal energy
  • Deep-sea hydrothermal vents: alkaline (Lost City type) vents provide chemiosmotic gradients potentially usable as energy sources, mineral surfaces for catalysis and concentration
  • Ice: freeze-concentration of organics; lipid membranes form spontaneously at phase boundaries
  • Mineral surfaces (pyrite, clay): adsorption concentrates organic molecules; surfaces catalyse polymerisation reactions; Günter Wächtershäuser’s “iron-sulfur world” hypothesis

9.2 Hypercycles: Eigen’s Model

Manfred Eigen and Peter Schuster (The Hypercycle, 1979) proposed the hypercycle as a solution to the “error catastrophe” — the problem that without sophisticated proofreading machinery, early replicating molecules would accumulate errors faster than selection could remove them.

A hypercycle is a catalytic network in which a set of self-replicating molecules \(I_1, I_2, \ldots, I_n\) form a cycle of mutual catalysis: \(I_1\) catalyses the replication of \(I_2\), \(I_2\) catalyses the replication of \(I_3\), …, \(I_n\) catalyses the replication of \(I_1\). Each molecule can remain short (low error catastrophe) while the network collectively maintains a larger total information content.

\[ I_1 \rightarrow I_2 \rightarrow I_3 \rightarrow \cdots \rightarrow I_n \rightarrow I_1 \]

Hypercycles display cooperative dynamics: if any member falls below a threshold, the whole cycle can crash; parasitic molecules that exploit the cycle’s catalytic products without contributing are a stability problem. Nevertheless, the hypercycle provides a plausible model for the early integration of independently replicating molecules into a cooperative network, a step toward the first cell.

9.3 Origin of the Genetic Code

The genetic code (mapping of codons to amino acids) is nearly universal, suggesting it was fixed early and is highly conserved. How did it originate?

Several hypotheses:

  • Stereochemical hypothesis: direct chemical affinity between certain codons/anticodons and the amino acids they encode. Some experimental evidence for a few amino acids (arginine, isoleucine) but not universal.
  • Coevolution hypothesis: the genetic code co-evolved with biosynthetic pathways; amino acids added later were assigned codons similar to those of biosynthetically related earlier amino acids
  • Frozen accident hypothesis (Crick): once a code became established and used in a large network of molecules, any change would be catastrophic (lethal mutations throughout the proteome), so the code became frozen regardless of whether it is optimal
  • Error minimisation hypothesis: the code is organised to minimise the phenotypic effects of point mutations and mistranslations — which implies the code is not a frozen accident but was shaped by selection for robustness. Statistical analyses confirm that the standard genetic code is remarkably (though not maximally) robust to mistranslation.

9.4 Exobiology and the Possibility of Life Elsewhere

Exobiology (astrobiology) extends the ALife perspective to the question of life in the universe. The key conceptual tool is the NASA working definition: life as a self-sustaining chemical system capable of Darwinian evolution. This definition is deliberately substrate-neutral — it says nothing about carbon, water, or DNA.

The habitable zone: the range of orbital distances from a star at which liquid water could exist on a planetary surface. However, this is a narrow Earth-centric criterion; subsurface oceans (Europa, Enceladus) extend habitability beyond the classical zone.

Candidate environments: Mars (past habitability, possible subsurface life), Europa (liquid ocean under ice, confirmed organic chemistry), Titan (hydrocarbon lakes — possible non-aqueous life chemistry?), Enceladus (active plumes, organic molecules, phosphates detected by Cassini).

The SETI context: If life is a general property of chemistry under appropriate conditions, it should be common in the universe. The Fermi paradox (where is everyone?) and the Drake equation frame the uncertainty. ALife contributes by suggesting that the minimum complexity required for life may be lower than previously thought, and that life’s chemical signatures could be detectable even without direct contact.

Assembly theory and biosignatures: Walker and Cronin argue that assembly index could serve as a substrate-independent biosignature detectable by mass spectrometry on future planetary missions — a direct application of ALife theoretical tools to astrobiology.


Chapter 10: Ethics, Risks, and Responsible Practice

10.1 Why Ethics Matters in Artificial Life

Artificial Life research spans a spectrum from purely theoretical (mathematical models of evolution) to practically realised (synthetic organisms, digital life systems, robots). As the field matures and intersects with synthetic biology, robotics, AI, and nanotechnology, the ethical stakes increase. SYDE 537 treats ethical analysis not as a peripheral afterthought but as an integral component of responsible ALife practice.

The central ethical tensions in ALife can be summarised:

  1. Ontological provocation: ALife challenges traditional distinctions between alive and non-alive, natural and artificial, organic and digital. This destabilises frameworks that assign moral status based on these distinctions.
  2. Novel risks: Creating self-replicating, evolving systems introduces risks not present in conventional software or chemistry — the risk of uncontrolled replication and evolution beyond the laboratory.
  3. Dual-use: Methods and findings useful for beneficial purposes (medicine, agriculture, computation) can be applied to harmful ones (bioweapons, malicious code, surveillance).
  4. Ecological consequences: Releasing engineered or digital life forms into natural systems could have unpredictable ecological consequences.
  5. Moral status of artificial life: If artificial organisms exhibit the properties we associate with life — and especially if they exhibit sentience or suffering — what moral obligations do we have toward them?

10.2 Biosafety and Biosecurity in Synthetic Biology

Biosafety addresses the prevention of unintentional harm from biological research — worker exposure, environmental release, accidental contamination. Standard biosafety levels (BSL-1 to BSL-4) regulate the physical containment required based on pathogen risk.

Biosecurity addresses the prevention of intentional misuse of biological research for harmful purposes. The “dual-use research of concern” (DURC) framework, developed after the 2001 anthrax attacks and the 2002 synthesis of poliovirus, requires institutional review of research that could be used to enhance pathogen transmissibility, virulence, or host range.

Specific concerns in synthetic biology and ALife:

  • Gain-of-function research: deliberately engineering pathogens with enhanced transmissibility or virulence. The 2011 H5N1 ferret-transmission papers (Fouchier, Kawaoka) triggered intense debate
  • De novo synthesis of dangerous pathogens: as gene synthesis becomes cheaper, the risk of synthesising pathogens from published genomes grows
  • Chimeric organisms: combining genetic material from disparate sources in ways that nature would not, potentially creating novel risks

JCVI and synthetic biology: the creation of JCVI-syn3.0 (the “minimal cell”) required institutional review and public engagement. The creation of Mycoplasma mycoides JCVI-syn1.0 (2010) — the first synthetic cell with a purely chemically synthesised genome — was accompanied by an ethics review. The team watermarked the genome with quotations and the names of contributing scientists, raising questions about ownership of synthetic life.

10.3 Digital Life: Risks and Responsibilities

Self-replicating software and digital organisms raise risks analogous to biological risks but in computational substrates:

  • Computer viruses and worms are self-replicating programs — a direct instantiation of one aspect of ALife in deployed computing systems. The Morris Worm (1988), which exploited vulnerabilities to spread across the early internet, infected thousands of machines and was among the first widely publicised demonstrations that self-replicating code could cause large-scale disruption.
  • Uncontrolled evolutionary dynamics: releasing evolving digital organisms into open networks could produce unforeseen behaviours if fitness landscapes change or if the organisms evolve to exploit unanticipated environments
  • Digital ecology: large-scale interactions between AI systems, bots, and evolving software in online environments already exhibit emergent dynamics — price crashes caused by algorithmic trading, flash crashes, echo chambers in recommendation systems

The ALife community has largely avoided deliberately creating dangerous self-replicating software, but the techniques and insights are directly relevant. Responsible practice requires thinking about containment before deployment.

10.4 Ecological Risks of Releasing Artificial Life

Biological invasion: Introduced species frequently cause severe ecological disruption — the cane toad in Australia, kudzu in North America, Caulerpa taxifolia in the Mediterranean. The key property of biological invaders is self-replication: unlike chemical pollutants, introduced organisms can multiply, spread, and evolve.

Synthetic organisms designed for environmental applications (bioremediation, carbon capture, nitrogen fixation) raise analogous concerns:

  • Could engineered organisms outcompete natural organisms?
  • Could synthetic genetic elements (engineered genes, gene drives) spread into natural populations via horizontal gene transfer?
  • Gene drives: CRISPR-based gene drives can in principle spread through an entire wild population in a small number of generations, potentially causing extinction. Applications include eradicating malaria-transmitting mosquitoes and controlling invasive rodents — but the ecological consequences of eliminating entire species are uncertain.

Containment strategies: Genetic firewall (using non-natural amino acids requiring synthetic supplements not available in nature), synthetic auxotrophy (organisms that cannot grow without supplied nutrients), kill switches (genetic circuits that induce cell death if the organism leaves the designed environment). None of these is perfect.

10.5 Intellectual Property and the Ownership of Life

J. Craig Venter’s Institute filed patents on the minimal synthetic genome and on synthetic organisms. This raises questions:

  • Can living organisms (even synthetic ones) be patented?
  • The Diamond v. Chakrabarty (1980) US Supreme Court decision held that genetically modified organisms could be patented, opening the era of biotechnology patents
  • Synthetic organisms that are entirely designed by humans (rather than derived from natural organisms) seem even more clearly patentable — but should they be?
  • Patents on synthetic organisms could restrict scientific research and concentrate biotechnology capabilities in wealthy institutions

10.6 Moral Status of Artificial Life

As ALife research progresses, the question of the moral status of artificial organisms becomes increasingly pressing. Standard criteria for moral status include: sentience (capacity for subjective experience), interests (having states that can be frustrated), autonomy (self-directed behaviour), and sociality (membership in a moral community).

Current digital organisms (Tierra, Avida organisms) are almost certainly not sentient in any meaningful sense. But several considerations complicate this:

  • If sufficiently complex self-organising information-processing systems can give rise to subjective experience in biological brains, the same might be possible in artificial systems
  • We have no reliable method for detecting subjective experience from outside — the “other minds” problem
  • The precautionary principle might argue for erring on the side of attributing moral status rather than withholding it

The ALife community has grappled with these questions since the field’s inception. Langton himself raised them in early papers. The emerging consensus is that responsible ALife practice requires ongoing ethical reflection as capabilities grow.

10.7 Responsible Innovation Frameworks

Several frameworks have been proposed for guiding responsible innovation in ALife and related fields:

Responsible Research and Innovation (RRI): an EU-developed framework emphasising anticipation (foreseeing potential consequences), reflexivity (researchers reflecting on their own assumptions), inclusion (stakeholder participation in research governance), and responsiveness (adapting research in light of emerging concerns).

The Asilomar model: The 1975 Asilomar Conference brought researchers together to develop a voluntary moratorium on risky recombinant DNA experiments until safety guidelines could be established — a model of scientists proactively regulating their own research. The 2015 Napa conference on CRISPR and the 2018 He Jiankui affair (germline gene editing of human embryos) have revived debates about whether such voluntary mechanisms are sufficient.

Dual-use research governance: National Institutes of Health (NIH) and National Science Advisory Board for Biosecurity (NSABB) processes for reviewing DURC; journals’ decisions about publishing potentially dangerous research.

The overarching principle for ALife practitioners: the power to create life-like systems imposes a responsibility to understand and mitigate the risks of doing so. This responsibility extends beyond individual researchers to institutions, funding bodies, and the broader scientific community.


Chapter 11: Selected Advanced Topics

11.1 Complexity and Interaction Machines

Classical computability theory focuses on machines that take an input, perform a computation, and produce an output — the Turing machine paradigm. Peter Wegner, and subsequently Nehaniv and colleagues, have argued that this paradigm is fundamentally inadequate for living systems and other systems that engage in ongoing interaction with their environment.

Interaction machines are computational models that engage in unbounded interaction with an environment — they receive inputs from, and send outputs to, an environment that is itself potentially adaptive and non-static. It has been shown (Wegner, 1997) that interaction machines are computationally more powerful than Turing machines: they can compute functions that no Turing machine can compute.

This is directly relevant to ALife: a living organism is precisely an interaction machine. It does not compute a “final answer” but engages in continuous, open-ended interaction with a complex environment. Understanding organisms as interaction machines places them beyond the scope of classical computability theory and motivates new theoretical frameworks.

Complex adaptive systems (CAS), studied extensively at the Santa Fe Institute, are systems with many interacting agents that adapt individually and collectively to their environment. Features of CAS include:

  • Nonlinearity: outputs are not proportional to inputs
  • Emergence: collective properties not predictable from individual properties
  • Self-organisation: order arising without external design
  • Adaptation: agents modify their behaviour based on experience
  • Edge of chaos dynamics: CAS tend to evolve toward the boundary between order and chaos, where computation is maximised and adaptability is greatest

Holland’s work on CAS, his development of classifier systems (rule-based adaptive agents) and echo (a multi-agent simulation of complex adaptive systems), laid the groundwork for understanding CAS in formal terms.

11.2 Nanomedicine and Biological Nanotechnology

Biological systems are nanotechnology in the most literal sense — molecular machines operating at the nanoscale. ALife research has both inspired and been inspired by nanoscience:

  • The ribosome is a molecular assembler that constructs proteins one amino acid at a time under programmatic control
  • Kinesin, dynein, and myosin are molecular motors that convert chemical energy (ATP hydrolysis) into directed mechanical work
  • DNA hybridisation provides programmable, reversible, sequence-specific binding — the basis of DNA nanotechnology (Seeman, 1982 onward)
  • DNA origami (Rothemund, 2006) enables arbitrary 2D and 3D nanoscale structures to be assembled from a single long scaffold strand and short “staple” strands following a computational design

Nanomedicine applications include:

  • Drug delivery nanoparticles: lipid nanoparticles (as in mRNA COVID vaccines), liposomes, polymeric nanoparticles, DNA cages
  • Molecular diagnostics: aptamer-based sensors, CRISPR-based diagnostics (SHERLOCK, DETECTR)
  • DNA computing for medical decisions: a DNA circuit that logically integrates multiple disease biomarkers and releases a therapeutic payload only when all conditions are met (Shapiro group, Weizmann Institute)

ALife contributes to nanomedicine by providing theoretical frameworks for designing self-assembling, self-replicating, and adaptive nanoscale systems.

11.3 Open-Ended Evolution: Conditions and Challenges

Open-ended evolution (OEE) — the indefinite generation of novelty, complexity, and diversity — is a defining feature of biological evolution but has proven extremely difficult to achieve in artificial systems. Understanding why is among the deepest theoretical challenges in ALife.

Characteristic behaviours of biological OEE:

  • Continuous increase in maximum complexity over evolutionary time (though mean complexity need not increase)
  • Indefinite coexistence of diversity at multiple taxonomic levels
  • Arrival of qualitative innovations (new body plans, new biochemistries, new modes of heredity)
  • No sign of convergence to a fixed-point attractor

Obstacles to OEE in artificial systems:

  1. Bounded search spaces: current digital systems operate in spaces that, while large, are finite and fixed. Biological evolution operates in a continuously expanding possibility space as new molecules and gene products become available.
  2. Fixed instruction sets: digital organisms’ instruction sets do not themselves evolve, limiting the qualitative novelty achievable
  3. Fitness landscape fixity: a fixed fitness function means evolution converges rather than indefinitely explores
  4. Lack of major transitions: digital systems have not, to our knowledge, spontaneously undergone transitions analogous to the origin of sex, multicellularity, or language

Proposed requirements for OEE (Banzhaf et al., 2016; Packard et al., 2019):

  • Unbounded constructive closure: the possibility space must grow as organisms evolve
  • Persistent disequilibrium: a mechanism (co-evolution, environmental change) that prevents stabilisation on a single attractor
  • Evolvability evolution: the capacity to generate variation must itself evolve
  • Transitions in individuality: mechanisms for qualitative changes in the unit of selection

Current best candidates for OEE include co-evolutionary systems (Avida with environmental change, evolutionary robotics with competitive scenarios) and systems with emergent self-organisation of developmental complexity (gene regulatory network models, developmental grammars).

11.4 Complexity Metrics in ALife

Quantifying complexity is a fundamental challenge. Several distinct measures are used:

Kolmogorov (algorithmic) complexity \(K(x)\): the length of the shortest computer program that outputs string \(x\). \(K(x)\) is uncomputable in general but approximated by compression length. A string with low \(K\) is highly regular (e.g., all zeros); a random string has \(K \approx |x|\). Living systems are neither maximally simple nor maximally random — they have intermediate complexity.

Effective complexity (Gell-Mann & Lloyd): the length of a compact description of the regularities of an object. Effective complexity is low for both simple and random objects; high for objects with significant but non-random structure — like organisms.

Logical depth (Bennett): the computation time required to derive an object from its shortest program. Deep objects are those whose regularities took a long time to compute — loosely capturing the “work done” by evolution to arrive at an organism.

Assembly index (Walker/Cronin): as described above — the minimum number of bond-forming steps to assemble a molecule from monomers. Operationally measurable and applicable to biosignature detection.

Statistical complexity (Crutchfield): the minimal amount of information needed to optimally predict the future of a process. Computable from the epsilon-machine (the minimal sufficient statistic of a process’s causal state).

Each measure captures different aspects of complexity; the appropriate choice depends on the application.


Chapter 12: Synthesis and Perspectives

12.1 Recurrent Themes in ALife

Several themes recur across the diverse topics of SYDE 537:

Emergence and self-organisation: Complex global patterns arising from simple local rules — cellular automata, swarm behaviour, reaction-diffusion patterns, collective evolution in digital organisms, major evolutionary transitions. ALife provides both formal models and simulation experiments that demonstrate emergence in controlled conditions.

The edge of chaos: Systems at the boundary between ordered and chaotic dynamics — whether in CAs (Wolfram Class IV, Langton’s lambda), Boolean networks (K near 2), or evolutionary dynamics — appear to exhibit maximal computational capacity, evolvability, and adaptability. This recurring finding suggests a deep connection between criticality and complex function.

Information and computation as substrate-independent: The genetic code, the ribosome, neural computation, and swarm intelligence can all be understood as information-processing systems. ALife insists that the substrate (carbon vs. silicon, wet vs. dry) is less important than the organisation of information processing — a principle with both scientific and philosophical implications.

Evolution as search: Evolutionary dynamics, genetic algorithms, swarm intelligence, and ant colony optimisation are all forms of distributed stochastic search. The relationship between biological evolution and computation is not metaphorical but structural: evolution implements a massively parallel, population-based, adaptive search algorithm on fitness landscapes.

Major transitions as increases in organisational complexity: Each major transition in evolution represents an increase in the hierarchical complexity of biological organisation — new levels of individuality, new modes of inheritance, new forms of cooperation. ALife models these transitions computationally, connecting the theoretical framework of Maynard Smith and Szathmáry to formal models of evolutionary dynamics.

12.2 Open Problems

Despite decades of progress, fundamental questions remain open:

  1. Origin of life: The precise sequence of prebiotic chemical steps leading to the first self-replicating cell remains unknown. Key open questions: How did the genetic code originate? How did the first membrane-enclosed systems arise? How did RNA catalysis reach sufficient efficiency?

  2. Open-ended evolution: Why is biological evolution apparently open-ended while artificial evolution systems are not? What are the necessary and sufficient conditions for OEE?

  3. Consciousness and ALife: Can artificial life systems be conscious? If so, would digital organisms on current hardware be conscious? What ethical obligations would follow?

  4. Predictability of evolution: Is evolutionary innovation predictable in principle (given sufficient knowledge of genetics, fitness landscapes, and environments), or is it fundamentally contingent? This bears on questions of convergent evolution and the reproducibility of evolution’s outcomes.

  5. Life elsewhere: Is Earth life typical or atypical? What fraction of habitable planets have developed life? What does the single example of Earth life tell us about the general properties of life?

  6. The hard problem of ALife: Langton’s programme of “life-as-it-could-be” faces the question: when, if ever, is a simulated or artificial system genuinely alive rather than merely a simulation of life? What would count as proof?

12.3 The Intellectual Legacy of ALife

ALife has contributed to multiple fields:

  • Evolutionary computation: genetic algorithms, genetic programming, evolution strategies, and swarm intelligence are direct transfers from ALife to engineering and operations research
  • Computational biology: ALife modeling frameworks (agent-based models, GRN models, L-systems) are now standard in computational biology and systems biology
  • Robotics: evolutionary robotics, behaviour-based robotics (Brooks), swarm robotics, and developmental robotics all draw on ALife principles
  • Complex systems science: ALife contributed empirically and theoretically to the Santa Fe Institute programme on complex adaptive systems
  • Synthetic biology: ALife’s “synthesis” approach, combined with molecular biology techniques, gave rise to synthetic biology — the design of novel biological systems from standardised parts
  • Philosophy of biology: ALife has enriched debates on the definition of life, the nature of biological functions, the meaning of information in biology, and the relationship between mechanism and emergence

ALife exemplifies the power of the computational perspective in science: treating physical systems as information-processing systems, and using computation both to model existing systems and to generate new ones, reveals patterns invisible to purely analytical or experimental approaches. For engineers, ALife provides a toolkit of design principles drawn from 4 billion years of evolutionary R&D — arguably the most thorough design process available.

A final reflection: the boundary between "simulation of life" and "genuine life" is precisely what ALife puts under pressure. If a digital organism in Avida evolves in response to selection, accumulates adaptations, and exhibits all the properties associated with life except carbon-based chemistry — is it alive? Langton's answer would be: the question reveals an arbitrary restriction. Life is a process, not a substance. The pattern matters; the substrate is incidental. Whether this position is ultimately defensible remains one of the deepest open questions in the field — and one that makes ALife not merely an interesting engineering tool but a profound philosophical provocation.
Back to top