·Mathematics & Probability
Section 1
The Core Idea
A casino offers you a game. On each round, a fair coin is flipped. Heads: your wealth increases by 50%. Tails: your wealth decreases by 40%. The expected value per round is positive — the arithmetic average of +50% and −40% is +5%. A statistician would tell you to play. An economist trained in expected utility theory would tell you to play. The game has a positive expected value, and rational agents maximise expected value.
Play this game a hundred times and you will be virtually bankrupt.
The arithmetic average of the outcomes across a large population of players at a single point in time — the ensemble average — is indeed positive. If a thousand people each play one round, the average wealth of the group increases by 5%. But the experience of any single player across many rounds is governed by a different quantity: the geometric growth rate, which accounts for the multiplicative compounding of sequential outcomes. The geometric rate of this game is the square root of (1.5 × 0.6) minus one — approximately −5.1% per round. Each round, on average, a single player's wealth shrinks by 5.1%. The population gets richer in aggregate while almost every individual within it gets poorer over time.
This divergence between what happens on average across a population and what happens to a single participant over time is the ergodicity problem. A system is ergodic when the time average of a single participant converges to the ensemble average across all participants. A system is non-ergodic when these two averages diverge — when the statistical expectation computed across a population at one moment fails to describe the lived experience of any individual within that population across time.
The concept originates in statistical mechanics. Ludwig Boltzmann introduced the ergodic hypothesis in the 1870s to describe systems of gas molecules where the time-averaged behaviour of a single molecule equals the space-averaged behaviour of all molecules at an instant. For ideal gases in thermal equilibrium, the hypothesis holds — the molecule that bounces around the container long enough visits every accessible state with a frequency proportional to that state's equilibrium probability. The system has no absorbing barriers — no state from which the molecule cannot escape.
Financial markets, career trajectories, evolutionary fitness, and nearly every system involving irreversible consequences and multiplicative dynamics are structurally different. They contain absorbing barriers — bankruptcy, death, permanent exclusion — states from which the participant cannot return. The presence of absorbing barriers is what makes the system non-ergodic. The molecule in the gas container cannot be "eliminated." The trader with a finite bankroll can. And that single structural difference invalidates the entire apparatus of ensemble-average reasoning when applied to individuals facing sequential, multiplicative risk.
The physicist Ole Peters identified this as the central error in three centuries of economic theory. Expected utility theory, formulated by Daniel Bernoulli in 1738 and formalised by von Neumann and Morgenstern in 1944, evaluates gambles by computing the weighted average of outcomes across possible states of the world — the ensemble average. This is the correct calculation for an entity that experiences all possible outcomes simultaneously. It is the wrong calculation for an entity that experiences outcomes sequentially, one at a time, where each outcome changes the capital base from which the next outcome is experienced.
The distinction matters because of multiplicative dynamics. In additive systems — where gains and losses are fixed dollar amounts independent of current wealth — the ensemble average and the time average converge. Winning $50 or losing $40 on each round, regardless of current wealth, produces a time average that matches the expected value. But in multiplicative systems — where gains and losses are proportional to current wealth — a single catastrophic loss compounds through every subsequent round. A 50% loss requires a 100% gain to recover. A 90% loss requires a 900% gain. The mathematics of sequential proportional losses creates a ratchet effect: the path to ruin runs downhill and accelerates, while the path to recovery runs uphill and decelerates.
This is not an academic distinction. It is the structural explanation for why intelligent, informed people go broke. The gambler who sizes bets according to expected value — maximising the arithmetic average of outcomes — systematically overbets relative to their ability to survive a sequence of losses. The investor who evaluates opportunities by their expected return, without accounting for the path-dependent reality that a 50% drawdown eliminates them from the game before the recovery can materialise, is confusing the ensemble average for their individual trajectory. The entrepreneur who takes a series of "positive expected value" risks, each of which has a meaningful probability of total ruin, is playing a game where the population-level statistics promise wealth while the individual-level mathematics guarantee bankruptcy.
The corrective is not to avoid risk. It is to recognise which systems are ergodic and which are not — and to size exposure so that you remain a participant in the system long enough for the long-run statistics to become relevant to your personal experience. The Kelly criterion, developed independently by John Kelly at Bell Labs in 1956 and applied by
Ed Thorp in both casinos and financial markets, provides the mathematical solution: the optimal bet size in a non-ergodic system is the fraction that maximises the expected geometric growth rate — not the expected arithmetic return. The Kelly fraction is always smaller, often dramatically smaller, than the fraction that maximises expected value. The difference is the margin of survival.
The 2008 financial crisis provided the most expensive real-world demonstration. Long-Term Capital Management had already collapsed in 1998 under the weight of leverage applied to ensemble-average calculations that assumed convergence would occur before capital ran out. A decade later, Bear Stearns and Lehman Brothers repeated the structural error at institutional scale — holding portfolios of mortgage-backed securities whose expected returns were positive in the ensemble (the average mortgage pays) while their time-series risk was catastrophic (a correlated default wave eliminates the holder before the average can materialise). The executives who approved the leverage ratios were computing expected values. The mathematics that destroyed their firms was the mathematics of sequential, multiplicative, non-ergodic loss applied to a capital base that could not absorb the path.
Warren Buffett has expressed the same insight without the mathematics for sixty years. "Rule number one: never lose money. Rule number two: never forget rule number one." The advice sounds tautological until you understand it through the lens of ergodicity. Buffett is not saying that losses never occur. He is saying that the first priority in any non-ergodic system is to ensure that no single loss — or correlated sequence of losses — removes you from the game permanently. The arithmetic of recovery is too punishing, and the compounding of survival is too powerful, to treat ruin risk as one variable among many. It is the variable. Everything else is a footnote to the question of whether you will still be playing when the favourable outcomes eventually arrive.
The concept extends beyond finance into any domain where participation is a precondition for benefit. An organism that takes repeated survival-threatening risks, each with a "positive expected fitness" in the ensemble, will be eliminated from the gene pool — evolution is a non-ergodic process where the dead cannot benefit from future reproductive opportunities. A reputation that is destroyed by a single catastrophic error cannot recover through subsequent excellent performance — career trajectories are non-ergodic when a single failure mode produces permanent exclusion. The ergodicity question is not "is this a good bet on average?" It is "will I survive to collect the average?"