·Mathematics & Probability
Section 1
The Core Idea
Reference class forecasting answers a single question: when projects like this one were attempted, what actually happened? Instead of building a bottom-up estimate from tasks and assumptions — the inside view — you find a class of similar past projects and use the distribution of their outcomes as your forecast. The method was developed by Daniel Kahneman and Amos Tversky and operationalised by Bent Flyvbjerg for megaprojects. The core claim: the inside view is systematically optimistic; the outside view corrects by anchoring to base rates.
A software team estimates a new platform will ship in six months. They list features, assign days per feature, add buffer. That is the inside view. Reference class forecasting asks: what fraction of similar platform projects at similar companies shipped within 2× their initial estimate? The answer is usually a majority. The forecast should therefore be twelve months or the distribution of that reference class, not six. The same logic applies to cost. Infrastructure projects, product launches, and M&A integrations have documented track records. Those track records are the reference class.
Flyvbjerg's work on transport and construction showed that cost overruns and schedule slips are the norm, not the exception. Reference class forecasting does not eliminate overruns; it produces forecasts that reflect how often overruns occur. The decision-maker gets a distribution: 50th percentile, 80th percentile, worst case. Planning then uses the percentile that matches the organisation's risk tolerance. Governments and enterprises that adopt reference class forecasting often set budgets at the 80th percentile of the reference class. Projects still overrun, but less often than when forecasts are anchored to the inside view.
The practical requirement is a well-defined reference class. "Similar" must be specified: same sector, same type of project, same scale band, same technology maturity. A first-of-its-kind project has a thin or contested reference class; the forecast carries more uncertainty. The discipline is explicit: name the reference class, cite the data, report the distribution. When no reference class exists, the forecast should say so. That transparency is the point — not false precision, but an honest outside-view anchor.