·General Thinking & Meta-Models
Section 1
The Core Idea
In 1747, a Scottish naval surgeon named James Lind had a problem. Scurvy was killing more British sailors than enemy action — roughly 1,400 men had died on George Anson's circumnavigation just five years earlier, most from the disease. Dozens of proposed remedies circulated: vinegar, seawater, sulfuric acid, cider, barley water, fresh air. Every ship's surgeon had a favourite. None of them had evidence.
Lind did something unprecedented. He selected twelve sailors with scurvy at similar stages of the disease, divided them into six pairs, and gave each pair a different treatment while keeping all other conditions constant. The pair receiving oranges and lemons recovered within a week. The other ten continued to deteriorate. It was the first controlled clinical trial in recorded history — and a demonstration that systematic testing could resolve questions that centuries of argument, authority, and tradition had failed to settle.
The scientific method is the formalisation of what Lind did intuitively: observe a phenomenon, form a testable explanation, design an experiment that could prove that explanation wrong, run the experiment, analyse the results, and revise the explanation accordingly. Then repeat. The cycle has no terminal point. Every conclusion is provisional — the best explanation the evidence currently supports, held until better evidence arrives.
The intellectual architecture took centuries to assemble. Francis Bacon laid the groundwork in Novum Organum (1620), arguing that knowledge should be built from systematic observation and induction rather than from Aristotelian deduction and received authority. Galileo demonstrated the principle physically, dropping objects from the Leaning Tower of Pisa (probably apocryphal) and rolling balls down inclined planes (definitely real) to show that heavier objects don't fall faster than lighter ones — contradicting Aristotle's claim, which had gone untested for nearly two millennia because nobody had thought to check.
Isaac Newton codified the approach in the
Principia (1687), deriving universal gravitation from Kepler's observational data and his own laws of motion. Antoine Lavoisier applied rigorous measurement to chemistry in the 1770s, disproving the phlogiston theory by showing that combustion required oxygen — and demonstrating that careful weighing could settle what philosophical argument could not. Each contribution added a structural element: Bacon contributed systematic induction, Galileo contributed controlled experimentation, Newton contributed mathematical formalisation, Lavoisier contributed quantitative precision.
Karl Popper sharpened the entire framework in The Logic of Scientific Discovery (1934, English translation 1959) with a single, devastating criterion: a theory is scientific only if it is falsifiable — if there exists an observation that could, in principle, prove it wrong. Popper's insight reorganised the hierarchy of intellectual merit. A theory that explains everything predicts nothing. A theory that makes specific, falsifiable predictions and survives repeated attempts to disprove it earns provisional trust proportional to the severity of the tests it has passed. The more opportunities a hypothesis has had to fail, and hasn't, the more seriously you should take it.
The method's history is also a history of institutional resistance to its conclusions. In 1847, the Hungarian physician Ignaz Semmelweis observed that maternity wards staffed by doctors had mortality rates five times higher than those staffed by midwives. His hypothesis: doctors were carrying "cadaverous particles" from autopsy rooms to delivery wards. He instituted a handwashing policy with chlorinated lime solution. Mortality dropped from roughly 10% to under 2% in months. The medical establishment rejected his findings for two decades — not because the data was weak, but because the implication (that doctors were killing their own patients) was socially intolerable. Semmelweis died in an asylum in 1865, thirty years before germ theory vindicated him completely. The data was never the problem. The method produced the right answer. The institution couldn't accept it.
This pattern — method produces evidence, institution resists conclusion — recurs with sufficient regularity that it should be treated as a structural feature, not an anomaly. The method is a machine for generating truths that are sometimes inconvenient. Its value is proportional to the degree of inconvenience it can withstand.
The non-obvious insight: the method's power doesn't come from generating correct answers. It comes from systematically eliminating incorrect ones. Every failed experiment narrows the space of possibilities. The process is subtractive, not additive — a disciplined pruning of the hypothesis tree through contact with evidence. Most people misunderstand this. They think science is about proving things right. It is about finding out what survives being proven wrong.
This subtractive logic is what makes the method transferable to business, investing, and decision-making. A founder running A/B tests, an investor stress-testing a thesis against disconfirming data, a manager running a controlled pilot before rolling out a new process — each is applying the same structural discipline: state a falsifiable hypothesis, design a test that could kill it, run the test honestly, and update.
The uncomfortable corollary: most organisations and most people don't actually do this. They form conclusions, then seek confirming evidence. They design experiments that cannot fail. They interpret ambiguous results as validation. The scientific method isn't hard to understand. It's hard to practise, because every step of the cycle creates opportunities for self-deception — and the human brain is exquisitely designed to take them.
Consider how rarely the method appears in corporate strategy. A company deciding to enter a new market typically builds a business case — a narrative that justifies the decision already made. What they almost never do is state the specific conditions under which they would abandon the initiative, measure against those conditions on a fixed cadence, and actually pull the plug when the kill criteria are met. That process — which is nothing more than the scientific method applied to capital allocation — would prevent the majority of failed expansions, unsuccessful product launches, and value-destroying acquisitions that characterise large-company strategy. The method is available. The organisational will to use it honestly is not.