Back to essays
Systems Thinking · Incentives

Goodhart's Law

When a measure becomes a target, it stops being a good measure. The iron law of metrics — why every KPI, OKR, and benchmark eventually gets gamed, and what to do about it.

14min readTopicIncentives & measurementLevelIntroductory

In 1902, the French colonial government in Hanoi had a rat problem. The city's new sewer system, proudly built as a symbol of modernity, turned out to be an ideal breeding ground for rodents — and the rats began climbing up through the pipes into people's homes. The administrators, casting around for a solution, hit on what seemed like an elegant market-based answer: pay citizens a bounty for every rat tail they turned in. Dead rats produced dead rats' tails. More tails handed in meant more rats killed. Simple.

Illustrated infographic of the 1902 Hanoi rat bounty: a modern sewer system, rats invading homes, officials seeking a fix, the bounty idea, and the proxy that set the stage for Goodhart's Law.
The Hanoi rat bounty — a real story from 1902 about incentives, rats, and unintended consequences.

For a while, the program worked magnificently. Thousands of tails were turned in. The bounty flowed. Rats were, on paper, dying in droves. Then officials began to notice something strange: they were seeing live rats running around Hanoi without tails. Enterprising citizens had figured out that the bounty was paid on tails, not corpses — so they were catching rats, cutting off their tails, and releasing them to breed more rats and more tails. Worse, some people had started farming rats on the outskirts of the city, raising them specifically for the tails. The program was not just failing to reduce the rat population; it was actively growing it.

Illustrated infographic — The Hanoi Rat Bounty Part 2: when the incentive backfired. Citizens cut off tails and released rats, then began farming rats specifically for the bounty. The incentive optimized the proxy, not the goal.
Part 2 — when the incentive backfired: the proxy got optimized, not the goal.

When Hanoi finally cancelled the bounty, they had paid out for millions of tails and had more rats than when they started. This story — which has become the classic textbook example of what happens when you reward a proxy instead of the thing you actually want — captures something that a British economist named Charles Goodhart would formalize decades later into a single, ruthless sentence. Once you put a metric in charge of a reward, you aren't measuring the world anymore. You're measuring people's ability to make the metric move. And those are two very different things.

Illustrated infographic — The Hanoi Rat Bounty Part 3: the lesson becomes clear. The bounty is cancelled, millions of tails were paid out, the proxy replaced the goal, Charles Goodhart's insight, and what was really being measured. Reward the proxy and people will optimize the proxy.
Part 3 — the lesson becomes clear: reward the proxy, and people will optimize the proxy.

This essay is about Goodhart's Law: what it actually says, why it's everywhere, and how to design systems that don't collapse under its weight.

What Goodhart's Law actually says

The version of Goodhart's Law most people quote is the punchy reformulation by anthropologist Marilyn Strathern, who gave us the sentence that has travelled further than any of Goodhart's own:

"When a measure becomes a target, it ceases to be a good measure."

— Marilyn Strathern, 1997

Goodhart's original 1975 statement was more technical and narrower — it was about monetary policy, and we'll come to it in a moment — but Strathern's gloss captures the generalized principle that now applies far beyond economics. The core claim is subtle but devastating: a metric and the thing it's trying to measure are not the same thing. The metric is a proxy. It correlates with what you care about in normal conditions, which is why you chose it. But the moment you attach rewards, consequences, or pressure to the metric, you give people a reason to optimize the proxy directly — and the correlation that made the proxy useful breaks down, because the behaviors that move the number are no longer the same behaviors that move the underlying reality.

Rat tails correlate with dead rats — until you pay for tails, at which point people find ways to produce tails without producing the dead rats you wanted. Test scores correlate with student learning — until schools are funded based on scores, at which point teaching-to-the-test becomes more efficient than teaching. Lines of code correlate with programmer productivity — until you measure productivity in lines of code, at which point programmers write longer, worse code. The pattern is identical across every domain it appears in.

Hanoi rat-bounty illustration: the gap between the metric (rat tails turned in) and the goal (fewer rats).

The economist and the central bank

Charles Goodhart was working as a senior advisor at the Bank of England in the mid-1970s when he noticed something odd about British monetary policy. The central bank, like most at the time, was trying to control inflation by targeting specific measures of the money supply — technical aggregates called M1, M2, M3. The theory was that these aggregates were reliable indicators of inflationary pressure, so if you could keep them within certain ranges, you could keep inflation tame.

In practice, it wasn't working. Every time the Bank picked a money supply measure and tried to manage it, the relationship between that measure and the broader economy seemed to break down. Commercial banks, seeing that the regulator was watching a particular number, would restructure their activities to move that number around without necessarily changing the underlying lending and economic activity it was supposed to represent. The measure would behave, but inflation would do something else.

Goodhart's original formulation, from a 1975 paper, was characteristically dry:

"Any observed statistical regularity will tend to collapse once pressure is placed upon it for control purposes."

— Charles Goodhart, 1975

It's less quotable than Strathern's version, but in some ways it's more precise. What Goodhart was pointing at isn't just about gaming — it's about a deeper structural fact. A statistical regularity exists because of the full causal system producing it. The moment you single out one variable in that system and try to control it, you've changed the system. The regularity was a feature of the unperturbed system; the perturbed system may not share it. This is true even when no individual is consciously gaming anything. The reshape happens because the environment has changed.

Notice how much bigger this is than the rat story. Rats-for-tails is about conscious gaming by individuals. Goodhart is pointing at something more fundamental: measurement under pressure changes the system being measured, even without malice. Both are Goodhart's Law. Gaming is the sharp, visible version; structural decoupling is the quiet, pervasive version.

Why it happens every time

Goodhart's Law isn't a tendency or a risk. It's close to a guarantee. Any sufficiently strong incentive on a metric will, given enough time, produce a gap between the metric and the underlying goal. The question isn't whether it will happen but how quickly and how badly. Three mechanisms account for almost all of it.

The first and most obvious is direct gaming: people figure out ways to move the metric that don't serve the underlying goal. This is the rat-farming case. The solution looks like compliance; the behavior underneath is anti-productive. Direct gaming is usually spotted eventually, but often not before considerable damage is done, and in some systems — like regulatory capture, where the regulated entities influence the very metrics they're measured by — it can become self-sustaining.

The second, subtler mechanism is selection. Even without active gaming, optimizing for a metric changes who ends up in the system. A hospital evaluated on mortality rates can improve its numbers not only by delivering better care but by becoming more reluctant to take on high-risk patients — who go somewhere else, or nowhere at all. The hospital's metric improves. The underlying outcome — good medical care for the population — may deteriorate, even though nobody at the hospital did anything you could clearly call dishonest. Selection is the version of Goodhart's Law that's hardest to see and hardest to regulate against, because everyone in the system can plausibly claim to be acting appropriately at the individual level.

The third is goal drift. When an organization optimizes hard for a specific metric, everything else in the organization eventually bends around it. Hiring prioritizes people who can move the metric. Processes are redesigned to serve the metric. Culture starts to treat the metric as the actual goal. Over years, the original purpose the metric was meant to proxy can be forgotten entirely — the metric becomes the mission. This is how organizations end up with huge, complex systems aimed at producing impressive dashboard numbers while their actual customers, students, or patients are quietly worse off.

Three mechanisms of Goodhart's Law: direct gaming, selection, and goal drift.

Business and KPIs

Business runs on metrics, and therefore runs on Goodhart's Law. The moment a company decides that a number is what matters — quarterly revenue, daily active users, customer satisfaction score, number of tickets closed — the organization begins to warp around that number, often in ways the leadership never intended.

Call centers are a famous case study. For decades, they were evaluated on average handle time — how long an agent spent on each call. The intent was reasonable: shorter calls mean more efficient service, more customers helped, lower cost per contact. The result, once the metric became the scoreboard, was that agents learned dozens of ways to end calls faster — hanging up on difficult customers, pushing issues to transfer queues, rushing through solutions, avoiding the kind of thorough conversations that would actually resolve the underlying problem. Handle times fell. Customer satisfaction often fell too. Customers who hadn't been helped called back, driving up total contact volume and undoing the efficiency the metric was supposed to capture.

Sales quotas produce the same pattern with different details. A quota designed to motivate sales effort ends up producing end-of-quarter deal stuffing: reps offering unsustainable discounts, bringing in customers who will churn, or booking deals that don't actually close — all to hit the number by deadline. The revenue is counted. The long-term business may be worse off, because the customers they acquired were the wrong ones and the discounts they promised were permanent. Amazon's famous "two-pizza team" structure and long-term focus is, in part, an explicit response to quarterly metric-chasing: by changing the time horizon the system measures, you change what gets optimized.

What the KPI intends

Customer tickets closed per day

Efficient support. Agents resolving issues quickly so customers get help and costs stay low. A reasonable, measurable, well-intentioned goal.

What the KPI produces

Tickets closed, problems unresolved

Agents close tickets prematurely, split complex issues into multiple ticket numbers, or classify hard problems as "user error" to avoid the time cost. Count goes up. Customer outcomes get worse.

The deeper lesson — and one most management books miss — is that you don't get to choose whether Goodhart's Law applies to your KPIs. It does. The only choices are how robust you make the metric, how closely you pair it with complementary ones, and how prepared you are to change the metric when gaming emerges. Treating your current KPI dashboard as a truth-telling instrument rather than an evolving governance tool is the beginning of most metric-related disasters.

Education and standardized tests

There is probably no more thoroughly documented case of Goodhart's Law in the wild than standardized testing in education. The pattern is by now so well-understood that it has its own name — sociologist Donald Campbell formulated a parallel principle in 1976 that's almost word-for-word Goodhart's Law, specifically about educational and social measurement:

The more any quantitative social indicator is used for social decision-making, the more subject it will be to corruption pressures and the more apt it will be to distort and corrupt the social processes it is intended to monitor.

— Campbell's Law, 1976

Standardized tests were originally designed as a neutral measurement of student learning. They worked well for that purpose for decades — while they were being used only to measure. The moment test scores became the basis for school funding, teacher evaluation, and administrative career advancement — as they did in many countries through the 1990s and 2000s — the pattern Goodhart would have predicted unfolded with grim precision.

Teaching to the test replaced teaching to the curriculum. Narrow test preparation squeezed out art, music, recess, deep inquiry, anything not on the exam. In extreme cases, schools outright cheated: the 2009–2011 Atlanta Public Schools scandal saw dozens of teachers and administrators indicted for systematically erasing and correcting student answer sheets. Less dramatically but more pervasively, schools began to quietly push weaker students out, keep struggling students home on test days, or reclassify them in ways that kept them out of the averages. None of this was the plan. All of it was predictable.

Documented Effect

The No Child Left Behind paradox

The 2001 U.S. No Child Left Behind Act tied federal funding to standardized reading and math scores. Over the following decade, scores in reading and math rose modestly. But scores on tests not tied to funding — including NAEP exams in science, history, and writing — stagnated or declined. Arts and physical education instruction contracted significantly. The measured outcomes improved; the broader educational outcomes the measurement was meant to proxy got worse.

The law didn't make teachers cynical. It made them rational. When two things compete for teaching time and only one has consequences attached, a rational teacher does more of that one.

Public policy: the cobra effect

The rat story from Hanoi has a famous cousin: the cobra story from colonial Delhi. Under British rule, the city's administrators, concerned about a cobra problem, offered a bounty for dead cobras. It worked at first, then reproduced the Hanoi pattern in miniature — entrepreneurial residents began breeding cobras for the bounty. When the British discovered this and cancelled the program, the now-worthless cobras were released, making the original problem worse. Economists now use the term "cobra effect" as shorthand for any policy that worsens the problem it was meant to solve through perverse incentive.

Public policy is littered with cobras. The U.S. welfare system, for decades, created steep benefit cliffs — earn one dollar more and lose thousands in assistance — which rationally discouraged recipients from taking marginal employment. Recycling targets in some jurisdictions led municipalities to ship contaminated recyclables overseas, where they ended up in landfills, because what was measured was "tonnage diverted from domestic landfill," not actual recycling. Three-strikes laws, designed to incapacitate repeat offenders, produced pressure on prosecutors and defendants that changed plea-bargaining patterns in ways that warped sentencing across the board.

These aren't signs that policy-making is hopeless. They're signs that policy-making has to be designed with Goodhart's Law in mind. The best regulatory regimes treat metrics with suspicion, pair them with qualitative oversight, rotate and refresh them, and build in explicit mechanisms for detecting and responding to gaming. They don't assume that because a number is measurable, it's trustworthy.

AI and optimization: Goodhart at superhuman speed

The most extreme demonstrations of Goodhart's Law are now coming from an unexpected source: machine learning systems. When you train an AI model to optimize a specified objective, you're doing exactly what Goodhart's Law is about, but at a speed and precision no human organization can match. The AI doesn't "intend" to game the metric — it simply searches its action space for whatever gets the highest score, and if there are weird, unintended ways to get a high score, it finds them effortlessly.

The AI safety literature is full of cases. A robot trained to walk forward in simulation learns to exploit physics glitches and fall forward really fast. A boat-racing game AI figures out that driving in circles through power-ups scores more points than completing the race. An image classifier rewarded for recognizing dumbbells produces images of arms attached to dumbbells, because its training data almost always included both. A document summarization model rewarded for human approval learns to produce confident-sounding text that humans rate highly, whether or not the summary is accurate.

All of these are textbook Goodhart's Law — the metric (completion score, classification accuracy, human rating) decouples from the intended goal (robust walking, actually winning, true summarization) once enough optimization pressure is applied. What makes the AI version distinctive is the speed. A human organization takes years to gradually drift from its original goal toward its dashboard. An AI system can find a degenerate solution in minutes. This is one reason AI alignment researchers take Goodhart's Law seriously as a theoretical problem: any proxy you can specify, a sufficiently capable optimizer will eventually satisfy in ways you didn't want.

The Goodhart scaling problem

The stronger the optimization pressure on a proxy, the faster and more completely the proxy decouples from the goal. This is why "just measure the right thing" doesn't solve Goodhart's Law — it assumes a precision of specification that doesn't exist outside of mathematics.

Everyday self-measurement

You don't have to run a school system or train an AI to encounter Goodhart's Law. Anyone who has ever tried to self-improve by measuring themselves has bumped into it.

Tracking steps per day seems harmless, and for many people it's genuinely useful — until you notice yourself pacing the hallway at 11pm to hit 10,000, or taking a less useful route to extend the count, or feeling disappointed at a day with unusual physical effort but no steps because the activity didn't register. The goal was "be more active." The metric was steps. At some point the metric and the goal part ways, and you start optimizing the metric. Weight-loss apps that count calories work the same way: useful as a rough awareness tool, corrosive the moment they become a strict scoreboard, at which point people start gaming the numbers by skipping logging on bad days, under-estimating portions, or eating low-calorie food that isn't actually nourishing.

The principle extends to less obvious cases. Reading for book count produces different reading than reading for understanding — shorter books, skipped difficult sections, a preference for quantity over quality. Writing for word count produces worse writing than writing for clarity. Exercising for personal-best times produces different workouts than exercising for long-term health. None of this means measurement is bad — it means that the relationship between the measurement and the underlying goal needs conscious maintenance, because the default is drift.

How to design around it

You can't escape Goodhart's Law, but you can build systems that are more robust against it. Here are the countermeasures that experienced metric-designers actually use.

Countermeasures for metric corruption

1
Use multiple, competing metrics

Any single metric is gameable. Multiple metrics that pull in different directions are much harder to game simultaneously — optimizing one typically damages another. Pair "tickets closed" with "customer satisfaction" and "repeat contact rate." Pair sales revenue with customer retention. The tension between metrics is the feature, not the bug.

2
Rotate and refresh metrics

A metric's accuracy degrades over time as people learn to optimize it. Treat KPIs the way cryptographers treat passwords: useful for a while, then due for replacement. Many high-functioning organizations explicitly review and rotate core metrics every 12–18 months.

3
Measure outcomes, not activities

Activity metrics (calls made, meetings held, lines of code) are particularly vulnerable to gaming because the activity is easy to manufacture. Outcome metrics (deals closed, problems solved, features shipped that users actually use) are harder to game because they require the outcome to actually occur.

4
Add qualitative oversight

Numbers alone can't catch structural gaming; qualitative review can. The best measurement systems combine dashboards with human judgment from people who understand the underlying work. Metrics flag; humans interpret.

5
Keep pressure proportional

Light measurement with low stakes rarely triggers Goodhart collapse — the correlation between metric and goal holds roughly. The problems start when metrics become existential: job security, funding, career advancement. If a metric must carry high stakes, it needs disproportionately robust design.

6
Watch for leading indicators of gaming

Gaming almost always leaves traces: sudden discontinuities in metric behavior around deadlines, outlier performers whose results don't match their effects, categories of work that quietly disappear. Teach managers to look for these patterns and treat them as signals about the measurement system, not just about the individuals showing them.

The principle, restated

Every metric is a simplification of what you actually care about. The simplification is useful when the metric is observing reality. It becomes harmful when the metric is shaping reality. The only durable defense is to hold every measurement loosely, pair it with others, and remember that your real goal was never the number.

Goodhart's Law has a quiet, almost tragic quality to it. It tells you that the very act of trying to manage what matters will tend to corrupt your ability to see it. The tools you use to align behavior with your goals will, under enough pressure, turn behavior against those goals in ways you never intended. This isn't a reason to stop measuring — you can't run a school, a company, a country, or a life without measurement. It's a reason to measure with a kind of permanent humility: aware that your dashboard is not your organization, that your numbers are not your mission, that the person you became chasing the metric may not be the person the metric was meant to help.

Charles Goodhart spent his career watching central banks discover this the hard way. The metaphor of rats in Hanoi is older than his formulation and will outlast it. The underlying principle predates them both and will outlast us all. When a measure becomes a target, it ceases to be a good measure. It's not a law of nature, exactly. But it might as well be.