← back to essays/

Goodhart's Law

2026-03-25

In 2016, YouTube changed its recommendation algorithm to optimize for watch time instead of clicks. The logic was sound: watch time seemed like a better proxy for "did the user actually enjoy this?" than whether they clicked on a thumbnail.

Watch time went up. Way up. But something else happened: the algorithm learned that conspiracy theories, outrage, and "rabbit hole" content kept people watching longer than anything else. Users weren't enjoying the experience more. Many reported feeling worse. But they couldn't stop watching. The metric improved. The thing the metric was supposed to represent got worse.

This pattern has a name: Goodhart's Law.

"When a measure becomes a target, it ceases to be a good measure."

Named after economist Charles Goodhart, who observed in 1975 that any statistical regularity tends to collapse once pressure is placed on it for control purposes. Marilyn Strathern later generalized it to the punchier version above.

How It Works

Every organization needs to measure things. You can't manage what you can't measure, as the saying goes. So you find a metric that correlates with the outcome you actually care about, and you optimize for it.

The problem: correlation is not identity. The metric is a proxy for the real goal, and there's always a gap between the two. When you push hard enough on the proxy, the system finds ways to improve the metric without improving the thing you care about, or even while making it worse.

The optimizer doesn't need to be malicious. It doesn't even need to be a person. Algorithms, incentive structures, and natural selection all do this. The system mechanically exploits the gap between the proxy and the true goal.

The Simulation

Pick a scenario below, then watch what happens as a system optimizes for a proxy metric over time. The two lines, the proxy and the true goal, start moving together and then diverge. The harder you optimize, the faster and wider the gap.

Social Media Platform

True goal: user satisfaction
Proxy: engagement (time on site)
Gaming: outrage and anxiety keep people scrolling

Software Team

True goal: code quality
Proxy: lines of code written
Gaming: verbose, redundant code; copy-paste over abstraction

Education

True goal: deep learning
Proxy: standardized test scores
Gaming: teaching to the test; memorization over understanding

Customer Support

True goal: customer satisfaction
Proxy: average handle time
Gaming: rushing calls; transferring instead of solving

Hospital Performance

True goal: patient health outcomes
Proxy: mortality rate
Gaming: refusing high-risk patients to keep numbers low

Policing

True goal: community safety
Proxy: arrest numbers
Gaming: targeting easy, minor offenses; ignoring complex cases

5/10
85%
5/10
Proxy metric True goal Gap (divergence)
Choose a scenario and click "Run simulation."

Explore the Dynamics

Try adjusting the three sliders and re-running:

Run It Many Times

One run might look like a smooth divergence. But in practice, systems are noisy. Run the simulation many times below and see the distribution of outcomes. Even with the same settings, the divergence point and final gap vary, which makes it even harder to detect in the real world.

Proxy runs True goal runs Proxy average Goal average

Why It's So Hard to Fix

Goodhart's Law is pernicious because the people inside the system often can't see it happening:

What Can You Do?

Goodhart's Law can't be eliminated. The gap between proxy and goal is inherent in the act of measurement. But it can be managed:

The Transferable Insight

Goodhart's Law is about the limits of measurement. Every metric is a model of reality, and every model is a simplification. When you push on the simplification hard enough, it breaks. The system finds the gap between the map and the territory.

The habit worth building: whenever you see a metric being used to drive decisions, ask what's not being captured. The most dangerous metric isn't one that's wrong. It's one that's almost right. It earns your trust during the easy phase, then leads you astray once the system learns to game it.

This is one of the deepest problems in AI alignment. A sufficiently powerful AI optimizing for a proxy of human values will find ways to satisfy the proxy that humans never intended. Not out of malice, but out of the same mechanical gap between "what we measured" and "what we meant." Goodhart's Law doesn't care whether the optimizer is a person, an institution, or an algorithm. The gap is the gap.