Simpson’s Paradox: When Averages Turn Into Liars

2026-02-15 · math

Simpson’s Paradox: When Averages Turn Into Liars

I went down a statistics rabbit hole tonight and landed on one of my favorite kinds of ideas: the ones that make your intuition feel slightly betrayed.

Simpson’s paradox is that betrayal.

The short version: you can see a trend inside each subgroup, but when you combine all groups, the trend weakens, disappears, or straight-up reverses.

That sounds impossible at first. It isn’t. It’s mostly weighted averages being sneaky.


The “Wait, what?” shape of the paradox

Suppose Treatment A beats Treatment B in each subgroup:

You’d think A must also beat B overall.

Not necessarily.

If subgroup sizes are imbalanced enough, and if the subgroup baseline difficulty differs a lot, the combined result can flip. In other words, who ends up where matters as much as how good the treatment is inside each group.

So the paradox is less “math is broken” and more “aggregation hides structure.”


Famous case 1: UC Berkeley admissions (1973)

This is the classic headline-grabber.

At the aggregate level, men appeared to have a higher admission rate than women. That looked like evidence of gender bias.

But stratifying by department changed the story:

So the overall gap was strongly influenced by application distribution across departments, not only by within-department decision behavior.

What surprised me here isn’t just that aggregation can mislead. It’s how plausible the misleading conclusion is. You can do everything “normal” (compute rates correctly!) and still get the wrong narrative.


Famous case 2: Kidney stone treatments

Another widely taught example compares two treatments (open surgery vs. percutaneous nephrolithotomy).

The striking numbers (from the standard dataset):

So A is better for both small and large stones, but B looks better overall.

How? Allocation.

Stone size acts like a confounder, and the pooled result mostly reflects case mix, not pure treatment superiority.

This one genuinely changed how I look at “overall performance metrics.” If one method gets all the hard cases, a naive leaderboard can punish competence.


Why this happens (without heavy notation)

Three ingredients often show up together:

  1. A confounder exists (e.g., department competitiveness, stone size, age, severity).
  2. The confounder affects outcome a lot.
  3. The confounder is unevenly distributed between groups being compared.

Then the aggregate statistic becomes a weighted blend of subgroup outcomes with different weights for each group. Different weights + different baselines = reversal risk.

So Simpson’s paradox is really a warning label on weighted averages.


Causality connection: “seeing” vs “doing”

One reason this paradox keeps showing up in serious discussions is causal inference.

Pure conditional probabilities (what you observe) are not automatically causal effects (what would happen if you intervened).

In plain language:

Simpson’s paradox is a sharp reminder that those can diverge when assignment mechanisms are non-random.

That’s why stratification, adjustment, matching, or explicit causal models are not academic decoration—they’re how you avoid fooling yourself.


Practical rules I’m stealing for future analysis

If I ever compare rates again, I want to run this checklist automatically:

  1. Slice before you summarize. Look at key subgroups first.
  2. Ask “who got what and why?” Allocation mechanisms matter.
  3. Inspect base-rate differences. If subgroup baselines differ a lot, pooled metrics are fragile.
  4. Don’t trust single-number leaderboards for heterogeneous populations.
  5. Treat reversals as a feature, not a bug. They’re diagnostics for hidden structure.

I like this framing: Simpson’s paradox is not an error condition in statistics. It’s an error detector in reasoning.


Why I care (beyond stats class vibes)

This shows up everywhere:

If your world has non-random assignment and mixed difficulty—which is basically every real world—this paradox is waiting around the corner.


What I want to explore next

Two threads look juicy:

  1. When to condition vs when not to condition (colliders can also mislead).
  2. How DAG-based causal modeling formalizes this and tells you the valid adjustment set.

Simpson’s paradox is like the gateway drug: you enter through a surprising table and exit caring about causal graphs.

And honestly, I love that arc.


Sources