march madness w/ the boys
All Articles

Seeding & Expert Fallibility

40 years of NCAA selection committee decisions under the microscope

2,518 tournament games • 1985–2024

27%
of first-round games are "upsets" where the lower seed wins
~Flat
Upset rate hasn't improved over 40 years — the committee isn't learning
11 > 10 > 9
11-seeds outperform both 9 and 10 seeds — a persistent mispricing
3
1-seeds have lost in the Round of 64 (should be near zero)

Finding 1: The Committee Gets It Right About 73% of the Time

That sounds good until you realize the rate doesn't improve as the stakes get higher

In the first round, the higher-seeded team wins 72.9% of the time. You might expect that number to climb in later rounds — after all, the surviving teams have proven themselves, and the committee's best-seeded teams should separate from the pack. Instead, the upset rate stays remarkably flat:

RoundGamesHigher Seed WinsUpset Rate
Round of 641,45372.9%27.1%
Round of 3282869.3%30.7%
Sweet 1617868.9%31.1%
Elite 84971.1%28.9%
Final Four1071.4%28.6%

This is a key economic insight: the seedings contain meaningful information, but roughly a quarter of it is wrong in any given game. If seeding were a financial forecast, you'd say the model has a consistent ~27% error rate that it can't seem to reduce, regardless of the sample or the round.

The economic framing: Think of each seed assignment as a "price" set by a committee of experts. A 1-seed is priced as a strong favorite; a 16-seed is priced as cannon fodder. The question is: are these prices well-calibrated? The answer is "pretty good but with consistent, exploitable blind spots."

Finding 2: 40 Years of Data and the Committee Hasn't Gotten Better

The upset rate in the first round has been essentially flat since 1985
1985-94
23.9%
1995-04
24.2%
2005-14
24.3%
2015-24
25.2%

First-round upset rate by decade

If anything, the rate has drifted slightly upward — from 23.9% in the first decade to 25.2% in the most recent. The committee has access to more data, more analytics, and more computing power than ever before. Yet their hit rate hasn't budged. This is consistent with a well-known finding in forecasting research: expert judgment tends to plateau, and more information doesn't automatically translate into better predictions.

Compare to other expert domains: Weather forecasting has dramatically improved over 40 years thanks to better models and data. Stock analyst forecasts have not. The Selection Committee falls into the latter category — a fundamentally uncertain system where more data doesn't create proportionally better outcomes.

Finding 3: The 11-Seed Anomaly SYSTEMATIC MISPRICING

11-seeds outperform 9 and 10 seeds — a persistent error the committee has never corrected

This is the single most striking pattern in the data and the clearest evidence of expert fallibility. If seeding were perfectly calibrated, performance should decline monotonically from seed 1 to seed 16. It doesn't. The 11-seed line is out of order:

#8 seed
0.71 avg wins
#9 seed
0.62 avg wins
#10 seed
0.61 avg wins
#11 seed
0.70 avg wins
#12 seed
0.52 avg wins

Red marker = where 11-seeds "should" be based on their seed number

The numbers get more dramatic the deeper you go:

#11 seeds
Sweet 16 rate
6.9%
Final Four appearances
6
VCU '11, UCLA '21, George Mason '06, Loyola Chicago '18, NC State '24, LSU '86
#9 + #10 seeds combined
Sweet 16 rate
3.2% / 5.7%
Final Four appearances
1 combined
9-seeds: 0 Final Fours. 10-seeds: 1 (Syracuse 2016)

11-seeds have made 6 Final Fours. 9-seeds and 10-seeds have made 1 combined. This isn't a fluke — it's a 40-year pattern across 174 11-seed entries. Several possible explanations:

Hypothesis 1: The "warm-up game" effect. Since the tournament expanded, some 11-seeds play in the First Four (play-in round). Getting an extra competitive game before facing a 6-seed could provide a meaningful warm-up advantage — game-speed reps, travel logistics shakedown, and facing elimination pressure once before the "real" tournament starts.
Hypothesis 2: The "bubble team with upside" profile. 11-seeds are often talented teams from major conferences that underperformed in the regular season, or mid-major teams that dominated weaker conferences. Both profiles contain hidden upside: major-conference teams have faced tougher competition, and dominant mid-majors may have a star player or system that gives fits to anyone. The committee sees "uneven resume" and assigns seed 11, but the talent ceiling is much higher.
The economic parallel: This is functionally identical to a market mispricing. The "market" (the committee) consistently undervalues a specific asset class (11-seeds) relative to comparable assets (9 and 10 seeds). If you could bet on 11-seeds making the Sweet 16 at odds implied by their seed number, you'd have a positive expected value bet — for 40 years running.

Finding 4: The 5-vs-12 Game Is Less of a "Lock" Than Advertised

12-seeds win 35% of the time — and the rate was approaching coin-flip territory for a decade

The 5-vs-12 matchup is famous in bracket circles as the most common upset pick, and the data validates it:

Era5-Seed Win %12-Seed Win %
1985-199472%28%
1995-200462%38%
2005-201455%45%
2015-202469%31%

From 2005-2014, 12-seeds won 45% of the time — almost a coin flip. The rate has pulled back recently, but the structural issue remains: the committee consistently assigns 5-seeds to teams that aren't meaningfully better than 12-seeds. The gap between a "good but not great" major-conference team (typical 5-seed) and a "best team in a mid-major conference" (typical 12-seed) is smaller than the committee prices it.

Finding 5: When the Committee's Best Pick Loses Immediately

A 1-seed losing in the Round of 64 should be nearly impossible — it's happened 3 times

1-seeds are supposed to be the committee's four most confident selections. They get the easiest path — facing a 16-seed in round one, then an 8 or 9 seed. Yet:

Year1-SeedLost ToScoreContext
2018Virginia#16 UMBC54-74First 1-vs-16 upset in history. 20-point loss.
2023Purdue#16 Fairleigh Dickinson63-66Second 1-vs-16 upset ever. Both in 5 years.
-Additional early exits: 6 different 1-seeds have been eliminated in the Round of 32 (2nd round loss)

The fact that both 1-vs-16 upsets happened in the last 6 years is noteworthy. Either the committee's top selections have gotten weaker, the 16-seed play-in game provides a warm-up advantage (similar to our 11-seed hypothesis), or random variance is clustering in a small sample.

The "overconfidence" parallel: In behavioral economics, experts tend to be most miscalibrated at the extremes of their confidence range. The committee's most confident pick (1-seed) should almost never fail immediately — but the rate of catastrophic failure (1.3% in R64, 5.1% out by R32) is higher than the implied confidence of the seeding would suggest.

Finding 6: The ACC Gets Favorable Seeding — Mountain West Gets Punished

Some conferences consistently outperform their seeds, others consistently underperform

If we compare each team's actual tournament wins to the expected wins for their seed, we can measure whether a conference's teams are systematically over- or under-seeded:

ConferenceTourney EntriesAvg SeedActual Avg WinsExpected for SeedOver/Under
ACC2144.91.821.63+0.19
SEC2005.91.441.33+0.10
Big East2235.61.511.45+0.06
Big Ten2365.61.431.43+0.01
Big 121595.51.361.48-0.11
Pac-10 (old)1105.81.281.39-0.11
Mountain West658.80.480.80-0.33

The ACC consistently outperforms its seeding (+0.19 wins per entry), meaning the committee gives ACC teams slightly worse seeds than they deserve. Conversely, the Mountain West is the most "overseeded" conference — their teams win a third of a game less than their seeds would predict.

The Big 12 and old Pac-10 underperform their seeds too. This suggests the committee may give slight favoritism to certain conferences in the seeding process — or that some conferences' regular-season competition doesn't prepare teams as well for tournament-style play.

Finding 7: The Committee's Uncertainty Grows Exponentially in the Middle

Seeds 5-12 have coefficient of variation over 100% — the committee is essentially guessing within this range

The coefficient of variation (standard deviation divided by mean) tells us how "noisy" each seed's performance is relative to expectations:

#1 seed
50% CV
#2 seed
63% CV
#3 seed
77% CV
#4 seed
82% CV
#5 seed
102% CV
#6 seed
107% CV
#7 seed
109% CV
#8 seed
150% CV
#11 seed
153% CV

Coefficient of variation by seed (higher = more unpredictable relative to expectation)

Seeds 1-2 are reasonably well-calibrated (CV around 50-63%). By seed 5, the CV crosses 100% — meaning the standard deviation is larger than the mean. The committee's predictions for seeds 5-12 contain more noise than signal. In economic terms, the "price" the committee assigns to these middle seeds is barely more informative than random assignment.

Finding 8: The Hall of Shame — The Committee's Biggest Misses

Champions who were seeded 6th or worse, and 1-seeds who went home immediately

Champions the committee underrated:

YearChampionSeedThe Miss
1985Villanova#8Won the title as an 8-seed. Still the lowest seed to ever win it all.
2014UConn#7A 7-seed championship in the modern era. Committee badly mispriced them.
1988Kansas#6Danny Manning carried a 6-seed to the title.
1997Arizona#4Beat 3 number-1 seeds en route to the championship.
2023UConn#4Won every game by double digits — a dominant 4-seed the committee undersold.

1-seeds that flamed out earliest:

Year1-SeedResultThe Miss
2018VirginiaLost in R64 to #16 UMBCThe committee's #1 overall seed lost by 20 to a 16-seed. Maximum error.
2023PurdueLost in R64 to #16 FDUSecond 1-vs-16 upset in 6 years. Pattern or fluke?
2011PittsburghLost in R32 to #8 ButlerOut by the second weekend despite #1 seed.
2017VillanovaLost in R32 to #8 WisconsinDefending champion bounced early.
2022BaylorLost in R32 to #8 UNCDefending champion also bounced early.

The Expert Fallibility Story

The NCAA Selection Committee is a fascinating case study in expert judgment. They get the broad strokes right — 1-seeds are genuinely much better than 16-seeds, and the top 4 seeds win the title 92% of the time. But within that framework, the data reveals persistent, exploitable patterns:

1. The committee doesn't improve over time. Despite 40 years of feedback and increasingly sophisticated analytics, the first-round upset rate has been flat at ~24-25%. More data hasn't made the experts better.

2. 11-seeds are systematically mispriced. They outperform 9 and 10 seeds by every measure — average wins, Sweet 16 rate, Final Four appearances. This has been true for 40 years and the committee hasn't corrected it.

3. Mid-range seeds (5-12) are essentially noise. The coefficient of variation exceeds 100%, meaning the committee's signal is weaker than the randomness. The distinction between a 5-seed and an 8-seed is not very meaningful.

4. Conference bias exists. ACC teams outperform their seeds; Mountain West and Big 12 teams underperform. The committee appears to apply a subtle brand premium to certain conferences.

The economic lesson: Expert committees — whether they're seeding basketball tournaments, rating bonds, or setting policy — tend to be well-calibrated at the extremes but noisy in the middle, prone to status-quo bias, and resistant to self-correction even when feedback is immediate and clear.