Wisdom of Crowds Explained: Why Prediction Markets Aggregate Better Than Experts

In March 2026, traders on Kalshi bet $4.04 million on whether Elon Musk would tweet more than 75 times in 24 hours. Others wagered on whether Trump would drink water during his State of the Union address (81% said no), whether he'd say "discombobulator" (9% probability), and how long a handshake would last. The weird specificity wasn't the point. The accuracy was. These markets, built on thousands of people betting real money, consistently outperformed pundit predictions and traditional polls throughout 2024 and 2025.

Apr 2026|17 min read

That accuracy runs on a principle economist Francis Galton discovered in 1906. At a county fair, 800 attendees guessed an ox's weight. Most had never weighed livestock. Nobody there qualified as an expert. Yet their median guess landed within nine pounds of the truth, closer than any single butcher or farmer managed. The collective estimate beat the best individual judgment.

Wisdom of crowds means aggregating diverse, independent guesses produces estimates more accurate than most experts generate alone. When enough people make predictions based on different information and reasoning, their errors tend to cancel out while their accurate signals reinforce each other. Prediction markets create a financial mechanism for this. Every trade is a judgment. Every price reflects thousands of judgments combined. Every participant risks money, which keeps them honest.

But the principle only works under specific conditions. When people copy each other, when a few loud voices dominate, when everyone shares the same blind spot, crowds don't get smarter. They amplify mistakes. Understanding when wisdom emerges and when it collapses explains why Polymarket correctly called the 2024 election when traditional polls missed, and why markets sometimes crash spectacularly when everyone bets the same direction.

When Crowds Get Smart: The Galton Principle in Practice

During the final five weeks before the 2024 election, Polymarket processed $2.4 billion in transactions across federal, state, and congressional races. Daily trading volume hit $478 million in March 2026, higher than stocks like Ralph Lauren or Molson Coors on traditional exchanges. The final odds correctly predicted the winner in 48 of 50 states, closer to the actual vote margins than any single polling firm achieved.

Traditional polls missed Trump's strength in swing states by an average of 3.2 percentage points. Polymarket's final odds were off by an average of 1.8 percentage points.

The difference wasn't that prediction market traders were smarter. They were structurally better at processing distributed information.

Here's how aggregation works: if 100 people guess a quantity and their errors are truly independent, the average guess becomes 10 times more accurate than a typical individual guess. With 10,000 people, the aggregate becomes 100 times more accurate. Philip Tetlock spent two decades tracking 28,000 predictions from 284 political experts for a government intelligence project. His finding: aggregated amateur forecasts on platforms like Metaculus beat the experts by 23% on average.

The crowds weren't smarter person-by-person. They canceled out each other's errors.

Metaculus, which has tracked over 12,000 questions since 2015, published accuracy data in 2023 showing their community predictions beat baseline models and expert panels by substantial margins. On questions about scientific breakthroughs, technological development, and political events, the aggregated crowd forecast beat 80% of individual expert forecasts. The mechanism: continuous updating as new information emerges, scoring incentives that reward accuracy over time, and a participant base ranging from professional researchers to curious amateurs.

Kalshi showed similar accuracy on non-political questions. Their March Madness bracket market correctly predicted 14 of 16 Sweet Sixteen teams in 2026, outperforming 94% of ESPN bracket entries. Federal Reserve rate decision markets matched the actual decision in 22 of 23 meetings from 2023 through early 2026. On economic indicators like monthly jobs numbers and CPI inflation, market-implied forecasts beat the consensus economist forecast in seven of twelve months measured through March 2026.

Markets don't just aggregate opinions. They weight those opinions by confidence. A trader willing to risk $10,000 on an outcome moves the price more than someone betting $10. This creates a natural filter: people with strong evidence bet more aggressively, while those with weak hunches bet cautiously. The price reflects not just the average opinion, but the average confidence-weighted opinion across all active participants.

The Four Conditions That Make Crowd Wisdom Work

James Surowiecki's 2004 book The Wisdom of Crowds outlined four conditions that determine whether aggregation produces accuracy or amplifies error:

Diversity of opinion. Each person brings different information sources. Some follow political polling, others track fundraising data, others parse candidate rhetoric.

Independence of judgment. People bet based on their own analysis rather than copying their neighbors.

Decentralization of information. No single participant sees the full picture.

Aggregation mechanism. A system that combines individual inputs into collective output. In prediction markets, the price mechanism automatically aggregates every judgment into a single number: the current odds.

When these conditions hold, crowd wisdom works. When they break, markets fail in specific, predictable ways.

When Aggregation Breaks: Three Failure Modes

Information cascades destroy independence. If you see 99 people bet one direction, you might follow even if your analysis suggests the opposite. Not because you're irrational, but because you assume those 99 people know something you don't. Analysis of Polymarket trading patterns in 2024 found evidence of herding behavior in thin markets with fewer than 500 active traders. When a large bet suddenly moved odds, subsequent smaller traders tended to bet the same direction regardless of whether fundamentals had changed.

The March 2026 State of the Union betting frenzy illustrated this failure mode. Markets correctly predicted most major speech elements: topics covered, approximate length, specific rhetorical choices. But they failed spectacularly on one question: whether JD Vance would clap during the speech. Odds jumped from 22% to 66% in 30 minutes based on a single large bet, only to crash back to 15% when the speech ended and Vance had barely applauded. The cascade wasn't caused by new information. It was pure herd behavior triggered by one confident trader.

Manipulation poisons the information pool. In February 2026, the CFTC filed enforcement proceedings for insider trading violations on Kalshi. One case involved a video editor for a major YouTube channel who bet $8,500 on specific outcomes related to that channel's content before public announcements. The bets earned $12,147 in profits before enforcement caught the violation. Other participants assumed the large bets reflected genuine information rather than insider knowledge, temporarily distorting market odds.

More serious: politicians trading on their own races. One candidate placed bets totaling $3,200 on their own electoral outcome, then posted evidence of the trades on social media. They paid a $2,246 penalty. These cases mattered not because of the dollar amounts, but because they demonstrated how informed insiders could manipulate prices and mislead other participants about true probabilities.

Groupthink emerges when participant diversity collapses. Research from Boston University's Markets and Policy Initiative examined prediction market accuracy across different question types. Markets populated primarily by one demographic group (tech industry professionals forecasting cryptocurrency questions, for example) showed worse calibration than markets with more diverse participants. These professionals didn't have bad information. They shared common blind spots. Without diverse perspectives to counter those blind spots, errors amplified instead of canceling out.

What the Numbers Actually Show

Berg, Nelson, and Rietz (2008) examined historical prediction market accuracy from Iowa Electronic Markets, Betfair, and Intrade across thousands of events between 2002 and 2012 in American Economic Review. Market prices corresponded to realized frequencies with remarkable precision. A 30% market probability translated to events occurring 31% of the time. A 70% probability meant 69% realization. The markets weren't just directionally correct. They were calibrated, meaning their probability estimates matched how often things actually happened.

But the track record isn't perfect. Analysis of prediction market accuracy in 2024 found systematic biases. Longshot candidates showed inflated odds: a 5% probability in markets often corresponded to 2-3% real-world frequency. Favorite bias appeared in sports betting crossover markets, with heavy favorites trading at odds slightly worse than their true probability.

These weren't failures of aggregation. They were behavioral patterns in how crowds assess extreme probabilities.

Research from Duke University compared individual analyst forecasts to crowd aggregates across multiple domains. The crowd aggregate beat the best individual analyst 70% of the time. Weighting experts more heavily than novices didn't improve accuracy. Simple averaging performed as well or better than complex weighting schemes. The diversity of perspective mattered more than the expertise of individual contributors.

The Regulatory Question: Does Oversight Help or Hurt?

The CFTC issued an advisory and advance notice of proposed rulemaking on March 12, 2026, proposing stricter oversight of event contract markets. The specific concern: prediction markets had grown from niche platforms processing millions in volume to mainstream financial instruments processing billions, without corresponding regulatory infrastructure. Chair Rostin Behnam stated the Commission needed clearer authority to prevent "gaming of the public interest and potential for market manipulation."

Arizona Attorney General Kris Mayes filed criminal charges against Kalshi on March 17, 2026, arguing the platform violated state gambling laws. Kalshi filed a preemptive lawsuit the same day seeking declaratory judgment that CFTC registration preempted state gambling enforcement. The CFTC moved for a preliminary injunction on April 8, creating a three-way legal battle that remained unresolved through early 2026.

Legislative response came in late March 2026 with a proposed bill titled "Prediction Markets Are Gambling Act." The bill sought to classify all real-money prediction markets as gambling regardless of CFTC registration, subjecting them to state gaming regulations. Leaked drafts suggested restrictions on political betting, age requirements, and deposit limits similar to those applied to sports betting platforms.

Restrictions that reduce participation also reduce diversity, one of the core conditions for accurate aggregation. If regulatory barriers limit prediction markets to sophisticated investors or licensed traders, the platforms lose the broad participant base that makes them accurate. Tetlock's government research found that amateur forecasters contributed essential diversity to crowd predictions. Critics argue amateurs introduce noise rather than signal, but the data showed amateur diversity improved accuracy even when individuals performed poorly.

The counterargument: regulation could improve accuracy by eliminating manipulation and insider trading. The CFTC enforcement actions from February 2026 showed how a few bad actors could distort prices and mislead participants. If stronger oversight prevents those distortions, the remaining crowd wisdom would operate on cleaner information.

The legal battles through March and April 2026 will determine whether these markets continue operating as open platforms where anyone can participate, or become restricted instruments accessible only to sophisticated investors. Restricted markets could cut Polymarket's active trader count from 45,000 to under 5,000, eliminating the diversity that produces accuracy.

How This Changes How You Read Odds

A 60% probability on a prediction market doesn't reflect one person's analysis. It reflects thousands of individual judgments, weighted by confidence, aggregated through trades. When you see that price, you're seeing the market's collective assessment based on all available information processed by all active participants.

Whether you can make money trading these markets depends on beating crowd wisdom repeatedly. That's hard precisely because the crowd is already incorporating most available information. If Kalshi shows 65% odds on a Fed rate hike and you think it's actually 70%, you need to ask: what information do I have that these thousands of other traders don't?

Sometimes the answer is "I have genuine edge." More often, it's "I'm overconfident in my analysis."

The exception: you can profit from temporary mispricing caused by cascades, manipulation, or insufficient liquidity. If a large bet moves odds 10 percentage points but fundamentals haven't changed, contrarian traders can capture value by betting against the herd. This requires understanding whether prices reflect information or behavioral quirks.

Traditional polling aggregates opinions through sampling. FiveThirtyEight's model during 2024 combined hundreds of polls using weighted averages based on pollster quality and recency. The aggregation was sophisticated, but it didn't incentivize accuracy. A pollster who missed badly faced reputational cost but no direct financial penalty. Prediction markets create immediate consequences: bet wrong and you lose money, bet right and you profit. This skin-in-the-game dynamic filters participants in ways polling cannot.

Expert consensus represents another aggregation method. Professional forecasters in fields like economics or meteorology build track records over time, and their predictions get aggregated by organizations like the Federal Reserve or the National Hurricane Center. These consensus forecasts perform well in established domains with clear historical patterns. But research showed they underperform crowd forecasts in novel situations without strong precedent, exactly the scenarios where prediction matters most.

Different aggregation methods fail in different ways. Polls can miss by sampling error or response bias. Expert consensus can miss by groupthink or shared assumptions. Prediction markets can miss by manipulation or cascade effects. Understanding these failure modes explains when to trust crowd wisdom and when to seek alternative information sources.

The Edge and the Limit

Collective intelligence produces better probability estimates than individual experts. But it doesn't produce certainty.

A 70% probability means the event happens 70% of the time and fails to happen 30% of the time. Crowds improve how well probabilities match reality, meaning their 70% predictions occur 70% of the time rather than 60% or 80%. But they don't eliminate the underlying uncertainty.

Some phenomena are genuinely unpredictable regardless of information quality. A coin flip remains 50-50 no matter how many people bet on it. Chaotic systems amplify small differences until prediction becomes impossible. Crowds can't overcome fundamental limits on knowability.

The 2024 election demonstrated both the strength and limit. Prediction markets correctly assessed Trump's chances in swing states because polling data, fundraising figures, and early voting patterns provided genuine signals that crowds could aggregate. But no prediction market correctly forecast the exact Electoral College margin or national popular vote share. Those specifics involved too much irreducible uncertainty for even perfect information aggregation to nail precisely.

How AI and Blockchain Change Crowd Diversity

Artificial intelligence introduces a new variable into crowd dynamics. Metaculus published research in March 2026 showing that GPT-4-based forecasting systems performed comparably to median human forecasters on structured prediction questions. This creates a potential enhancement to crowd wisdom: AI systems can process far more data than individual humans, then participate in markets alongside human traders. The resulting aggregate would combine human judgment with machine pattern recognition.

The risk: AI forecasters might introduce correlated errors. If all AI systems share similar training data and architectural assumptions, they might amplify each other's mistakes instead of canceling them out. Research from Anthropic in early 2026 found that multiple AI systems making predictions on the same question showed correlation coefficients of 0.67 compared to 0.23 for multiple humans, suggesting AI participation could reduce effective diversity.

Blockchain-based prediction markets represent another evolution. Polymarket operates on Polygon blockchain infrastructure, creating transparent price history and trade records that can't be retroactively altered. This transparency allows researchers to analyze market accuracy and identify manipulation attempts after events resolve. But it also introduces new risks: smart contract bugs could lock funds or produce incorrect resolutions, destroying trust in the aggregation mechanism.

The institutional adoption question remains unresolved. Will corporations and governments integrate prediction markets into decision-making processes, or will regulatory pressure keep them as fringe platforms? Intelligence research showed agencies could improve forecast accuracy by incorporating crowd predictions. But the March 2026 regulatory actions suggest political resistance to markets that enable betting on elections and policy outcomes.

Frequently Asked Questions

What is the wisdom of crowds principle? Wisdom of crowds is the phenomenon where aggregating many independent guesses produces more accurate estimates than individual expert judgments. Francis Galton discovered this in 1906 when 800 county fair attendees guessed an ox's weight. Their median guess landed within nine pounds of the actual weight, closer than any single expert. The principle works because individual errors cancel out while accurate signals reinforce each other. Prediction markets apply this by converting every trade into a judgment, with prices automatically aggregating thousands of individual predictions weighted by confidence.

When do crowds make better predictions than experts? Crowds outperform experts when four conditions exist: diversity of opinion, independence of judgment, decentralized information, and an effective aggregation mechanism. Philip Tetlock's two-decade study tracking 28,000 predictions from 284 political experts found aggregated amateur forecasts beat experts by 23% on average. Metaculus data from over 12,000 questions showed community predictions beat 80% of individual expert forecasts on scientific, technological, and political questions. The crowd advantage comes from processing distributed information that no single expert can access, not from superior intelligence of individual participants.

What causes prediction markets to fail? Prediction markets fail when the four conditions for crowd wisdom break down. Information cascades occur when people copy others' bets instead of making independent judgments. This happened in March 2026 State of the Union betting when odds jumped from 22% to 66% in 30 minutes based on herd behavior. Manipulation by insiders poisons the information pool, as seen in CFTC enforcement cases where traders with privileged information distorted market prices. Groupthink emerges when participants lack diversity and share common blind spots, causing errors to amplify instead of canceling out.

How accurate are prediction markets compared to polls? Prediction markets showed superior accuracy to traditional polling in the 2024 election. Polymarket correctly predicted winners in 48 of 50 states with odds averaging 1.8 percentage points from actual vote margins, while traditional polls missed by 3.2 percentage points on average. Berg, Nelson, and Rietz (2008) examined thousands of events and found market prices corresponded to realized frequencies with remarkable precision. A 30% market probability translated to events occurring 31% of the time. However, markets show systematic biases: longshot candidates have inflated odds, with 5% market probabilities often corresponding to 2-3% real-world frequency.

Can you make money by betting against prediction market crowds? Beating crowd wisdom repeatedly is difficult because markets already incorporate most available information from thousands of participants. You need genuine information edge that other traders don't have, which is rare. The exception is temporary mispricing from cascades, manipulation, or low liquidity. If a large bet moves odds 10 percentage points without fundamental changes, contrarian traders can profit by betting against the herd. Research showed simple averaging of crowd predictions performed as well as complex weighting schemes, meaning the collective assessment is hard to beat consistently without privileged information.

Will AI improve or hurt prediction market accuracy? AI introduces both opportunities and risks to crowd wisdom. Metaculus research from March 2026 showed GPT-4-based forecasting systems performed comparably to median human forecasters, suggesting AI could enhance aggregation by processing more data than humans. However, Anthropic research found multiple AI systems showed correlation coefficients of 0.67 compared to 0.23 for humans on the same questions. This means AI forecasters might introduce correlated errors instead of independent judgments, reducing the diversity that makes crowd wisdom work. The net effect depends on whether AI adds new information or just amplifies existing patterns.

The Mechanism, Not the Magic

On a February afternoon in 1906, Francis Galton stood at a country fair watching 800 people guess the weight of an ox. He collected their guesses, calculated the median, and discovered something that changed how we think about collective judgment. The crowd, collectively, knew more than any individual expert.

That principle now powers markets processing billions in volume, predicting everything from elections to Fed rate decisions to whether a politician will drink water during a speech. The accuracy isn't magic. It's architecture: diverse people making independent judgments, aggregated through a price mechanism that weights confidence and incorporates new information continuously.

Sometimes the architecture works. Polymarket called 48 of 50 states in 2024 more accurately than any polling firm. Kalshi predicted 14 of 16 March Madness Sweet Sixteen teams. The crowd sees what individuals miss.

Sometimes it breaks. Cascades amplify herd behavior. Insiders manipulate prices. Groupthink eliminates diversity. The crowd becomes a mob, amplifying errors instead of canceling them.

Platform reviews of Polymarket and Kalshi's structure show how different design choices affect participant diversity and independence. Some platforms attract narrow demographics, creating groupthink risk. Others build features that encourage contrarian betting and diverse participation. The regulatory battles through 2026 will determine whether these markets continue operating as open platforms where anyone can participate, or become restricted instruments accessible only to sophisticated investors.

The crowds know more than you think. They also know less than the price sometimes suggests. Your job isn't to trust them blindly or dismiss them entirely. Your job is to understand the architecture that makes them smart, recognize the conditions that make them dumb, and know which situation you're looking at when you read the odds.