The first (and possibly last) debate between the two major party nominees for president of the United States is in the books.
It’s too early to know exactly how the debate will affect polling averages and the various statistical models that rely on them. But movements in prediction markets give us a glimpse of what might be on the horizon.
The figure below shows prices for the Harris contract on PredictIt and Polymarket over a twenty-four hour period that encompasses the debate, adjusted to allow for their interpretation as probabilities and to facilitate comparison with statistical models.1
Five hours before the debate, the markets disagreed on the outcome of the election, with PredictIt seeing the race as a coin flip and Polymarket giving Trump the edge. Following the debate the markets continue to disagree, but now Polymarket sees the race as evenly poised, while PredictIt gives Harris the edge.
Despite these disagreements, the two markets responded in very similar fashion to the debate—they moved in the same direction to roughly the same degree. One hour into the debate, the likelihood of a Harris victory had risen from 50 to 54 on PredictIt and from 47 to 50 on Polymarket. Prices fluctuated around these higher levels thereafter.
Statistical models such as those published by FiveThirtyEight, Silver Bulletin, and the Economist cannot respond to such events instantaneously—it will take several days for the effect of the debate (if any) to make itself felt in horse-race polls, and the models will respond when the polls do.
This is one advantage of markets as forecasting mechanisms—they can take into account novel and qualitative sources of information without delay, and continue to make coherent predictions even when models lose continuity and reliability in the face of unprecedented events.
But markets have drawbacks too—they can be subject to overreaction and manipulation. So I will be looking to see how model forecasts evolve over the coming days. There will be confounding factors, of course, so this is not a clean experiment. But what markets are predicting is an increased likelihood of victory by about three to four percentage points for Harris once the models have adjusted to the effects of the debate.
While the markets responded to the debate in similar fashion, the disagreement between them regarding the election outcome has not narrowed. This rasies the question of how such disagreement can be sustained in the face of financial incentives. Couldn’t traders bet against Trump on Polymarket and against Harris on PredictIt, locking in a certain gain of about four percent over two months, or more than twenty-six percent at an annualized rate? And wouldn’t the pursuit of such arbitrage opportunities bring prices across markets into alignment?
There are several obstacles to executing such a strategy. PredictIt is restricted to verified residents of the US who fund accounts with cash, while trading on Polymarket is crypto-based and the exchange does not accept cash deposits from US residents. This leads to market segmentation and limits cross-market arbitrage. In addition, PredictIt has a limit of $850 on position size in any given contract, as well as a punishing fee structure.
This leads to a natural question: which market should one take more seriously as a forecasting mechanism? Polymarket has no limits on position size and enjoys much greater trading volume, with almost 900 million dollars having been wagered in the presidential winner market to date (compared with about 37 million in the corresponding market on PredictIt). But larger volume does not always mean better forecasting, and position limits can improve forecasting accuracy by ensuring that no trader is large relative to the market as a whole.
While arbitrage places limits on the extent to which markets can disagree, there is no such constraint on statistical models. Here the disagreement is substantially greater—the probability of a Trump victory ranges from 45 percent on FiveThirtyEight to 49 percent on the Economist and 62 percent on Silver Bulletin.
Why the striking difference across models that use basically the same ingredients? One reason is a questionable “convention bounce adjustment” in the Silver Bulletin model, without which its disagreement with FiveThirtyEight would be negligible.
But there also seem to be some deep differences in the underlying correlation structure in these models that I find extremely puzzling. For example, according to the Silver Bulletin model, Trump is more likely to win New Hampshire (30 percent) than Harris is to win Arizona (23 percent). The other two models rank these two states very differently, with a Harris victory in Arizona being significantly more likely than a a Trump victory in New Hampshire. Convention bounce adjustments aside, the correlation structure across states in the Silver Bulletin model just doesn’t seem plausible to me.
In any case, once the election is over, we can evaluate the forecasting performance of the various models and markets based on a profitability test of predictive accuracy. Our analysis of this is already underway, and I look forward to reporting the results in due course.
Update: Andrew Gelman has written a detailed and thoughtful response to this post.
The adjustment is necessary because the prices of the two contracts don’t always sum to one, on account of fees and the fact that the they are traded independently. To convert prices into probabilities, one can divide the price of each contract by the sum of the prices of the two contracts. For example, at the end of the period shown, Harris and Trump contracts were trading at 55 cents and 47 cents respectively on Predictit. This translates to probabilities of 54 and 46 percent. The adjusted prices of the Trump contracts are (by construction) a mirror image of the Harris prices:
I’ve noticed that there are notable differences among the betting markets. It’s interesting to get some insight as to the differences. Still, why do those differences translate, say, to Predictit consistently, having Haris a larger favorite than.Polymarket?
Also interesting that the Real Clear betting market aggregate has just recently dropped Predictit from their list.
https://www.realclearpolling.com/betting-odds/2024/president
I've been curious about the difference between the PredictIt and Polymarket numbers, too. Thanks for the additional info.
May not be relevant, but I think Silver has said his current model retains most of the features of the old 538 models he spearheaded. Andrew Gelman had an interesting post looking at strange edge cases in those models during the last election: https://statmodeling.stat.columbia.edu/2020/10/24/reverse-engineering-the-problematic-tail-behavior-of-the-fivethirtyeight-presidential-election-forecast/.