This is a thread on presidential election forecasting, polls, and fundamentals, based on my research and experience forecasting the 2012 and 2016 elections.

Let's walk through some of the debates and why it gets contentious.
If we had to guess, right now, who's going to win in 2020, it's probably Joe Biden. Why? Because Trump is unpopular, the economy's terrible, and Biden is leading in the polls. https://projects.economist.com/us-2020-forecast/president
But that's a very rough claim. It doesn't tell us how certain we should be, or say anything about what states Biden/Trump is going to win. It's also not very rigorous -- what if only 1 or 2 of those conditions were true? Then what would we think?
Traditionally what election forecasters have done is looked at past elections and compared some measure of those years' political/economic climate (the "fundamentals") to the election outcome. Then they extrapolate from those correlations to predict the current year's election.
For example, presidents who are more popular in June tend to get more votes in November. When second quarter GDP growth is higher, on average, the candidate from the incumbent president's party does better. https://twitter.com/DrewLinzer/status/1289290422754279424
This sort of analysis is useful from an academic standpoint where researchers are trying to test theories about why people vote the way they do. But from a statistical forecasting perspective, there are major challenges/problems with this approach.
For starters, there's hardly any historical data to extrapolate from -- typically fewer than 20 past presidential elections are considered comparable. With that small sample size, any forecasts are going to be very imprecise.
Second, it's not clear which fundamental factors are the best predictors, and there's not enough data to test or optimize this. A lot depends on the judgment of the analyst. Put simply, if you select different fundamental factors, you will get different forecasts.
In 2015, @benlauderdale and I did our best to address this in a research paper -- built a Bayesian model with lots of inputs, and still concluded: "Until more elections have been observed, our expectations about forthcoming U.S. presidential elections cannot be very strong."
One solution to this forecasting problem is to bring in another source of data: pre-election polls. Then the fundamentals can be used as vaguely informative prior expectations that get updated as more polling becomes available. I wrote about this in JASA https://votamatic.org/wp-content/uploads/2013/07/Linzer-JASA13.pdf
But the polls are no silver bullet either. Polling isn't available every day or in every state, and the polls have errors -- random and systematic. Now forecasters have to "fill gaps" in each state's polling AND determine how much info each poll should contribute to the forecast.
Some analysts try to estimate the direction of the error in the polls, and "correct" for it. With limited data, this is difficult too. The adjustments may seem sensible on their face, but a lot of times they only add complexity and unless done carefully, can make forecasts worse.
What election forecasters are left with is a bunch of uncertain fundamentals and polling data of uncertain quality.

The wisest approach may simply be to add more uncertainty in both the Democratic and Republican directions, recognizing the weaknesses of all of the evidence.
Forecasters differing on the "right" amount of uncertainty is how we ended up with 2016, where every major forecaster's Electoral Vote forecast was Clinton 323, Trump 215, but the chances of Clinton winning ranged from 71% to 99%. https://www.nytimes.com/interactive/2016/upshot/presidential-polls-forecast.html
Two identical forecasts can come up with very different probabilities of winning, by expanding the uncertainty around each point estimate. https://twitter.com/DrewLinzer/status/804824244748099584
Election forecasters aren't oracles, they're analysts with systems for processing data. All else equal, good forecasters devise systems that are empirically defensible, robust to outliers, theoretically well-motivated, honest about uncertainty, and preferably mostly transparent.
Election forecasting serves a lot of important purposes. We're lucky to have a range of smart people doing it quantitatively, in rigorous but somewhat different ways. For all the challenges and debates, the alternative is pure punditry and guesswork, and that's way worse. /end
You can follow @DrewLinzer.
Tip: mention @twtextapp on a Twitter thread with the keyword “unroll” to get a link to it.

Latest Threads Unrolled: