Pollsters seemed to finally get it right in 2024. After years of bad misses, they said the presidential election would be close, and it was.
In fact, the industry did not solve its problems last year. In 2016, pollsters famously underestimated Donald Trump by about 3.2 points on average. In 2024, after eight years of introspection, they underestimated Trump by … 2.9 points. Many of the most accurate pollsters last year were partisan Republican outfits; many of the least accurate were rigorous university polls run by political scientists.
Polls can’t be perfect; after all, they come with a margin of error. But they should not be missing in the same direction over and over. And chances are the problem extends beyond election polling to opinion surveys more generally. When Trump dismisses his low approval ratings as “fake polls,” he might just have a point.
For years, the media have been covering the travails of the polling industry, always with the premise that next time might be different. That premise is getting harder and harder to accept.
Polling used to be simple. You picked up the phone and dialed random digits. People answered their landline and answered your survey. Then, you published the results. In 2000, nearly every national pollster used this methodology, known as random-digit dialing, and their average error was about two points. In subsequent elections, they got even closer, and the error, small as it was, shifted from overestimating Bush in 2000 to underestimating him in 2004—a good sign that the error was random.
Then came the Great Polling Miss of 2016. National polls actually came pretty close to predicting the final popular-vote total, but at the state level, particularly in swing states, they missed badly, feeding into the narrative that Hillary Clinton’s win was inevitable.
The 2016 miss was widely blamed on education polarization. College graduates preferred Clinton and were more likely to respond to polls. So, going forward, most pollsters began adjusting, or “weighting,” their results to counteract the underrepresentation of non-college-educated voters. In 2018, the polls nailed the midterms, and pollsters rejoiced.
That reaction turned out to be premature. The 2020 election went even worse for the polling industry than 2016 had. On average, pollsters had underestimated Trump again, this time by four points. Joe Biden won, but by a much slimmer margin than had been predicted.
This sent pollsters searching for a solution yet again. If weighting by education didn’t work, then there must be something specific about Trump voters—even Trump voters with a college degree—that made them less likely to answer a poll. So, many pollsters figured, the best way to solve this would be weighting by whether the respondent had previously voted for Trump, or identified as a Republican. This was a controversial move in polling circles. The proportion of the electorate that is Democratic or Republican, or Trump-voting, changes from election to election; that’s why polls exist in the first place. Could such elaborate modeling turn polls into something more like predictions than surveys?
“This is where some of the art and science get a little mixed up,” Michael Bailey, a Georgetown professor who studies polling, told me. If you weight a sample to be 30 percent Republican, 30 percent Democrat, and 40 percent independent—because that’s roughly how people self-identify when asked—you are making an assumption about how the three groups will behave, not merely matching a poll to population demographics such as age, gender, and education.
These assumptions vary from pollster to pollster, often reflecting their unconscious biases. And for most pollsters, these biases seem to point in the same direction: underestimating Trump and overestimating his opponent. “Most pollsters, like most other people in the expert class, are probably not huge fans of Trump,” the election-forecasting expert Nate Silver told me. This personal dislike may not seem to matter much—after all, this should be a science—but every decision about weighting is a judgment call. Will suburban women show up to vote in 2024? Will young men? What about people who voted for Trump in 2020? All three of these respondent groups have a different weight in an adjusted sample, and the weight that a pollster chooses reflects what the pollster, not the respondents, thinks about the election. Some pollsters might even adjust their weights after the fact if they see a result they find hard to believe. The problem is that sometimes, things that are hard to believe happen, such as Latino voters moving 16 points to the right.
This dynamic might explain a curious exception to the trend last year. Overall, most polls missed yet again: The average error was a three-point underestimate of Trump, the same as 2016. But Republican-aligned pollsters did better. In fact, according to Silver’s model (others have similar results), four of the five most accurate pollsters in 2024, and seven of the top 10, were right-leaning firms—not because their methods were different, but because their biases were.
The most basic problem in 2024 was the same as in 2016: nonresponse bias, the name for the error that is introduced by the fact that people who take polls are different from those who don’t.
A pollster can weight their way out of this problem if the difference between those who respond and those who don’t is an observable demographic characteristic, such as age and gender. If the difference is not easily observable, and it’s correlated with how people vote, then the problem becomes extremely difficult to surmount.
Take the fact that Trump voters tend to be, on average, less trusting of institutions and less engaged with politics. Even if you perfectly sample the right proportion of men, the right proportions of each age group and education level, and even the right proportion of past Trump voters, you will still pick up the most engaged and trusting voters within each of those groups—who else would spend 10 minutes filling out a poll?—and such people were less likely to vote for Trump in 2024. So after all that weighting and modeling, you still wind up with an underestimate of Trump. (This probably explains why pollsters did quite well in 2018 and 2022: disengaged voters tend to turn out less during midterm elections.)
This problem almost certainly afflicts presidential-approval polls too, though there’s no election to test their accuracy against. Low-trust voters who don’t answer polls don’t suddenly transform into reliable respondents once the election’s over. According to Nate Silver’s Silver Bulletin poll aggregator, Trump’s approval is currently six percentage points underwater. But if those approval polls are plagued by the same nonresponse bias as election surveys were last year—which could well be the case—then he’s at only negative 3 percent. That might not seem like a big difference, but it would make Trump’s approval rate historically pedestrian, in line with where Gerald Ford was at roughly this point in his presidency, rather than historically low.
Jason Barabas, a Dartmouth College political scientist, knows something about nonresponse bias. Last year, he directed the new Dartmouth Poll, described by the college as “an initiative aimed at establishing best practices for polling in New Hampshire.” Barabas and his students mailed out more than 100,000 postcards across New Hampshire, each with a unique code to complete a poll online. This method is not cheap, but it delivers randomness, like old-school random-digit dialing.
The Dartmouth Poll also applied all the latest statistical techniques. It was weighted on gender, age, education, partisanship, county, and congressional district, and then fed through a turnout model based on even more of the respondent’s biographical details. The methodology was set beforehand, in keeping with scientific best practices, so that Barabas and his research assistant couldn’t mess with the weights after the fact to get a result that fit with their expectations. They also experimented with ways to increase response rates: Some respondents were motivated by the chance to win $250, some were sent reminders to respond, and some received a version of the poll that was framed in terms of “issues” rather than the upcoming election.
In the end, none of it mattered. Dartmouth’s polling was a disaster. Its final survey showed Kamala Harris up by 28 points in New Hampshire. That was wrong by an order of magnitude; she would win the state by 2.8 points the next day. A six-figure budget, sophisticated methodology, the integrity necessary to preregister their methodology, and the bravery necessary to still release their outlier poll—all that, only to produce what appears to have been the most inaccurate poll of the entire 2024 cycle, and one of the worst results in American polling history.
Barabas isn’t totally sure what happened. But he and his students do have one theory: their poll’s name. Trust in higher education is polarized on political lines. Under this theory, Trump-voting New Hampshirites saw a postcard from Dartmouth, an Ivy League school with a mostly liberal faculty and student body, and didn’t respond—whereas anti-Trump voters in the state leaped at the opportunity to answer mail from their favorite institution. The Dartmouth Poll is an extreme example, but the same thing is happening basically everywhere: People who take surveys are people who have more trust in institutions, and people who have more trust in institutions are less likely to vote for Trump.
Once a pollster wraps their head around this point, their options become slim. They could pay poll respondents in order to reach people who wouldn’t otherwise be inclined to answer. The New York Times tried this in collaboration with the polling firm Ipsos, paying up to $25 to each respondent. They found that they reached more moderate voters who usually don’t answer the phone and who were more likely to vote for Trump, but said the differences were “relatively small.”
Or pollsters can get more creative with their weights. Jesse Stinebring, a co-founder of the Democratic polling firm Blue Rose Research, told me that his company asks whether respondents “believe that sometimes a child needs a good hard spanking”—a belief disproportionately held by the type of American who doesn’t respond to surveys—and uses the answer alongside the usual weights.
Bailey, the Georgetown professor, has an even more out-there proposal. Say you run a poll with a 5 percent response rate that shows Harris winning by four points, and a second poll with a 35 percent response rate that shows her winning by one point. In that situation, Bailey says, you can infer that every 10 points of response rate increases Trump’s margin by one percentage point. So if the election has a 65 percent turnout rate, that should mean a two-point Trump victory. It’s “a new way of thinking,” Bailey admitted, in a bit of an understatement. But can you blame him?
To be clear, political polls can be valuable even if they underestimate Republicans by a few points. For example, Biden likely would have stayed in the 2024 race if polls hadn’t shown him losing to Trump by an insurmountable margin—one that was, in retrospect, almost certainly understated.
The problem is that people expect the most from polls when elections are close, but that is when polls are the least reliable, given the inevitability of error. And if the act of answering a survey, or engaging in politics at all, correlates so strongly with one side, then pollsters can only do so much.
The legendary Iowa pollster Ann Selzer has long hated the idea of baking your own assumptions into a poll, which is why she used weights for only a few variables, all demographic. For decades, this stubborn refusal to guess in advance earned her both accurate poll results and the adoration of those who study polling: In 2016, a 538 article called her “The Best Pollster in Politics.”
Selzer’s final poll of 2024 showed Harris leading Iowa by three percentage points. Three days later, Trump would win the state by 13 points, a stunning 16-point miss.
A few weeks after the election, Selzer released an investigation into what might have gone wrong. “To cut to the chase,” she concluded, “I found nothing to illuminate the miss.” The same day the analysis was published, she retired from election polling.
The post Polling Was Quietly Still Bad in 2024 appeared first on The Atlantic.