An oracle erred badly. The most impressive results were turned in by a little-known company in Brazil. A nagging problem reemerged, and some media critics turned profane in their assessments.
So it went for pollsters in the 2024 presidential election. Their collective performance, while not stellar, was improved from that of four years earlier. Overall, polls signaled a close outcome in the race between former President Donald Trump and Vice President Kamala Harris.
That is what the election produced: a modest win for Trump.
With votes still being counted in California and a few other states more than a week after Election Day, Trump had received 50.1% of the popular vote to Harris’ 48.1%, a difference of 2 points. That margin was closer than Joe Biden’s win by 4.5 points over Trump in 2020. It was closer than Hillary Clinton’s popular vote victory in 2016, closer than Barack Obama’s wins in 2008 and 2012.
There were, moreover, no errors among national pollsters quite as dramatic as CNN’s estimate in 2020 that Biden led Trump by 12 points.
This time, CNN’s final national poll said the race was deadlocked – an outcome anticipated by six other pollsters, according to data compiled by RealClearPolitics.
Sign up for The Fulcrum newsletter
The most striking discrepancy this year was the Marist College poll, conducted for NPR and PBS. It estimated Harris held a 4-point lead nationally at campaign’s end.
‘Oracle’ of Iowa’s big miss
In any event, a sense lingered among critics that the Trump-Harris election had resulted in yet another polling embarrassment, another entry in the catalog of survey failures in presidential elections, which is the topic of my latest book, “Lost in a Gallup.”
Comedian Jon Stewart gave harsh voice to such sentiments, saying of pollsters on his late-night program on election night, “I don’t ever want to fucking hear from you again. Ever. … You don’t know shit about shit, and I don’t care for you.”
Megyn Kelly, a former Fox News host, also denounced pollsters, declaring on her podcast the day after the election: “Polling is a lie. They don’t know anything.”
Two factors seemed to encourage such derision – a widely discussed survey of Iowa voters released the weekend before the election and Trump’s sweep of the seven states where the outcome turned.
The Iowa poll injected shock and surprise into the campaign’s endgame, reporting that Harris had taken a 3-point lead in the state over Trump. The result was likened to a “bombshell” and its implications seemed clear: If Harris had opened a lead in a state with Iowa’s partisan profile, her prospects of winning elsewhere seemed strong, especially in the Great Lakes swing states of Wisconsin, Michigan and Pennsylvania.
The survey was conducted for the Des Moines Register by J. Ann Selzer, a veteran Iowa-based pollster with an outstanding reputation in opinion research. In a commentary in The New York Times in mid-September, Republican pollster Kristen Soltis Anderson declared Selzer “the oracle of Iowa.” Rachel Maddow of MSNBC praised Selzer’s polls before the election for their “uncanny predictive accuracy.” Ratings released in June by data guru Nate Silver gave Selzer’s polls an A-plus grade.
But this time, Selzer’s poll missed dramatically.
Trump carried Iowa by 13 points, meaning the poll was off by 16 points – a stunning divergence for an accomplished pollster.
“Even the mighty have been humbled” by Trump’s victory, the Times of London said of Selzer’s polling failure.
Selzer said afterward she will “be reviewing data from multiple sources with hopes of learning why that (discrepancy) happened.”
It is possible, other pollsters suggested, that Selzer’s reliance on telephone-based surveying contributed to the polling failure. “Phone polling alone … isn’t going to reach low-propensity voters or politically disengaged nonwhite men,” Tom Lubbock and James Johnson wrote in a commentary for The Wall Street Journal.
These days, few pollsters rely exclusively on the phone to conduct election surveys; many of them have opted for hybrid approaches that combine, for example, phone, text and online sampling techniques.
Surprise sweep of swing states
Trump’s sweep of the seven vigorously contested swing states surely contributed to perceptions that polls had misfired again.
According to RealClearPolitics, Harris held slender, end-of-campaign polling leads in Michigan and Wisconsin, while Trump was narrowly ahead in Arizona, Georgia, Pennsylvania, North Carolina and Nevada.
Trump won them all, an outcome no pollster anticipated – except for AtlasIntel of Sao Paulo, Brazil, a firm “about which little is known,” as The New Republic noted.
AtlasIntel estimated Trump was ahead in all seven swing states by margins that hewed closely to the voting outcomes. In none of the swing states did AtlasIntel’s polling deviate from the final vote tally by more than 1.3 points, an impressive performance.
AtlasIntel did not respond to email requests I sent requesting information about its background and polling technique. The company describes itself as “a leading innovator in online polling” and says it uses “a proprietary methodology,” without revealing much about it.
Its founder and chief executive is Andrei Roman, who earned a doctorate in government at Harvard University. Roman took to X, formerly Twitter, in the election’s aftermath to post a chart that touted AtlasIntel as “the most accurate pollster of the US Presidential Election.”
It was a burst of pollster braggadocio reminiscent of a kind that has emerged periodically since the 1940s. That was when polling pioneer George Gallup placed two-page advertising spreads in the journalism trade publication “Editor & Publisher” to assert the accuracy of his polls in presidential elections.
Underestimating Trump’s support again
A significant question facing pollsters this year – their great known unknown – was whether modifications made to sampling techniques would allow them to avoid underestimating Trump’s support, as they had in 2016 and 2020.
Misjudging Trump’s backing is a nagging problem for pollsters. The results of the 2024 election indicate that the shortcoming persists. By margins ranging from 0.9 points to 2.7 points, polls overall understated Trump’s support in the seven swing states, for example.
Some polls misjudged Trump’s backing by even greater margins. CNN, for example, underestimated Trump’s vote by 4.3 points in North Carolina, by more than 6 points in Michigan and Wisconsin as well as Arizona.
Results that misfire in the same direction suggest that adjustments to sampling methodologies were inadequate or ineffective for pollsters in seeking to reach Trump backers of all stripes.
Campbell, is professor emeritus of communication at the American University School of Communication.
This article is republished from The Conversation under a Creative Commons license. Read the original article.