Skip to content
Search

Latest Stories

Top Stories

Better but not stellar: Pollsters faced familiar complaints, difficulties in assessing Trump-Harris race

Hand holding a mobile phone showing CNN's "Magic Wall."

CNN’s Magic Wall map with U.S. presidential, seen on a mobile phone on Nov.

Beata Zawrzel/NurPhoto via Getty Images

An oracle erred badly. The most impressive results were turned in by a little-known company in Brazil. A nagging problem reemerged, and some media critics turned profane in their assessments.

So it went for pollsters in the 2024 presidential election. Their collective performance, while not stellar, was improved from that of four years earlier. Overall, polls signaled a close outcome in the race between former President Donald Trump and Vice President Kamala Harris.


That is what the election produced: a modest win for Trump.

With votes still being counted in California and a few other states more than a week after Election Day, Trump had received 50.1% of the popular vote to Harris’ 48.1%, a difference of 2 points. That margin was closer than Joe Biden’s win by 4.5 points over Trump in 2020. It was closer than Hillary Clinton’s popular vote victory in 2016, closer than Barack Obama’s wins in 2008 and 2012.

There were, moreover, no errors among national pollsters quite as dramatic as CNN’s estimate in 2020 that Biden led Trump by 12 points.

This time, CNN’s final national poll said the race was deadlocked – an outcome anticipated by six other pollsters, according to data compiled by RealClearPolitics.

Sign up for The Fulcrum newsletter

The most striking discrepancy this year was the Marist College poll, conducted for NPR and PBS. It estimated Harris held a 4-point lead nationally at campaign’s end.

‘Oracle’ of Iowa’s big miss

In any event, a sense lingered among critics that the Trump-Harris election had resulted in yet another polling embarrassment, another entry in the catalog of survey failures in presidential elections, which is the topic of my latest book, “Lost in a Gallup.”

Comedian Jon Stewart gave harsh voice to such sentiments, saying of pollsters on his late-night program on election night, “I don’t ever want to fucking hear from you again. Ever. … You don’t know shit about shit, and I don’t care for you.”

Megyn Kelly, a former Fox News host, also denounced pollsters, declaring on her podcast the day after the election: “Polling is a lie. They don’t know anything.”

Two factors seemed to encourage such derision – a widely discussed survey of Iowa voters released the weekend before the election and Trump’s sweep of the seven states where the outcome turned.

The Iowa poll injected shock and surprise into the campaign’s endgame, reporting that Harris had taken a 3-point lead in the state over Trump. The result was likened to a “bombshell” and its implications seemed clear: If Harris had opened a lead in a state with Iowa’s partisan profile, her prospects of winning elsewhere seemed strong, especially in the Great Lakes swing states of Wisconsin, Michigan and Pennsylvania.

The survey was conducted for the Des Moines Register by J. Ann Selzer, a veteran Iowa-based pollster with an outstanding reputation in opinion research. In a commentary in The New York Times in mid-September, Republican pollster Kristen Soltis Anderson declared Selzer “the oracle of Iowa.” Rachel Maddow of MSNBC praised Selzer’s polls before the election for their “uncanny predictive accuracy.” Ratings released in June by data guru Nate Silver gave Selzer’s polls an A-plus grade.

But this time, Selzer’s poll missed dramatically.

Trump carried Iowa by 13 points, meaning the poll was off by 16 points – a stunning divergence for an accomplished pollster.

“Even the mighty have been humbled” by Trump’s victory, the Times of London said of Selzer’s polling failure.

Selzer said afterward she will “be reviewing data from multiple sources with hopes of learning why that (discrepancy) happened.”

It is possible, other pollsters suggested, that Selzer’s reliance on telephone-based surveying contributed to the polling failure. “Phone polling alone … isn’t going to reach low-propensity voters or politically disengaged nonwhite men,” Tom Lubbock and James Johnson wrote in a commentary for The Wall Street Journal.

These days, few pollsters rely exclusively on the phone to conduct election surveys; many of them have opted for hybrid approaches that combine, for example, phone, text and online sampling techniques.

Surprise sweep of swing states

Trump’s sweep of the seven vigorously contested swing states surely contributed to perceptions that polls had misfired again.

According to RealClearPolitics, Harris held slender, end-of-campaign polling leads in Michigan and Wisconsin, while Trump was narrowly ahead in Arizona, Georgia, Pennsylvania, North Carolina and Nevada.

Trump won them all, an outcome no pollster anticipated – except for AtlasIntel of Sao Paulo, Brazil, a firm “about which little is known,” as The New Republic noted.

AtlasIntel estimated Trump was ahead in all seven swing states by margins that hewed closely to the voting outcomes. In none of the swing states did AtlasIntel’s polling deviate from the final vote tally by more than 1.3 points, an impressive performance.

AtlasIntel did not respond to email requests I sent requesting information about its background and polling technique. The company describes itself as “a leading innovator in online polling” and says it uses “a proprietary methodology,” without revealing much about it.

Its founder and chief executive is Andrei Roman, who earned a doctorate in government at Harvard University. Roman took to X, formerly Twitter, in the election’s aftermath to post a chart that touted AtlasIntel as “the most accurate pollster of the US Presidential Election.”

It was a burst of pollster braggadocio reminiscent of a kind that has emerged periodically since the 1940s. That was when polling pioneer George Gallup placed two-page advertising spreads in the journalism trade publication “Editor & Publisher” to assert the accuracy of his polls in presidential elections.

Underestimating Trump’s support again

A significant question facing pollsters this year – their great known unknown – was whether modifications made to sampling techniques would allow them to avoid underestimating Trump’s support, as they had in 2016 and 2020.

Misjudging Trump’s backing is a nagging problem for pollsters. The results of the 2024 election indicate that the shortcoming persists. By margins ranging from 0.9 points to 2.7 points, polls overall understated Trump’s support in the seven swing states, for example.

Some polls misjudged Trump’s backing by even greater margins. CNN, for example, underestimated Trump’s vote by 4.3 points in North Carolina, by more than 6 points in Michigan and Wisconsin as well as Arizona.

Results that misfire in the same direction suggest that adjustments to sampling methodologies were inadequate or ineffective for pollsters in seeking to reach Trump backers of all stripes.The Conversation

Campbell, is professor emeritus of communication at the American University School of Communication.

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Read More

People standing outside the Capitol

Dozens of members of Congress have had their likeness used in nonconsensual intimate imagery, otherwise known as deepfake porn. The majority of those impacted are women.

Kent Nishimura/Getty Images

AI enters Congress: Sexually explicit deepfakes target women lawmakers

Originally published by The 19th.

More than two dozen members of Congress have been the victims of sexually explicit deepfakes — and an overwhelming majority of those impacted are women, according to a new study that spotlights the stark gender disparity in this technology and the evolving risks for women’s participation in politics and other forms of civic engagement.

Keep ReadingShow less
Nvidia building and logo

The world came to a near standstill last month as everyone awaited Nvidia’s financial outlook.

Cheng Xin/Getty Images

Is AI too big to fail?

This is the first entry in “Big Tech and Democracy,” a series designed to assist American citizens in understanding the impact technology is having — and will have — on our democracy. The series will explore the benefits and risks that lie ahead and offer possible solutions.

In the span of two or so years, OpenAI, Nvidia and a handful of other companies essential to the development of artificial intelligence have become economic behemoths. Their valuations and stock prices have soared. Their products have become essential to Fortune 500 companies. Their business plans are the focus of the national security industry. Their collapse would be, well, unacceptable. They are too big to fail.

The good news is we’ve been in similar situations before. The bad news is we’ve yet to really learn our lesson.

Keep ReadingShow less

Berwyn Collaborative: Understanding Community Needs

“We have good people here, and if we have help highlighting our good people, we can connect more, collaborate more, be more creative, and resist harder,” said Berwyn resident Isabel Gonzalez Smith.

On a breezy November Saturday afternoon, members of the Cook County suburban city, had the opportunity to meet with local journalists and be heard at the Liberty Cultural Center in Berwyn, IL.

Keep ReadingShow less
disinformation spelled out
TolikoffPhotography/Getty Images

Listening in a time of disinformation

The very fabric of truth is unraveling at an alarming rate; Howard Thurman's wisdom about listening for the sound of the genuine is not just relevant but urgent. In the face of the escalating crisis of disinformation, distortion and the unsettling normalization of immoral and unethical practices, particularly in electoral politics and executive leadership, the need to cultivate the art of discernment and informed listening is more pressing than ever.
Keep ReadingShow less