Skip to content
Search

Latest Stories

Top Stories

AI shouldn’t scare us – but fearmongering should

OpenAI logo on a screen
NurPhoto/Getty Images

Lee is a public interest technologist and researcher in the Boston area, and public voices fellow with The OpEd Project.

The company behind ChatGPT, OpenAI, recently started investigating claims that its artificial intelligence platform is getting lazier. Such shortcomings are a far cry from the firing and rehiring saga of OpenAI’s CEO, Sam Altman, last month. Pundits speculated that Altman’s initial ousting was due to a project called Q*, which – unlike ChatGPT – was able to solve grade-school arithmetic. Q* was seen as a step towards artificial general intelligence (AGI) and therefore a possible existential threat to humanity. I disagree.


As a technologist who has published research employing Q-learning and worked under one of its pioneers, I was dumbfounded to scroll through dozens of these outrageous takes. Q-learning, a decades-old algorithm belonging to a branch of AI known as “ reinforcement learning (RL),” is not new and is certainly not going to lead to the total destruction of humankind. Saying so is disingenuous and dangerous. The ability for Q* to solve elementary school equations says more about ChatGPT’s inability to do so than its supposedly fearsome capabilities – which are on par with a calculator. Like the proverbial monster under the bed, humanity’s real threat is not AI – it’s the fearmongering around it.

The supposed existential threat of AI is rooted in the assumption that AI systems will become conscious and superintelligent – i.e., that AI will become AGI. A fringe theory then claims a conscious, superintelligent AGI could, either through malevolence or by accident, kill us all. Proponents of this extreme view, who use an extreme extension of utilitarianism known as longtermism, claim our ultimate imperative is thus to prevent “ extinction threats ” like AGI in order to prevent the total annihilation of humanity. If this sounds like a stretch of the imagination, it is.

This AI doomerism, espoused by people like OpenAI’s now former interim CEO, Emmett Shear, assumes that AGI is even a likely scenario. But as someone who has conducted research on cognition for over a decade, I’m not worried AI will become sentient. And AI experts, including one of the pioneers, agree. A chasm remains that cannot be bridged between human-like performance and human-like understanding. Even if an AI system appears to produce human-like behavior, copying is not comprehension – a speaking parrot is still a parrot. Further, there are still many tasks requiring abstraction where even state-of-the-art AI models fall well short of human performance, and many aspects of human cognition that remain ineffable, like consciousness.

Heeding false alarms over killer AGI has real-world, present-day consequences. It shifts otherwise valuable research priorities, avoids accountability for present harms, and distracts legislators from pushing for real solutions. Billions of dollars, university departments and whole companies have now pivoted to “AI safety.” By focusing on hypothetical threats, we forgo real threats like climate change, ironically likely sped up by the massive amounts of water used by servers running AI models. We ignore the ways marginalized communities are currently harmed by AI systems like automated hiring and predictive policing. We forget about ways to address these harms, like passing legislation to regulate tech companies and AI. And we entrench the power of the tech industry by focusing on its chosen solution and excusing it from culpability for these harms.

When it comes to the mysterious Q*, I’m sure the addition of Q-learning will improve ChatGPT’s performance. After all, an ongoing line of research, thankfully less over-hyped, already exists to use RL to improve large language models like ChatGPT, called reinforcement learning with human feedback. And a decade ago, RL already helped train AI systems to play Atari and beat the world champion of Go. These accomplishments were impressive, but are engineering feats. At the end of the day, it’s precisely the current impacts of human-engineered systems that we need to worry about. The threats are not in the future, they’re in the now.

In “The Wizard of Oz,” the protagonists are awed by the powerful Oz, an intimidating mystical figure that towers over them physically and metaphorically throughout their journey. Much later, the ruse is revealed: The much-feared wizard was simply a small, old man operating a set of cranks and levers.

Don’t let the doomers distract you. Q-learning, as with the rest of AI, is not a fearful, mystical being – it’s just an equation set in code, written by humans. Tech CEOs would like you to buy into their faulty math and not the real implications of their current AI products. But their logic doesn’t add up. Instead, we urgently need to tackle real problems by regulating the tech industry, protecting people from AI technologies like facial recognition and providing meaningful redress from AI harms. That is what we really owe the future.

Read More

Media criticism
News media's vital to democracy, Americans say; then a partisan divide yawns
Tero Vesalainen/Getty Images

Public Media Under Fire: Why Project 2025 Is Reshaping NPR and PBS

This past spring and summer, The Fulcrum published a 30-part, nonpartisan series examining Project 2025—a sweeping policy blueprint for a potential second Trump administration. Our analysis explored the proposed reforms and their far-reaching implications across government. Now, as the 2025 administration begins to take shape, it’s time to move from speculation to reality.

In this follow-up, we turn our focus to one of the most consequential—and quietly unfolding—chapters of that blueprint: Funding cuts from NPR and PBS.

Keep ReadingShow less
Medical Schools Are Falling Behind in the Age of Generative AI

"To prepare tomorrow’s doctors, medical school deans, elected officials, and health care regulators must invest in training that matches the pace and promise of this technology," writes Dr. Robert Pearl.

Getty Images, ArtistGNDphotography

Medical Schools Are Falling Behind in the Age of Generative AI

While colleges across the nation are adapting their curricula to harness the power of generative AI, U.S. medical schools remain dangerously behind.

Most students entering medicine today will graduate without ever being trained to use GenAI tools effectively. That must change. To prepare tomorrow’s doctors – and protect tomorrow’s patients – medical school deans, elected officials, and health care regulators must invest in training that matches the pace and promise of this technology.

Keep ReadingShow less
Bay Area Social Media Post Claims ICE Cannot Enter Library, Fuels Misinformation

South Novato Library, California

Pricila Flores

Bay Area Social Media Post Claims ICE Cannot Enter Library, Fuels Misinformation

Bay Area community advocates are cautioning community members to be wary of what they see, interact with, and post on social media regarding information about the United States Immigration and Customs Enforcement (ICE) and immigration, following a rumor that targeted the Marin County Library.

‘South Novato Library has safe rooms that cannot be accessed by border patrol or ICE without a court order,’ an Instagram story post reads, with photos of a room in the library next to the text alongside the library address. The graphic claims Immigration and Customs Enforcement would not have the right to enter the pictured room without a court-ordered warrant.

Despite the graphic becoming a popular share among the local community of Novato, a Marin County city located just north of San Francisco, the information is false.

Keep ReadingShow less
AI Progress Delayed Is Progress Denied
Students in a college classroom.
Getty Images, Klaus Vedfelt

AI Progress Delayed Is Progress Denied

Earlier this summer, I recorded an episode of the Scaling Laws podcast with MacKenzie Price, founder of Alpha Schools—schools “where kids crush academics in two hours, build life skills through workshops, and thrive beyond the classroom.” The secret is AI, but likely not the sort of AI that comes to mind.

Students at Alpha Schools work with “adaptive AI” that allows 1:1 learning at the pace necessary to master a subject, moving at the speed of the student’s learning rather than that of the entire class. By relying on AI to set that tempo, the school shaves hours off the traditional classroom model and reallocates that time to activities that allow students to more fully explore their interests, from horseback riding to documentary filmmaking.

Keep ReadingShow less