Skip to content
Search

Latest Stories

Top Stories

AI shouldn’t scare us – but fearmongering should

OpenAI logo on a screen
NurPhoto/Getty Images

Lee is a public interest technologist and researcher in the Boston area, and public voices fellow with The OpEd Project.

The company behind ChatGPT, OpenAI, recently started investigating claims that its artificial intelligence platform is getting lazier. Such shortcomings are a far cry from the firing and rehiring saga of OpenAI’s CEO, Sam Altman, last month. Pundits speculated that Altman’s initial ousting was due to a project called Q*, which – unlike ChatGPT – was able to solve grade-school arithmetic. Q* was seen as a step towards artificial general intelligence (AGI) and therefore a possible existential threat to humanity. I disagree.


As a technologist who has published research employing Q-learning and worked under one of its pioneers, I was dumbfounded to scroll through dozens of these outrageous takes. Q-learning, a decades-old algorithm belonging to a branch of AI known as “reinforcement learning (RL),” is not new and is certainly not going to lead to the total destruction of humankind. Saying so is disingenuous and dangerous. The ability for Q* to solve elementary school equations says more about ChatGPT’s inability to do so than its supposedly fearsome capabilities – which are on par with a calculator. Like the proverbial monster under the bed, humanity’s real threat is not AI – it’s the fearmongering around it.

Sign up for The Fulcrum newsletter

The supposed existential threat of AI is rooted in the assumption that AI systems will become conscious and superintelligent – i.e., that AI will become AGI. A fringe theory then claims a conscious, superintelligent AGI could, either through malevolence or by accident, kill us all. Proponents of this extreme view, who use an extreme extension of utilitarianism known as longtermism, claim our ultimate imperative is thus to prevent “extinction threats” like AGI in order to prevent the total annihilation of humanity. If this sounds like a stretch of the imagination, it is.

This AI doomerism, espoused by people like OpenAI’s now former interim CEO, Emmett Shear, assumes that AGI is even a likely scenario. But as someone who has conducted research on cognition for over a decade, I’m not worried AI will become sentient. And AI experts, including one of the pioneers, agree. A chasm remains that cannot be bridged between human-like performance and human-like understanding. Even if an AI system appears to produce human-like behavior, copying is not comprehension – a speaking parrot is still a parrot. Further, there are still many tasks requiring abstraction where even state-of-the-art AI models fall well short of human performance, and many aspects of human cognition that remain ineffable, like consciousness.

Heeding false alarms over killer AGI has real-world, present-day consequences. It shifts otherwise valuable research priorities, avoids accountability for present harms, and distracts legislators from pushing for real solutions. Billions of dollars, university departments and whole companies have now pivoted to “AI safety.” By focusing on hypothetical threats, we forgo real threats like climate change, ironically likely sped up by the massive amounts of water used by servers running AI models. We ignore the ways marginalized communities are currently harmed by AI systems like automated hiring and predictive policing. We forget about ways to address these harms, like passing legislation to regulate tech companies and AI. And we entrench the power of the tech industry by focusing on its chosen solution and excusing it from culpability for these harms.

When it comes to the mysterious Q*, I’m sure the addition of Q-learning will improve ChatGPT’s performance. After all, an ongoing line of research, thankfully less over-hyped, already exists to use RL to improve large language models like ChatGPT, called reinforcement learning with human feedback. And a decade ago, RL already helped train AI systems to play Atari and beat the world champion of Go. These accomplishments were impressive, but are engineering feats. At the end of the day, it’s precisely the current impacts of human-engineered systems that we need to worry about. The threats are not in the future, they’re in the now.

In “The Wizard of Oz,” the protagonists are awed by the powerful Oz, an intimidating mystical figure that towers over them physically and metaphorically throughout their journey. Much later, the ruse is revealed: The much-feared wizard was simply a small, old man operating a set of cranks and levers.

Don’t let the doomers distract you. Q-learning, as with the rest of AI, is not a fearful, mystical being – it’s just an equation set in code, written by humans. Tech CEOs would like you to buy into their faulty math and not the real implications of their current AI products. But their logic doesn’t add up. Instead, we urgently need to tackle real problems by regulating the tech industry, protecting people from AI technologies like facial recognition and providing meaningful redress from AI harms. That is what we really owe the future.

Read More

Dictionary entry for "democracy"
Lobro78.Getty Images

Paving the path forward to strengthening democracy

Kristina Becvar and David L. Nevins, co-publishers of The Fulcrum, announced recently that effective Jan. 1, Hugo Balta, The Fulcrum’s director of solutions journalism and DEI initiatives, will serve as executive editor. What follows is a message from Balta about his new responsibility.

In the aftermath of this year’s contentious presidential election, it is imperative to heal a democracy fractured by polarization, emphasizing the importance of dialogue, accountability, and inclusive and transparent governance.

Journalism plays a pivotal role in upholding democratic values and ensuring the health of democratic systems. As our country faces complex challenges, the significance of a free and independent press becomes increasingly evident.

Keep ReadingShow less
Hugo Balta

The Fulcrum's new executive editor: Hugo Balta

As co-publishers of The Fulcrum, we are proud to announce that, effective Jan. 1, Hugo Balta, The Fulcrum’s director of solutions journalism and DEI initiatives, will serve as executive editor.

Hugo is an award-winning, 30-year multimedia journalism veteran with multiple market and platform experience, including leadership positions in NBC, Telemundo, ABC, CBS, and PBS, among other storied news networks. A nationally recognized diversity in journalism advocate, he is the recipient of the 2024 Cecilia Vaisman Award from Northwestern University Medill School of Journalism, Media, Integrated Marketing Communications. Hugo is the only person to serve twice as president of the National Association of Hispanic Journalists. Hugo and his family live in Chicago.

Keep ReadingShow less
Cell phone showing logos of Google, Amazon, Meta, Apple and Microsfot
Jaque Silva/NurPhoto via Getty Images

Big Tech is suppressing industrial liberty

This is the second entry in “Big Tech and Democracy,” a series designed to assist American citizens in understanding the impact technology is having — and will have — on our democracy. The series explores the benefits and risks that lie ahead and offers possible solutions.

Industrial liberty — once a cornerstone of American antitrust policy — has faded into obscurity in the shadow of Big Tech’s overwhelming dominance. In short, industrial liberty refers to your ability to use and benefit from your skills, your knowledge and your passion. It manifests as entrepreneurs and small-business owners, through patents and innovations, and as everyday folks finding good work every day. This erosion of this specific sort of liberty not only undermines the principles of competition but also stifles the aspirational spirit that has for so long distinguished the American public.

Keep ReadingShow less
People standing outside the Capitol

Dozens of members of Congress have had their likeness used in nonconsensual intimate imagery, otherwise known as deepfake porn. The majority of those impacted are women.

Kent Nishimura/Getty Images

AI enters Congress: Sexually explicit deepfakes target women lawmakers

Originally published by The 19th.

More than two dozen members of Congress have been the victims of sexually explicit deepfakes — and an overwhelming majority of those impacted are women, according to a new study that spotlights the stark gender disparity in this technology and the evolving risks for women’s participation in politics and other forms of civic engagement.

Keep ReadingShow less