Skip to content
Search

Latest Stories

Follow Us:
Top Stories

The AI Race We Need: For a Better Future, Not Against Another Nation

Opinion

The AI Race We Need: For a Better Future, Not Against Another Nation

The concept of AI hovering among the public.

Getty Images, J Studios

The AI race that warrants the lion’s share of our attention and resources is not the one with China. Both superpowers should stop hurriedly pursuing AI advances for the sake of “beating” the other. We’ve seen such a race before. Both participants lose. The real race is against an unacceptable status quo: declining lifespans, increasing income inequality, intensifying climate chaos, and destabilizing politics. That status quo will drag on, absent the sorts of drastic improvements AI can bring about. AI may not solve those problems but it may accelerate our ability to improve collective well-being. That’s a race worth winning.

Geopolitical races have long sapped the U.S. of realizing a better future sooner. The U.S. squandered scarce resources and diverted talented staff to close the alleged missile gap with the USSR. President Dwight D. Eisenhower rightfully noted, “Every gun that is made, every warship launched, every rocket fired signifies, in the final sense, a theft from those who hunger and are not fed, those who are cold and are not clothed.” He realized that every race comes at an immense cost. In this case, the country was “spending the sweat of its laborers, the genius of its scientists, the hopes of its children.”


President John F. Kennedy failed to heed the guidance of his predecessor. He initiated yet another geopolitical contest by publicly challenging the USSR to a space race. Privately, he too knew that such a race required substantial trade-offs. Before Sputnik, Kennedy scoffed at spending precious funds on space endeavors. Following the Bay of Pigs Invasion, Kennedy reversed course. In his search for a political win, he found space. The rest, of course, is history. It’s true that the nation’s pursuit of the moon generated significant direct and indirect benefits. What’s unknowable, though, is what benefits could have been realized if Kennedy pursued his original science agenda: large-scale desalination of seawater. That bold endeavor would have also created spin-off improvements in related fields.

Decades from now, the true “winner” of the AI race will be the country that competes in the only race that really matters—tackling the most pressing economic, social, and political problems. The country that wins that race will have a richer, healthier, and more resilient population. That country will endure when crises unfold. Others will crumble.

AI development and deployment involve finite resources. The chips, energy, and expertise that go into creating leading AI models are in short supply. Chips accumulated by OpenAI, Anthropic, Google, and other massive AI labs to train the next frontier model are chips not being used to address more socially useful ends. Likewise, an AI expert working on a new AI-driven missile system is an expert not working on how AI can solve problems that have long been put on the back burner in the name of winning the geopolitical race of the moment.

Imagine the good that could come about if instead of prioritizing the pursuit of an unreachable AI frontier, we turned already impressive models toward the problems that will shape our long-term communal success. Early signs suggest that a pivot to this race would immediately improve the status quo. First, consider the potential for rapid improvements in health brought about by better, more affordable drugs. According to the Boston Consulting Group, AI-based discoveries or designs have spurred 67 clinical trials of new drugs. AstraZeneca reported that AI had cut its drug discovery process from years to months.

Second, consider the possibility of providing every student with personalized tutoring—setting us on a path to again become the most educated and productive workforce the world over. AI programs deployed in Bhutan helped students learn math skills in a fraction of the time when compared to classmates who received traditional math instruction. Closer to home, Khanmigo— an AI platform designed by the Khan Academy —is giving students personalized lessons in 266 school districts across the United States.

Third, and finally, consider a world in which traffic fatalities were halved thanks to the broader adoption of autonomous vehicles. Autonomous Vehicle (AV) companies have leveraged AI to make rapid advances in the ability of their vehicles to drive in all conditions. Further focus on these efforts may finally make AVs the majority of cars on the road and, as a result, save thousands of lives.

To redirect our AI race toward societal benefit, we need concrete policy changes. Federal research funding should prioritize AI applications targeting our most pressing challenges—healthcare access, energy development, and educational opportunities. Complementing this approach, tax incentives could reward companies that deploy AI for measurable social impact rather than pure market dominance. Additionally, public-private partnerships, similar to the one between Texas A&M and NVIDIA involving the creation of a high-performance supercomputer, could create innovation hubs focused specifically on using AI to solve regional problems, from drought management in the Southwest to infrastructure resilience on the coasts.

The choice before us is clear: we can continue the myopic pursuit of AI superiority for its own sake, or we can choose the wiser race—one toward a more innovative and prosperous future. History will not judge us by which nation first reached some arbitrary artificial intelligence threshold but by how we wielded this transformative technology to solve problems that have plagued humanity for generations. By redirecting our finite resources—chips, energy, and human ingenuity—toward these challenges, we can ensure that the true winners of the AI revolution will be all of us, not merely one flag or another. That is a victory worth pursuing with the full measure of our national commitment and creativity.


Kevin Frazier is an AI Innovation and Law Fellow at Texas Law and Author of the Appleseed AI substack.


Read More

Government Cyber Security Breach

An urgent look at the risks of unregulated artificial intelligence—from job loss and environmental strain to national security threats—and the growing political battle to regulate AI in the United States.

Getty Images, Douglas Rissing

AI Has Put Humanity on the Ballot

AI may not be the only existential threat out there, but it is coming for us the fastest. When I started law school in 2022, AI could barely handle basic math, but by graduation, it could pass the bar exam. Instead of taking the bar myself, I rolled immediately into a Master of Laws in Global Business Law at Columbia, where I took classes like Regulation of the Digital Economy and Applied AI in Legal Practice. By the end of the program, managing partners were comparing using AI to working with a team of associates; the CEO of Anthropic is now warning that it will be more capable than everyone in less than two years.

AI is dangerous in ways we are just beginning to see. Data centers that power AI require vast amounts of water to keep the servers cool, but two-thirds are in places already facing high water stress, with researchers estimating that water needs could grow from 60 billion liters in 2022 to as high as 275 billion liters by 2028. By then, data centers’ share of U.S. electricity consumption could nearly triple.

Keep ReadingShow less
Posters are displayed next to Sen. Ted Cruz (R-TX) as he speaks at a news conference to unveil the Take It Down Act to protect victims against non-consensual intimate image abuse, on Capitol Hill on June 18, 2024 in Washington, DC.

A lawsuit against xAI over AI-generated deepfakes targeting teenage girls exposes a growing crisis in schools. As laws struggle to keep up, this story explores AI accountability, teen safety, and what educators and parents must do now.

Getty Images, Andrew Harnik

Deepfakes: The New Face of Cyberbullying and Why Parents, Schools, and Lawmakers Must Act

As a former teacher who worked in a high school when Snapchat was born, I witnessed the birth of sexting and its impact on teens. I recall asking a parent whether he was checking his daughter’s phone for inappropriate messages. His response was, “sometimes you just don’t want to know.” But the federal lawsuit filed last week against Elon Musk's xAI has put a national spotlight on AI-generated deepfakes and the teenage girls they target. Parents and teachers can’t ignore the crisis inside our schools.

AI Companies Built the Tool. The Grok Lawsuit Says They Own the Damage.

Whether the theory of French prosecutors–that Elon Musk deliberately allowed the sexualized image controversy to grow so that it would drive up activity on the platform and boost the company’s valuation–is true or not, when a company makes the decision to build a tool and knows that it can be weaponized but chooses to release it anyway, they are making a risk-based decision believing that they can act without consequence. The Grok lawsuit could make these types of business decisions much more costly.

Keep ReadingShow less
Sketch collage image of businessman it specialist coding programming app protection security website web isolated on drawing background.

Amazon’s court loss over Just Walk Out highlights a deeper issue: employers are increasingly collecting workers’ biometric data without meaningful consent. Explore the growing conflict between workplace surveillance, privacy rights, and outdated U.S. laws.

Getty Images, Deagreez

The Quiet Rise of Employee Surveillance

Amazon’s loss in court over its attempt to shield the source code behind its Just Walk Out technology is a small win for shoppers, but the bigger story is how employers are quietly collecting biometric data from their own workers.

From factories to Fortune 500 companies, employers are demanding fingerprints, palmprints, retinal scans, facial scans, or even voice prints. These biometric technologies are eroding the boundary between workplace oversight and employee autonomy, often without consent or meaningful regulation.

Keep ReadingShow less
Close up of a woman wearing black, modern spectacles Smart glasses and reality concept with futuristic screen

Apple’s upcoming AI-powered wearables highlight growing privacy risks as the right to record police faces increasing threats. The death of Alex Pretti raises urgent questions about surveillance, civil liberties, and accountability in the digital age.

Getty Images, aislan13

AI Wearables and the Rising Risk of Recording Police

Last month, Apple announced the development of three wearable smart devices, all equipped with built-in cameras. The company has its sights set on 2027 for the release of their new smart glasses, AI pendant, and AirPods with built-in camera, all of which will be AI-functional for users. As the market for wearable products offering smart-recording capabilities expands, so does the risk that comes with how users choose to use the technology.

In Minneapolis in January, Alex Pretti was killed after an encounter with federal agents while filming them with his phone. He was not a suspect in a crime. He was not interfering, but was doing what millions of Americans now instinctively do when they see state power in motion: witnessing.

Keep ReadingShow less