Skip to content
Search

Latest Stories

Top Stories

Is AI too big to fail?

Nvidia building and logo

The world came to a near standstill last month as everyone awaited Nvidia’s financial outlook.

Cheng Xin/Getty Images

This is the first entry in “ Big Tech and Democracy,” a series designed to assist American citizens in understanding the impact technology is having — and will have — on our democracy. The series will explore the benefits and risks that lie ahead and offer possible solutions.

In the span of two or so years, OpenAI, Nvidia and a handful of other companies essential to the development of artificial intelligence have become economic behemoths. Their valuations and stock prices have soared. Their products have become essential to Fortune 500 companies. Their business plans are the focus of the national security industry. Their collapse would be, well, unacceptable. They are too big to fail.

The good news is we’ve been in similar situations before. The bad news is we’ve yet to really learn our lesson.


In the mid-1970s, a bank known for its conservative growth strategy decided to more aggressively pursue profits. The strategy worked. In just a few years the bank became the largest commercial and industrial lender in the nation. The impressive growth caught the attention of others — competitors looked on with envy, shareholders with appreciation and analysts with bullish optimism. As the balance sheet grew, however, so did the broader economic importance of the bank. It became too big to fail.

Regulators missed the signs of systemic risk. A kick of the bank’s tires gave no reason to panic. But a look under the hood — specifically, at the bank’s loan-to-assets ratio and average return on loans — would have revealed a simple truth: The bank had been far too risky. The tactics that fueled its “go-go” years rendered the bank over exposed to sectors suffering tough economic times. Rumors soon spread that the bank was in a financially sketchy spot. It was the Titanic, without the band, to paraphrase an employee.

When the inevitable run on the bank started, regulators had no choice but to spend billions on keeping the bank afloat — saving it from sinking and bringing the rest of the economy with it. Of course, a similar situation played out during the Great Recession — risky behavior by a few bad companies imposed bailout payments on the rest of us.

AI labs are similarly taking gambles that have good odds of making many of us losers. As major labs rush to release their latest models, they are not stopping to ask if we have the social safety nets ready if things backfire. Nor are they meaningfully contributing to building those necessary safeguards. Instead, we find ourselves in a highly volatile situation. Our stock market seemingly pivots on earnings of just a few companies — the world came to a near standstill last month as everyone awaited Nvidia’s financial outlook. Our leading businesses and essential government services are quick to adopt the latest AI models despite real uncertainty as to whether they will operate as intended. If any of these labs took a financial tumble or any of the models were significantly flawed, the public would likely again be asked to find a way to save the risk takers.

This outcome may be likely but it’s not inevitable. The Dodd Frank Act passed in response to the Great Recession and intended to prevent another Too Big to Fail situation in the financial sector has been roundly criticized for its inadequacy. We should learn from its faults in thinking through how to make sure AI goliaths don’t crush all of us Davids. Some sample steps include mandating and enforcing more rigorous testing of AI models before deployment. It would also behoove us to prevent excessive reliance on any one model by the government — this could be accomplished by requiring public service providers to maintain analog processes in the event of emergencies. Finally, we can reduce the economic sway of a few labs by fostering more competition in the space.

Too Big to Fail scenarios have happened on too many occasions. There’s no excuse for allowing AI labs to become so large and so essential that we collectively end up paying for their mistakes.

Frazier is an assistant professor at the Crump College of Law at St. Thomas University and a Tarbell fellow.

Read More

Fear of AI Makes for Bad Policy
Getty Images

Fear of AI Makes for Bad Policy

Fear is the worst possible response to AI. Actions taken out of fear are rarely a good thing, especially when it comes to emerging technology. Empirically-driven scrutiny, on the other hand, is a savvy and necessary reaction to technologies like AI that introduce great benefits and harms. The difference is allowing emotions to drive policy rather than ongoing and rigorous evaluation.

A few reminders of tech policy gone wrong, due, at least in part, to fear, helps make this point clear. Fear is what has led the US to become a laggard in nuclear energy, while many of our allies and adversaries enjoy cheaper, more reliable energy. Fear is what explains opposition to autonomous vehicles in some communities, while human drivers are responsible for 120 deaths per day, as of 2022. Fear is what sustains delays in making drones more broadly available, even though many other countries are tackling issues like rural access to key medicine via drones.

Keep ReadingShow less
A child looking at a smartphone.

With autism rates doubling every decade, scientists are reexamining environmental and behavioral factors. Could the explosion of social media use since the 1990s be influencing neurodevelopment? A closer look at the data, the risks, and what research must uncover next.

Getty Images, Arindam Ghosh

The Increase in Autism and Social Media – Coincidence or Causal?

Autism has been in the headlines recently because of controversy over Robert F. Kennedy, Jr's statements. But forgetting about Kennedy, autism is headline-worthy because of the huge increase in its incidence over the past two decades and its potential impact on not just the individual children but the health and strength of our country.

In the 1990s, a new definition of autism—ASD (Autism Spectrum Disorder)—was universally adopted. Initially, the prevalence rate was pretty stable. In the year 2,000, with this broader definition and better diagnosis, the CDC estimated that one in 150 eight-year-olds in the U.S. had an autism spectrum disorder. (The reports always study eight-year-olds, so this data was for children born in 1992.)

Keep ReadingShow less
Tech, Tribalism, and the Erosion of Human Connection
Ai technology, Artificial Intelligence. man using technology smart robot AI, artificial intelligence by enter command prompt for generates something, Futuristic technology transformation.
Getty Images - stock photo

Tech, Tribalism, and the Erosion of Human Connection

One of the great gifts of the Enlightenment age was the centrality of reason and empiricism as instruments to unleash the astonishing potential of human capacity. Great Enlightenment thinkers recognized that human beings have the capacity to observe the universe and rely on logical thinking to solve problems.

Moreover, these were not just lofty ideals; Benjamin Franklin and Denis Diderot demonstrated that building our collective constitution of knowledge could greatly enhance human prosperity not only for the aristocratic class but for all participants in the social contract. Franklin’s “Poor Richard’s Almanac” and Diderot and d’Alembert’s “Encyclopédie” served as the Enlightenment’s machines de guerre, effectively providing broad access to practical knowledge, empowering individuals to build their own unique brand of prosperity.

Keep ReadingShow less
The limits of free speech protections in American broadcasting

FCC Chairman Brendan Carr testifies in Washington on May 21, 2025.

The limits of free speech protections in American broadcasting

The chairman of the Federal Communications Commission is displeased with a broadcast network. He makes his displeasure clear in public speeches, interviews and congressional testimony.

The network, afraid of the regulatory agency’s power to license their owned-and-operated stations, responds quickly. They change the content of their broadcasts. Network executives understand the FCC’s criticism is supported by the White House, and the chairman implicitly represents the president.

Keep ReadingShow less