Skip to content
Search

Latest Stories

Follow Us:
Top Stories

Facebook’s own internal documents offer a blueprint for making social media safer for teens

Teenager on social media
Dan Kenyon/Getty Images

Twenge is a p rofessor of psychology at San Diego State University.

Right at the time social media became popular, teen mental health began to falter. Between 2010 and 2019, rates of depression and loneliness doubled in the U.S. and globally, suicide rates soared for teens in the U.S. and emergency room admissions for self-harm tripled among U.S. 10- to 14-year-old girls. Social scientists like myself have been warning for years that the ubiquity of social media might be at the root of the growing mental health crisis for teens.

Yet when Facebook CEO Mark Zuckerberg was asked during a congressional hearing in March to acknowledge the connection between social media and these troubling mental health trends, he replied, "I don't think that the research is conclusive on that."

Just six months later, The Wall Street Journal reported that Facebook had been doing its own research for years on the negative effects of Instagram, the company's photo-sharing app popular with teens and young adults. Six internal documents summarizing the research, leaked by a whistle-blower, were posted in full on Sept. 29, 2021.

The details in the 209 pages are revealing. They suggest not only that Facebook knew how Instagram could be harmful, but that the company also was aware of possible solutions to mitigate those harms. Facebook's own research strongly suggests that social media should be subject to more stringent regulation and include more guardrails to protect the mental health of its users.


There are two primary ways the company can do this: enforcing time limits and increasing the minimum age of users.

A ticking time bomb for mental health

Academic research shows that the more hours a day a teen spends on social media, the more likely she or he is to be depressed or to self-harm.

That's important because many teens, especially girls, spend large amounts of time on social media.

One study in the U.K. found that one-quarter of 15-year-old girls spent more than five hours a day using social media – and 38% of those girls were clinically depressed. Comparatively, among girls who used social media less than one hour a day, only 15% were depressed.

Although the internal Facebook research didn't examine links between time on Instagram and mental health, they did ask teens about what were, in their view, the worst aspects of Instagram. One of the things teens disliked the most about the app was how much time they spent on it.

Teens, the report said, had "an addict's narrative about their use. … They wish they could spend less time caring about it but they can't help themselves."

They knew they were spending too much time online, but had a hard time controlling how much time they spent. One-third of teens suggested Instagram should remind them to take a break or encourage them to get off the app.

That would be a step in the right direction, but simple nudges might not be enough to get teens to close the app and keep it closed. And while parents can already set time limits using the parental controls included on most smartphones, many of them don't know how to use these controls or are unaware how much time teens are spending on social media.

So better regulations might need to put teeth into time limits, such as limiting the number of hours teens under 18 can spend on social media apps. A blackout period overnight might also be useful, as many teens use their smartphones at night when they should be sleeping.

ID, please

One internal Facebook study of more than 50,000 people from 10 countries found that half of teen girls compare their appearance to others' on Instagram. Those appearance-based comparisons, the study found, peaked when users were 13 to 18 and were much less common among adult women.

This is key, as body image issues seem to be one of the biggest reasons why social media use is linked to depression among teen girls. It also dovetails with research I reported in my book, "iGen," finding that social media use is more strongly linked to unhappiness among younger teens than older ones.

This suggests another avenue for regulation: age minimums. A 1998 law called the Children's Online Privacy Protection Rule already sets the age minimum for social media accounts at 13. That limit is problematic for two reasons. First, 13 is a developmentally challenging time, right as boys and girls are going through puberty and bullying is at its peak.

Second, the age minimum is not regularly enforced. Kids 12 and under can simply lie about their age to sign up for an account, and they're rarely kicked off the platform for being underage. During a Facebook event with Instagram head Adam Mosseri, the young celebrity JoJo Siwa noted she had been using Instagram since she was 8 years old, forcing Mosseri to acknowledge that it's easy to lie about your age.

The problem is how to enforce an age limit online for a population that is too young for IDs. Raising the minimum age to create a social media account to 16, 17 or 18 could solve two problems at once: It would prevent kids from signing up until they're a bit more developed and mature, and it would be easier to enforce. For example, potential users might be asked to submit a photo of their state-issued ID, which most teens have by 16.

Verifying age would also make it easier to construct a safer app for younger users that might, say, hide follower counts or restrict access to celebrity accounts, both of which Facebook's research found negatively impacted girls' body images.

Curtailing that fear of missing out

It's tempting to think regulations like these would cause teens to riot in the streets – after all, they love keeping up with their friends on social media. But the teens interviewed by Facebook for its internal research were well aware of social media's downsides.

"The reason why our generation is so messed up and has higher anxiety and depression than our parents' is because we have to deal with social media. Everyone feels like they have to be perfect," one teen girl told the researchers. Other teens have spoken publicly about the negative effects of social media.

More stringent regulation would help with another issue teens know all too well: the unwritten mandate to use social media or be left out.

"Young people are acutely aware that Instagram can be bad for their mental health yet are compelled to spend time on the app for fear of missing out," Facebook's internal research concluded.

If age limits were enforced, the peer pressure of being on social media would vanish; no or few classmates would be there. Regulating time on the app could also help if teens knew their friends wouldn't constantly be online.

Facebook's research demonstrates something else: The company was aware of the issues with Instagram but chose not to set these limits itself. Congress is now considering taking action.

Until they do, it will be up to parents and teens themselves to set limits. That won't be easy, but teens will be safer for it.

This article is republished from The Conversation under a Creative Commons license. Click here to read the original article.

The Conversation


Read More

The robot arm is assembling the word AI, Artificial Intelligence. 3D illustration

AI has the potential to transform education, mental health, and accessibility—but only if society actively shapes its use. Explore how community-driven norms, better data, and open experimentation can unlock better AI.

Getty Images, sarawuth702

Build Better AI

Something I think just about all of us agree on: we want better AI. Regardless of your current perspective on AI, it's undeniable that, like any other tool, it can unleash human flourishing. There's progress to be made with AI that we should all applaud and aim to make happen as soon as possible.

There are kids in rural communities who stand to benefit from AI tutors. There are visually impaired individuals who can more easily navigate the world with AI wearables. There are folks struggling with mental health issues who lack access to therapists who are in need of guidance during trying moments. A key barrier to leveraging AI "for good" is our imagination—because in many domains, we've become accustomed to an unacceptable status quo. That's the real comparison. The alternative to AI isn't well-functioning systems that are efficiently and effectively operating for everyone.

Keep ReadingShow less
Government Cyber Security Breach

An urgent look at the risks of unregulated artificial intelligence—from job loss and environmental strain to national security threats—and the growing political battle to regulate AI in the United States.

Getty Images, Douglas Rissing

AI Has Put Humanity on the Ballot

AI may not be the only existential threat out there, but it is coming for us the fastest. When I started law school in 2022, AI could barely handle basic math, but by graduation, it could pass the bar exam. Instead of taking the bar myself, I rolled immediately into a Master of Laws in Global Business Law at Columbia, where I took classes like Regulation of the Digital Economy and Applied AI in Legal Practice. By the end of the program, managing partners were comparing using AI to working with a team of associates; the CEO of Anthropic is now warning that it will be more capable than everyone in less than two years.

AI is dangerous in ways we are just beginning to see. Data centers that power AI require vast amounts of water to keep the servers cool, but two-thirds are in places already facing high water stress, with researchers estimating that water needs could grow from 60 billion liters in 2022 to as high as 275 billion liters by 2028. By then, data centers’ share of U.S. electricity consumption could nearly triple.

Keep ReadingShow less
Posters are displayed next to Sen. Ted Cruz (R-TX) as he speaks at a news conference to unveil the Take It Down Act to protect victims against non-consensual intimate image abuse, on Capitol Hill on June 18, 2024 in Washington, DC.

A lawsuit against xAI over AI-generated deepfakes targeting teenage girls exposes a growing crisis in schools. As laws struggle to keep up, this story explores AI accountability, teen safety, and what educators and parents must do now.

Getty Images, Andrew Harnik

Deepfakes: The New Face of Cyberbullying and Why Parents, Schools, and Lawmakers Must Act

As a former teacher who worked in a high school when Snapchat was born, I witnessed the birth of sexting and its impact on teens. I recall asking a parent whether he was checking his daughter’s phone for inappropriate messages. His response was, “sometimes you just don’t want to know.” But the federal lawsuit filed last week against Elon Musk's xAI has put a national spotlight on AI-generated deepfakes and the teenage girls they target. Parents and teachers can’t ignore the crisis inside our schools.

AI Companies Built the Tool. The Grok Lawsuit Says They Own the Damage.

Whether the theory of French prosecutors–that Elon Musk deliberately allowed the sexualized image controversy to grow so that it would drive up activity on the platform and boost the company’s valuation–is true or not, when a company makes the decision to build a tool and knows that it can be weaponized but chooses to release it anyway, they are making a risk-based decision believing that they can act without consequence. The Grok lawsuit could make these types of business decisions much more costly.

Keep ReadingShow less
Sketch collage image of businessman it specialist coding programming app protection security website web isolated on drawing background.

Amazon’s court loss over Just Walk Out highlights a deeper issue: employers are increasingly collecting workers’ biometric data without meaningful consent. Explore the growing conflict between workplace surveillance, privacy rights, and outdated U.S. laws.

Getty Images, Deagreez

The Quiet Rise of Employee Surveillance

Amazon’s loss in court over its attempt to shield the source code behind its Just Walk Out technology is a small win for shoppers, but the bigger story is how employers are quietly collecting biometric data from their own workers.

From factories to Fortune 500 companies, employers are demanding fingerprints, palmprints, retinal scans, facial scans, or even voice prints. These biometric technologies are eroding the boundary between workplace oversight and employee autonomy, often without consent or meaningful regulation.

Keep ReadingShow less