Teens in Trial to Limit Social Media Use: A Shift Towards Real-life Interaction Daniel de la Hoz/Getty Images
A groundbreaking study is exploring the effects of reduced social media usage on teens’ mental health and well-being. While results are not expected until mid-2027, ongoing discussions suggest that some governments might institute bans on social media for teenagers before the outcomes are known.
The merit of such a ban is still up for debate in the courts. Despite limited evidence, Australia has introduced regulations for minors under 16, and the UK government is considering similar measures.
This trial prioritizes young people’s voices by involving them in the planning process. Historically, children and adolescents have been excluded from critical discussions concerning social media design and management.
“Involving kids is crucial,” states Pete Etchells from Bath Spa University, UK, who is not directly involved in the study.
“There is ample evidence pointing to the potential harms of social media on young users, some of which can be severe,” notes Amy Orben, co-leader of the trial, emphasizing the uncertainty regarding the broader impact of social media time.
To obtain clearer answers, large-scale studies are necessary. The IRL trial takes place in Bradford, England, aiming to recruit around 4,000 participants aged 12 to 15 across 10 schools. A bespoke app will be used to monitor social media engagement.
Half of the participants will face specific time limits on certain apps like TikTok, Instagram, and YouTube, with no restrictions on messaging apps like WhatsApp. “Total usage will be capped at one hour a day, with a curfew from 9 PM to 7 AM,” explains Dan Lewar from the Bradford Health Data Science Center, who co-leads the trial. This is significant, considering that the average social media usage for this age group is about three hours daily.
Importantly, participants will be randomized by grade level, allowing 8th graders to serve as the control group while 9th graders undergo restrictions. The aim is to create similar circumstances for both groups. “If a child’s social media is restricted, but their friends are active online post-curfew, they may feel excluded,” Orben explains.
Lewar emphasizes that the trial was designed collaboratively with teens. “They opposed a blanket ban,” he notes.
The comprehensive study will span six weeks around October, with preliminary results anticipated in mid-2027.
Orben emphasizes that this trial will yield more precise data on teenage social media habits through app monitoring rather than relying on self-reported information. The team will also gather data on anxiety, sleep quality, socializing, happiness, body image, school absenteeism, and experiences of bullying.
Etchells asserts the necessity of understanding whether restrictions or bans are beneficial or detrimental to youth. “The honest answer is we don’t know. That’s why research like this is critical.”
This initiative is welcomed due to the absence of high-quality studies in this area. A recent report from the UK Department for Science, Innovation, and Technology highlighted the need for quality causal evidence linking young people’s mental health to digital technology use, especially concerning social media, smartphones, and AI chatbots.
As stated by Margarita Panayiotou from the University of Manchester, engaging with youth is essential in social media research. Her findings show that teens often find ways to circumvent outright bans, making testing restrictions a more viable option. This approach may also be more ethical, as the harm caused by a ban is not yet understood.
“Teens view social media as a space for self-discovery,” says Panayiotou, highlighting concerns about platform distrust, feelings of loss of control, and unintentional overuse. They also report struggles with online judgment, body comparisons, and cyberbullying.
According to Etchells and Panayiotou, the primary challenge for governments is to compel tech companies to ensure safer social media environments for youth.
The Online Safety Act 2023 (OSA) mandates that technology firms like TikTok, Facebook, WhatsApp, and Instagram (owned by Meta), as well as Google (which owns YouTube), enhance user safety. “Effective enforcement of OSA could address many existing issues,” asserts Etchells.
Topics:
- Mental Health/
- Social Media
Source: www.newscientist.com

