The Times Australia
Fisher and Paykel Appliances
The Times World News

.

Coroner finds social media contributed to 14-year-old Molly Russell’s death. How should parents and platforms react?

  • Written by Tama Leaver, Professor of Internet Studies, Curtin University
Coroner finds social media contributed to 14-year-old Molly Russell’s death. How should parents and platforms react?

Last week, London coroner Andrew Walker delivered his findings[1] from the inquest into 14-year-old schoolgirl Molly Russell’s death, concluding she “died from an act of self harm while suffering from depression and the negative effects of online content”.

The inquest heard Molly had used social media, specifically Instagram and Pinterest, to view large amounts of graphic content related to self-harm, depression and suicide in the lead-up to her death in November 2017.

The findings are a damning indictment of the big social media platforms. What should they be doing in response? And how should parents react in light of these events?

Social media use carries risk

The social media landscape of 2022 is different to the one Molly experienced in 2017. Indeed, the initial public outcry after her death saw many changes[2] to Instagram and other platforms to try and reduce material that glorifies depression or self-harm.

Instagram, for example, banned[3] graphic self-harm images, made it harder to search for non-graphic self-harm material, and started providing information about getting help when users made certain searches.

BBC journalist Tony Smith noted[4] that the press team for Instagram’s parent company Meta requested that journalists make clear these sorts of images are no longer hosted on its platforms. Yet Smith found some of this content was still readily accessible today.

Also, in recent years Instagram has[5] been[6] found[7] to host pro-anorexia accounts and content encouraging eating disorders. So although platforms may have made some positive changes over time, risks still remain.

That said, banning social media content is not necessarily the best approach.

Read more: Instagram can make teens feel bad about their body, but parents can help. Here's how[8]

What can parents do?

Here are some ways parents can address concerns about their children’s social media use.

Open a door for conversation, and keep it open

It’s not always easy to get young people to open up about what they’re feeling, but it’s clearly important to make it as easy and safe as possible for them to do so.

Research has shown[9] creating a non-judgemental space for young people to talk about how social media makes them feel will encourage them to reach out if they need help. Also, parents and young people can often learn from each other through talking about their online experiences.

Try not to overreact

Social media can be an important, positive and healthy part[10] of a young person’s life. It is where their peers and social groups are found, and during lockdowns was the only way many young people could support and talk to each other[11].

Completely banning social media may prevent young people from being a part of their peer groups, and could easily do more harm than good[12].

Negotiate boundaries together

Parents and young people can agree on reasonable rules for device and social media use. And such agreements can be very powerful.

They also present opportunities for parents and carers to model positive behaviours. For example, both parties might reach an agreement to not bring their devices to the dinner table, and focus on having conversations instead.

Another agreement might be to charge devices in a different room overnight so they can’t be used during normal sleep times.

What should social media platforms do?

Social media platforms have long faced a crisis of trust and credibility. Coroner Walker’s findings tarnish their reputation even further.

Now’s the time for platforms to acknowledge the risks present in the service they provide and make meaningful changes. That includes accepting regulation by governments.

More meaningful content moderation is needed

During the pandemic, more and more content moderation was automated. Automated systems are great when things are black and white, which is why they’re great at spotting extreme violence or nudity. But self-harm material is often harder to classify, harder to moderate[13] and often depends on the context it’s viewed in.

For instance, a picture of a young person looking at the night sky, captioned “I just want to be one with the stars”, is innocuous in many contexts and likely wouldn’t be picked up by algorithmic moderation. But it could flag an interest in self-harm if it’s part of a wider pattern of viewing.

Human moderators do a better job determining this context, but this also depends on how they’re resourced and supported. As social media scholar Sarah Roberts writes in her book Behind the Screen[14], content moderators for big platforms often work in terrible conditions, viewing many pieces of troubling content per minute, and are often traumatised themselves.

If platforms want to prevent young people seeing harmful content, they’ll need to employ better-trained, better-supported and better-paid moderators.

Harm prevention should not be an afterthought

Following the inquest findings, the new Prince and Princess of Wales astutely tweeted “online safety for our children and young people needs to be a prerequisite, not an afterthought”.

For too long, platforms have raced to get more users, and have only dealt with harms once negative press attention became unavoidable. They have been left to self-regulate for too long.

The foundation[15] set up by Molly’s family is pushing hard for the UK’s Online Safety Bill[16] to be accepted into law. This bill seeks[17] to reduce the harmful content young people see, and make platforms more accountable for protecting them from certain harms. It’s a start, but there’s already more that could be done[18].

In Australia the eSafety Commissioner has pushed for Safety by Design[19], which aims to have protections built into platforms from the ground up.

If this article has raised issues for you, or if you’re concerned about someone you know, call Lifeline[20] on 13 11 14.

References

  1. ^ delivered his findings (www.insider.com)
  2. ^ many changes (www.theguardian.com)
  3. ^ banned (about.instagram.com)
  4. ^ noted (twitter.com)
  5. ^ has (www.bbc.com)
  6. ^ been (www.buzzfeednews.com)
  7. ^ found (www.techtransparencyproject.org)
  8. ^ Instagram can make teens feel bad about their body, but parents can help. Here's how (theconversation.com)
  9. ^ has shown (thefoodmedic.co.uk)
  10. ^ healthy part (www.danah.org)
  11. ^ each other (psyarxiv.com)
  12. ^ harm than good (www.washingtonpost.com)
  13. ^ harder to moderate (www.wired.com)
  14. ^ Behind the Screen (yalebooks.yale.edu)
  15. ^ foundation (mollyrosefoundation.org)
  16. ^ Online Safety Bill (assets.publishing.service.gov.uk)
  17. ^ seeks (www.gov.uk)
  18. ^ more that could be done (5rightsfoundation.com)
  19. ^ Safety by Design (www.esafety.gov.au)
  20. ^ Lifeline (www.lifeline.org.au)

Read more https://theconversation.com/coroner-finds-social-media-contributed-to-14-year-old-molly-russells-death-how-should-parents-and-platforms-react-191757

Times Magazine

Can bigger-is-better ‘scaling laws’ keep AI improving forever? History says we can’t be too sure

OpenAI chief executive Sam Altman – perhaps the most prominent face of the artificial intellig...

A backlash against AI imagery in ads may have begun as brands promote ‘human-made’

In a wave of new ads, brands like Heineken, Polaroid and Cadbury have started hating on artifici...

Home batteries now four times the size as new installers enter the market

Australians are investing in larger home battery set ups than ever before with data showing the ...

Q&A with Freya Alexander – the young artist transforming co-working spaces into creative galleries

As the current Artist in Residence at Hub Australia, Freya Alexander is bringing colour and creativi...

This Christmas, Give the Navman Gift That Never Stops Giving – Safety

Protect your loved one’s drives with a Navman Dash Cam.  This Christmas don’t just give – prote...

Yoto now available in Kmart and The Memo, bringing screen-free storytelling to Australian families

Yoto, the kids’ audio platform inspiring creativity and imagination around the world, has launched i...

The Times Features

Here’s what new debt-to-income home loan caps mean for banks and borrowers

For the first time ever, the Australian banking regulator has announced it will impose new debt-...

Why the Mortgage Industry Needs More Women (And What We're Actually Doing About It)

I've been in fintech and the mortgage industry for about a year and a half now. My background is i...

Inflation jumps in October, adding to pressure on government to make budget savings

Annual inflation rose[1] to a 16-month high of 3.8% in October, adding to pressure on the govern...

Transforming Addiction Treatment Marketing Across Australasia & Southeast Asia

In a competitive and highly regulated space like addiction treatment, standing out online is no sm...

Aiper Scuba X1 Robotic Pool Cleaner Review: Powerful Cleaning, Smart Design

If you’re anything like me, the dream is a pool that always looks swimmable without you having to ha...

YepAI Emerges as AI Dark Horse, Launches V3 SuperAgent to Revolutionize E-commerce

November 24, 2025 – YepAI today announced the launch of its V3 SuperAgent, an enhanced AI platf...

What SMEs Should Look For When Choosing a Shared Office in 2026

Small and medium-sized enterprises remain the backbone of Australia’s economy. As of mid-2024, sma...

Anthony Albanese Probably Won’t Lead Labor Into the Next Federal Election — So Who Will?

As Australia edges closer to the next federal election, a quiet but unmistakable shift is rippli...

Top doctors tip into AI medtech capital raise a second time as Aussie start up expands globally

Medow Health AI, an Australian start up developing AI native tools for specialist doctors to  auto...