Google AI
The Times Australia
The Times World News

.

Terrorist content lurks all over the internet – regulating only 6 major platforms won’t be nearly enough

  • Written by: Marten Risius, Senior Lecturer in Business Information Systems, The University of Queensland
Terrorist content lurks all over the internet – regulating only 6 major platforms won’t be nearly enough

Australia’s eSafety commissioner has sent legal notices[1] to Google, Meta, Telegram, WhatsApp, Reddit and X (formerly Twitter) asking them to show what they’re doing to protect Australians from online extremism. The six companies have 49 days to respond[2].

The notice comes at a time when governments are increasingly cracking down on major tech companies to address online harms like child sexual abuse material[3] or bullying[4].

Combating online extremism presents unique challenges different from other content moderation problems. Regulators wanting to establish effective and meaningful change must take into account what research has shown us about extremism and terrorism.

Extremists are everywhere

Online extremism and terrorism have been pressing concerns for some time. A stand-out example was the 2019 Christchurch terrorist attack on two mosques in Aotearoa New Zealand, which was live streamed on Facebook. It led to the “Christchurch Call” to action[5], aimed at countering extremism through collaborations between countries and tech companies.

But despite such efforts, extremists still use online platforms[6] for networking and coordination, recruitment and radicalisation, knowledge transfer, financing and mobilisation to action.

In fact, extremists use the same online infrastructure as everyday users: marketplaces, dating platforms, gaming sites, music streaming sites and social networks. Therefore, all regulation to counter extremism needs to consider the rights of regular users, as well.

Read more: Christchurch attacks 5 years on: terrorist’s online history gives clues to preventing future atrocities[7]

The rise of ‘swarmcasting’

Tech companies have responded with initiatives like the Global Internet Forum to Counter Terrorism[8]. It shares information on terrorist online content among its members (such as Facebook, Microsoft, YouTube, X and others) so they can take it down on their platforms. These approaches aim to automatically identify and remove[9] terrorist or extremist content.

However, a moderation policy focused on individual pieces of content on individual platforms fails to capture much of what’s out there.

Terrorist groups commonly use a “swarmcasting” multiplatform approach[10], leveraging 700 platforms or more to distribute their content.

Swarmcasting involves using “beacons” on major platforms such as Facebook, Twitter and Telegram to direct people to locations with terrorist material. This beacon can be a hyperlink to a blog post on a website like Wordpress or Tumblr that then contains further links to the content, perhaps hosted on Google Drive, JustPaste.It, BitChute and other places where users can download it.

So, while extremist content may be flagged and removed from social media, it remains accessible online thanks to swarmcasting.

Close-up of a mouse cursor hovering over a click here link on a computer screen.
Extremist content can be ‘hidden’ behind collections of hyperlinks to cloud sites or other hosts. Jakub Krechowicz/Shutterstock[11]

Putting up filters isn’t enough

The process of identifying and removing extremist content is far from simple. For example, at a recent US Supreme Court hearing over internet regulations, a lawyer argued[12] platforms could moderate terrorist content by simply removing anything that mentioned “al Qaeda”.

However, internationally recognised terrorist organisations, their members and supporters do not solely distribute policy-violating extremist content. Some may be discussing non-terrorist activities, such as those who engage in humanitarian efforts.

Other times their content is borderline (awful but lawful), such as misogynistic dog whistles, or even “hidden” in a different format[13], such as memes.

Accordingly, platforms can’t always cite policy violations and are compelled to use other methods to counter such content. They report using various content moderation techniques such as redirecting users, pre-bunking misinformation[14], promoting counterspeech and offering warnings[15], or implementing shadow bans[16]. Despite these efforts, online extremism continues to persist.

Read more: Disinformation threatens global elections – here's how to fight back[17]

What is extremism, anyway?

All these problems are further compounded by the fact we lack a commonly accepted definition[18] for terrorism or extremism. All definitions currently in place are contentious.

Academics attempt to seek clarity by using relativistic definitions[19], such as

extremism itself is context-dependent in the sense that it is an inherently relative term that describes a deviation from something that is (more) ‘ordinary’, ‘mainstream’ or ‘normal’.

However, what is something we can accept as a universal normal? Democracy is not the global norm, nor are equal rights. Not even our understanding of central tenets of human rights[20] is globally established.

What should regulators do, then?

As the eSafety commissioner attempts to shed light on how major platforms counter terrorism, we offer several recommendations for the commissioner to consider.

1. Extremists rely on more than just the major platforms to disseminate information. This highlights the importance of expanding the current inquiries beyond just the major tech players.

2. Regulators need to consider the differences between platforms that resist compliance, those that comply halfheartedly, and those that struggle to comply, such as small content storage providers. Each type of platform requires different regulatory approaches[21] or assistance.

3. Future regulations should encourage platforms to transparently collaborate with academia. The global research community is well positioned to address these challenges[22], such as by developing actionable definitions of extremism and novel countermeasures.

References

  1. ^ has sent legal notices (www.abc.net.au)
  2. ^ have 49 days to respond (www.esafety.gov.au)
  3. ^ child sexual abuse material (theconversation.com)
  4. ^ bullying (www.cbsnews.com)
  5. ^ “Christchurch Call” to action (www.beehive.govt.nz)
  6. ^ extremists still use online platforms (www.rand.org)
  7. ^ Christchurch attacks 5 years on: terrorist’s online history gives clues to preventing future atrocities (theconversation.com)
  8. ^ Global Internet Forum to Counter Terrorism (gifct.org)
  9. ^ automatically identify and remove (gifct.org)
  10. ^ “swarmcasting” multiplatform approach (static.rusi.org)
  11. ^ Jakub Krechowicz/Shutterstock (www.shutterstock.com)
  12. ^ a lawyer argued (law.stanford.edu)
  13. ^ in a different format (onlinelibrary.wiley.com)
  14. ^ pre-bunking misinformation (www.pbs.org)
  15. ^ offering warnings (www.bbc.com)
  16. ^ implementing shadow bans (theconversation.com)
  17. ^ Disinformation threatens global elections – here's how to fight back (theconversation.com)
  18. ^ commonly accepted definition (www.unodc.org)
  19. ^ relativistic definitions (www.ijcv.org)
  20. ^ central tenets of human rights (blogs.lse.ac.uk)
  21. ^ requires different regulatory approaches (ksp.techagainstterrorism.org)
  22. ^ to address these challenges (gifct.org)

Read more https://theconversation.com/terrorist-content-lurks-all-over-the-internet-regulating-only-6-major-platforms-wont-be-nearly-enough-226219

Times Magazine

A Report From France: The Mood of a Nation

France occupies a unique place in the global imagination. To many outsiders, it remains the land ...

“More Choice” Or Fewer Choices? Australia’s New Vehicle Emission Rules

The Changing Face Of Motoring When the Federal Government announced Australia’s new fuel efficien...

Female founders to benefit from new funding to turn their ideas into viable ventures

The University of Newcastle Integrated Innovation Network (I2N) has been selected by the NSW Governm...

GLOBAL SPORTS MARKETING HEAVYWEIGHTS CONVERGE IN BRISBANE FOR INAUGURAL VICTORY LAP

Australia’s premier sports marketing and creative summit, Victory Lap, has revealed its lineup of in...

The 2026 Met Gala: Fashion, Power and the Theatre of Exclusivity

Each year, on the first Monday in May, the global fashion industry converges on the steps of Metro...

Australian Wine Guide

A Quick but Informed Guide to the Varieties and Popular Brands of Australian WinesDon’t let a wine...

The Times Features

The Overlooked Link Between Flat Tennis Balls and Tenni…

Tennis elbow is the sport's most common injury. Up to 50% of recreational players will experience it...

The Australian Government will hand down the 2026/27 Federal Budget on Tuesday 12 May, and with co...

64% of Aussie kids are influencing family holiday plans…

Forget coats and heaters- think t-shirts, thongs, sunscreen and swimming. Whales aren’t the only one...

Health Insurance Recent Government Changes — And What T…

Part of the confusion surrounding private health insurance is that governments regularly adjust th...

A Report From France: The Mood of a Nation

France occupies a unique place in the global imagination. To many outsiders, it remains the land ...

The More Things Change: Change Can Hurt

The only constant in life is change. It sounds wise because it is true. Nothing stays still fore...

Seeking Financial Advice Before Investing: How Australi…

Australians are constantly reminded to “seek financial advice” before making investment decisions...

Female founders to benefit from new funding to turn the…

The University of Newcastle Integrated Innovation Network (I2N) has been selected by the NSW Governm...

MoleMap ANZ continues growth trajectory with acquisitio…

MoleMap, Australia and New Zealand’s leading skin cancer detection and surveillance service, has...