The Times Australia
The Times World News

.

Algorithms are pushing AI-generated falsehoods at an alarming rate. How do we stop this?

  • Written by Stan Karanasios, Associate Professor, The University of Queensland
Algorithms are pushing AI-generated falsehoods at an alarming rate. How do we stop this?

Generative artificial intelligence (AI) tools are supercharging the problem of misinformation, disinformation and fake news. OpenAI’s ChatGPT, Google’s Gemini, and various image, voice and video generators have made it easier than ever to produce content, while making it harder to tell what is factual or real.

Malicious actors looking to spread disinformation can use AI tools to largely automate the generation of convincing and misleading text[1].

This raises pressing questions: how much of the content we consume online is true and how can we determine its authenticity? And can anyone stop this?

It’s not an idle concern. Organisations seeking to covertly influence public opinion or sway elections can now scale their operations[2] with AI to unprecedented levels. And their content is being widely disseminated by search engines and social media.

Read more: What is Sora? A new generative AI tool could transform video production and amplify disinformation risks[3]

Fakes everywhere

Earlier this year, a German study[4] on search engine content quality noted “a trend toward simplified, repetitive and potentially AI-generated content” on Google, Bing and DuckDuckGo.

Traditionally, readers of news media could rely on editorial control to uphold journalistic standards and verify facts. But AI is rapidly changing this space.

In a report published this week, the internet trust organisation NewsGuard identified 725 unreliable websites[5] that publish AI-generated news and information “with little to no human oversight”.

Last month, Google released an experimental AI tool[6] for a select group of independent publishers in the United States. Using generative AI, the publisher can summarise articles pulled from a list of external websites that produce news and content relevant to their audience. As a condition of the trial, the users have to publish three such articles per day.

Platforms hosting content and developing generative AI blur the traditional lines that enable trust in online content.

Can the government step in?

Australia has already seen tussles between government and online platforms over the display and moderation of news and content.

In 2019, the Australian government amended the criminal code[7] to mandate the swift removal of “abhorrent violent material” by social media platforms.

The Australian Competition and Consumer Commission’s (ACCC) inquiry into power imbalances between Australian news media and digital platforms led to the 2021 implementation of a bargaining code[8] that forced platforms to pay media for their news content.

While these might be considered partial successes, they also demonstrate the scale of the problem and the difficulty of taking action.

Our research[9] indicates these conflicts saw online platforms initially open to changes and later resisting them, while the Australian government oscillated from enforcing mandatory measures to preferring voluntary actions.

Ultimately, the government realised that relying on platforms’ “trust us” promises wouldn’t lead to the desired outcomes.

Read more: Why Google and Meta owe news publishers much more than you think – and billions more than they’d like to admit[10]

The takeaway from our study is that once digital products become integral to millions of businesses and everyday lives, they serve as a tool for platforms, AI companies and big tech to anticipate and push back against government.

With this in mind, it is right to be sceptical of early calls for regulation of generative AI by tech leaders like Elon Musk[11] and Sam Altman. Such calls have faded as AI takes a hold on our lives and online content.

A challenge lies in the sheer speed of change, which is so swift that safeguards to mitigate the potential risks to society are not yet established. Accordingly, the World Economic Forum’s 2024 Global Risk Report has predicted mis- and disinformation as the greatest threats[12] in the next two years.

The problem gets worse through generative AI’s ability to create multimedia content. Based on current trends, we can expect an increase in deepfake incidents[13], although social media platforms like Facebook are responding to these issues. They aim to automatically identify and tag[14] AI-generated photos, video and audio.

Read more: The OpenAI saga demonstrates how big corporations dominate the shaping of our technological future[15]

What can we do?

Australia’s eSafety commissioner is working on ways to regulate and mitigate[16] the potential harm caused by generative AI while balancing its potential opportunities.

A key idea is “safety by design”, which requires tech firms to place these safety considerations at the core of their products.

Other countries like the US are further ahead with the regulation of AI. For example, US President Joe Biden’s recent executive order on the safe deployment of AI[17] requires companies to share safety test results with the government, regulates red-team testing[18] (simulated hacking attacks), and guides watermarking on content.

We call for three steps to help protect against the risks of generative AI in combination with disinformation.

1. Regulation needs to pose clear rules[19] without allowing for nebulous “best effort” aims or “trust us” approaches.

2. To protect against large-scale disinformation operations, we need to teach media literacy in the same way we teach maths.

3. Safety tech or “safety by design” needs to become a non-negotiable part of every product development strategy.

People are aware AI-generated content is on the rise. In theory, they should adjust their information habits accordingly. However, research shows users generally tend to underestimate[20] their own risk of believing fake news compared to the perceived risk for others.

Finding trustworthy content shouldn’t involve sifting through AI-generated content to make sense of what is factual.

References

  1. ^ convincing and misleading text (cyber.fsi.stanford.edu)
  2. ^ scale their operations (cyber.fsi.stanford.edu)
  3. ^ What is Sora? A new generative AI tool could transform video production and amplify disinformation risks (theconversation.com)
  4. ^ a German study (www.techradar.com)
  5. ^ identified 725 unreliable websites (www.newsguardtech.com)
  6. ^ released an experimental AI tool (www.adweek.com)
  7. ^ amended the criminal code (www.aph.gov.au)
  8. ^ a bargaining code (www.accc.gov.au)
  9. ^ Our research (journals.sagepub.com)
  10. ^ Why Google and Meta owe news publishers much more than you think – and billions more than they’d like to admit (theconversation.com)
  11. ^ Elon Musk (fortune.com)
  12. ^ greatest threats (www.weforum.org)
  13. ^ deepfake incidents (www.nbcnews.com)
  14. ^ automatically identify and tag (about.fb.com)
  15. ^ The OpenAI saga demonstrates how big corporations dominate the shaping of our technological future (theconversation.com)
  16. ^ is working on ways to regulate and mitigate (www.esafety.gov.au)
  17. ^ on the safe deployment of AI (www.theguardian.com)
  18. ^ red-team testing (en.wikipedia.org)
  19. ^ to pose clear rules (www.linkedin.com)
  20. ^ generally tend to underestimate (www.ncbi.nlm.nih.gov)

Read more https://theconversation.com/algorithms-are-pushing-ai-generated-falsehoods-at-an-alarming-rate-how-do-we-stop-this-224626

Times Magazine

Headless CMS in Digital Twins and 3D Product Experiences

Image by freepik As the metaverse becomes more advanced and accessible, it's clear that multiple sectors will use digital twins and 3D product experiences to visualize, connect, and streamline efforts better. A digital twin is a virtual replica of ...

The Decline of Hyper-Casual: How Mid-Core Mobile Games Took Over in 2025

In recent years, the mobile gaming landscape has undergone a significant transformation, with mid-core mobile games emerging as the dominant force in app stores by 2025. This shift is underpinned by changing user habits and evolving monetization tr...

Understanding ITIL 4 and PRINCE2 Project Management Synergy

Key Highlights ITIL 4 focuses on IT service management, emphasising continual improvement and value creation through modern digital transformation approaches. PRINCE2 project management supports systematic planning and execution of projects wit...

What AI Adoption Means for the Future of Workplace Risk Management

Image by freepik As industrial operations become more complex and fast-paced, the risks faced by workers and employers alike continue to grow. Traditional safety models—reliant on manual oversight, reactive investigations, and standardised checklist...

From Beach Bops to Alpine Anthems: Your Sonos Survival Guide for a Long Weekend Escape

Alright, fellow adventurers and relaxation enthusiasts! So, you've packed your bags, charged your devices, and mentally prepared for that glorious King's Birthday long weekend. But hold on, are you really ready? Because a true long weekend warrior kn...

Effective Commercial Pest Control Solutions for a Safer Workplace

Keeping a workplace clean, safe, and free from pests is essential for maintaining productivity, protecting employee health, and upholding a company's reputation. Pests pose health risks, can cause structural damage, and can lead to serious legal an...

The Times Features

Duke of Dural to Get Rooftop Bar as New Owners Invest in Venue Upgrade

The Duke of Dural, in Sydney’s north-west, is set for a major uplift under new ownership, following its acquisition by hospitality group Good Beer Company this week. Led by resp...

Prefab’s Second Life: Why Australia’s Backyard Boom Needs a Circular Makeover

The humble granny flat is being reimagined not just as a fix for housing shortages, but as a cornerstone of circular, factory-built architecture. But are our systems ready to s...

Melbourne’s Burglary Boom: Break-Ins Surge Nearly 25%

Victorian homeowners are being warned to act now, as rising break-ins and falling arrest rates paint a worrying picture for suburban safety. Melbourne residents are facing an ...

Exploring the Curriculum at a Modern Junior School in Melbourne

Key Highlights The curriculum at junior schools emphasises whole-person development, catering to children’s physical, emotional, and intellectual needs. It ensures early year...

Distressed by all the bad news? Here’s how to stay informed but still look after yourself

If you’re feeling like the news is particularly bad at the moment, you’re not alone. But many of us can’t look away – and don’t want to. Engaging with news can help us make ...

The Role of Your GP in Creating a Chronic Disease Management Plan That Works

Living with a long-term condition, whether that is diabetes, asthma, arthritis or heart disease, means making hundreds of small decisions every day. You plan your diet against m...