The Times Australia
Fisher and Paykel Appliances
The Times World News

.

‘Godfather of AI’ now fears it’s unsafe. He has a plan to rein it in

  • Written by Armin Chitizadeh, Lecturer, School of Computer Science, University of Sydney

This week the US Federal Bureau of Investigation revealed two men suspected[1] of bombing a fertility clinic in California last month allegedly used artificial intelligence (AI) to obtain bomb-making instructions. The FBI did not disclose the name of the AI program in question.

This brings into sharp focus the urgent need to make AI safer. Currently we are living in the “wild west” era of AI, where companies are fiercely competing to develop the fastest and most entertaining AI systems. Each company wants to outdo competitors and claim the top spot. This intense competition often leads to intentional or unintentional shortcuts[2] – especially when it comes to safety.

Coincidentally, at around the same time of the FBI’s revelation, one of the godfathers of modern AI, Canadian computer science professor Yoshua Bengio, launched a new nonprofit organisation[3] dedicated to developing a new AI model specifically designed to be safer than other AI models – and target those that cause social harm.

So what is Bengio’s new AI model? And will it actually protect the world from AI-faciliated harm?

An ‘honest’ AI

In 2018, Bengio, alongside his colleagues Yann LeCun and Geoffrey Hinton, won the Turing Award for groundbreaking research they had published three years earlier on deep learning[4]. A branch of machine learning, deep learning attempts to mimic the processes of the human brain by using artificial neural networks to learn from computational data and make predictions.

Bengio’s new nonprofit organisation, LawZero[5], is developing “Scientist AI”. Bengio has said[6] this model will be “honest and not deceptive”, and incorporate safety-by-design principles.

According to a preprint paper[7] released online earlier this year, Scientist AI will differ from current AI systems in two key ways.

First, it can assess and communicate its confidence level in its answers, helping to reduce the problem of AI giving overly confident and incorrect responses.

Second, it can explain its reasoning to humans, allowing its conclusions to be evaluated and tested for accuracy.

Interestingly, older AI systems had this feature[8]. But in the rush for speed and new approaches, many modern AI models[9] can’t explain their decisions. Their developers have sacrificed explainability for speed.

Bengio also intends “Scientist AI” to act as a guardrail against unsafe AI. It could monitor other, less reliable and harmful AI systems — essentially fighting fire with fire.

This may be the only viable solution to improve AI safety. Humans cannot properly monitor systems such as ChatGPT, which handle over a billion queries daily. Only another AI can manage this scale.

Using an AI system against other AI systems is not just a sci-fi concept – it’s a common practice in research to compare and test different level of intelligence in AI systems[10].

Adding a ‘world model’

Large language models and machine learning are just small parts of today’s AI landscape.

Another key addition Bengio’s team are adding to Scientist AI is the “world model[11]” which brings certainty and explainability. Just as humans make decisions based on their understanding of the world, AI needs a similar model to function effectively.

The absence of a world model in current AI models is clear.

One well-known example is the “hand problem[12]”: most of today’s AI models can imitate the appearance of hands but cannot replicate natural hand movements, because they lack an understanding of the physics — a world model — behind them.

Another example is how models such as ChatGPT struggle with chess, failing to win and even making illegal moves[13].

This is despite simpler AI systems, which do contain a model of the “world” of chess, beating even the best human players[14].

These issues stem from the lack of a foundational world model in these systems, which are not inherently designed to model the dynamics of the real world[15].

A man with grey and white hair wearing a suit speaking into a microphone.
Yoshua Bengio is recognised as one of the godfathers of AI. Alex Wong/Getty Images

On the right track – but it will be bumpy

Bengio is on the right track, aiming to build safer, more trustworthy AI by combining large language models with other AI technologies.

However, his journey isn’t going to be easy. LawZero’s US$30 million in funding[16] is small compared to efforts such as the US$500 billion project[17] announced by US President Donald Trump earlier this year to accelerate the development of AI.

Making LawZero’s task harder is the fact that Scientist AI – like any other AI project – needs huge amounts of data to be powerful, and most data are controlled by major tech companies[18].

There’s also an outstanding question. Even if Bengio can build an AI system that does everything he says it can, how is it going to be able to control other systems that might be causing harm?

Still, this project, with talented researchers behind it, could spark a movement toward a future where AI truly helps humans thrive. If successful, it could set new expectations for safe AI, motivating researchers, developers, and policymakers to prioritise safety.

Perhaps if we had taken similar action when social media first emerged, we would have a safer online environment for young people’s mental health. And maybe, if Scientist AI had already been in place, it could have prevented people with harmful intentions from accessing dangerous information with the help of AI systems.

References

  1. ^ revealed two men suspected (www.cnbc.com)
  2. ^ often leads to intentional or unintentional shortcuts (theconversation.com)
  3. ^ launched a new nonprofit organisation (www.theguardian.com)
  4. ^ had published three years earlier on deep learning (scholar.google.com)
  5. ^ LawZero (lawzero.org)
  6. ^ Bengio has said (www.theguardian.com)
  7. ^ preprint paper (arxiv.org)
  8. ^ older AI systems had this feature (journals.sagepub.com)
  9. ^ modern AI models (seon.io)
  10. ^ compare and test different level of intelligence in AI systems (link.springer.com)
  11. ^ world model (medium.com)
  12. ^ hand problem (www.britannica.com)
  13. ^ ChatGPT struggle with chess, failing to win and even making illegal moves (www.chess.com)
  14. ^ beating even the best human players (www.sciencefocus.com)
  15. ^ are not inherently designed to model the dynamics of the real world (arxiv.org)
  16. ^ US$30 million in funding (time.com)
  17. ^ US$500 billion project (theconversation.com)
  18. ^ most data are controlled by major tech companies (www.theguardian.com)

Read more https://theconversation.com/godfather-of-ai-now-fears-its-unsafe-he-has-a-plan-to-rein-it-in-258288

Active Wear

Times Magazine

World Kindness Day: Commentary from Kath Koschel, founder of Kindness Factory.

What does World Kindness Day mean to you as an individual, and to the Kindness Factory as an organ...

In 2024, the climate crisis worsened in all ways. But we can still limit warming with bold action

Climate change has been on the world’s radar for decades[1]. Predictions made by scientists at...

End-of-Life Planning: Why Talking About Death With Family Makes Funeral Planning Easier

I spend a lot of time talking about death. Not in a morbid, gloomy way—but in the same way we d...

YepAI Joins Victoria's AI Trade Mission to Singapore for Big Data & AI World Asia 2025

YepAI, a Melbourne-based leader in enterprise artificial intelligence solutions, announced today...

Building a Strong Online Presence with Katoomba Web Design

Katoomba web design is more than just creating a website that looks good—it’s about building an onli...

September Sunset Polo

International Polo Tour To Bridge Historic Sport, Life-Changing Philanthropy, and Breath-Taking Beau...

The Times Features

Australian Startup Business Operators Should Make Connections with Asian Enterprises — That Is Where Their Future Lies

In the rapidly shifting global economy, Australian startups are increasingly finding that their ...

How early is too early’ for Hot Cross Buns to hit supermarket and bakery shelves

Every year, Australians find themselves in the middle of the nation’s most delicious dilemmas - ...

Ovarian cancer community rallied Parliament

The fight against ovarian cancer took centre stage at Parliament House in Canberra last week as th...

After 2 years of devastating war, will Arab countries now turn their backs on Israel?

The Middle East has long been riddled by instability. This makes getting a sense of the broader...

RBA keeps interest rates on hold, leaving borrowers looking further ahead for relief

As expected, the Reserve Bank of Australia (RBA) has kept the cash rate steady at 3.6%[1]. Its b...

Crystalbrook Collection Introduces ‘No Rings Attached’: Australia’s First Un-Honeymoon for Couples

Why should newlyweds have all the fun? As Australia’s crude marriage rate falls to a 20-year low, ...

Echoes of the Past: Sue Carter Brings Ancient Worlds to Life at Birli Gallery

Launching November 15 at 6pm at Birli Gallery, Midland, Echoes of the Past marks the highly anti...

Why careless adoption of AI backfires so easily

Artificial intelligence (AI) is rapidly becoming commonplace, despite statistics showing[1] th...

How airline fares are set and should we expect lower fares any time soon?

Airline ticket prices may seem mysterious (why is the same flight one price one day, quite anoth...