The Times Australia
Google AI
The Times World News

.

AI is not a magic wand – it has built-in problems that are difficult to fix and can be dangerous

  • Written by Niusha Shafiabady, Associate Professor in Computational Intelligence, Charles Darwin University
AI is not a magic wand – it has built-in problems that are difficult to fix and can be dangerous

By now, all of us have heard and read a lot about artificial intelligence (AI). You’ve likely used some of the countless AI tools[1] that are becoming available. For some, AI feels like a magic wand that predicts the future.

But AI is not perfect. A supermarket meal planner in Aotearoa New Zealand gave customers poisonous recipes[2], a New York City chatbot advised people to break the law[3], and Google’s AI Overview is telling people to eat rocks[4].

At its core, an AI tool is a particular system that addresses a particular problem[5]. With any AI system, we should match our expectations to its abilities – and many of those come down to how the AI was built.

Let’s explore some inherent shortcomings of AI systems.

Trouble in the real world

One of the inherent issues for all AI systems is that they are not 100% accurate in real-world settings. For example, a predictive AI system will be trained using data points from the past.

If the AI then comes across something new – not similar to anything in the training data – it most likely won’t be able to make the correct decision.

As a hypothetical example, let’s take a military plane equipped with an AI-powered autopilot system. This system will function thanks to its training “knowledge base”. But an AI really isn’t a magic wand, it’s just mathematical computations. An adversary could create obstacles the plane AI cannot “see” because they are not in the training data, leading to potentially catastrophic consequences.

Unfortunately, there is not much we can do about this problem apart from trying to train the AI for all possible circumstances that we know of. This can sometimes be an insurmountable task.

Bias in the training data

You may have heard about AI making biased decisions[6]. Usually, bias happens when we have unbalanced data. In simple terms, this means that when training the AI system, we are showing it too many examples of one type of outcome and very few of another type.

Let’s take the example of an AI system trained to predict the likelihood a given individual will commit a crime[7]. If the crime data used for training the system mostly contains people from group A (say, a particular ethnicity) and very few from group B, the system won’t learn about both groups equally.

As a result, its predictions for group A will make it seem these people are more likely to commit crimes compared to people from group B. If the system is used uncritically, the presence of this bias can have severe ethical consequences[8].

Thankfully, developers can address this issue by “balancing” the data set. This can involve different approaches, including the use of synthetic[9] data – computer-generated, pre-labelled data built for testing and training AI[10] that has checks built into it to protect against bias.

A group of people are passing each other in the street with a surveillance overlay of white box outlines.
Having balanced data is critical to prevent AI systems from perpetuating bias. Comuzi/© BBC/Better Images of AI/Surveillance View A., CC BY[11][12]

Being out of date

Another issue with AI can arise when it’s been trained “offline”[13] and isn’t up to date with the dynamics of the problem it is meant to work on.

A simple example would be an AI system developed to predict daily temperature in a city. Its training data contain all the past information on temperature data for this location.

After the AI has finished training and is deployed, let’s say a severe climactic event disrupts the usual weather dynamics. Since the AI system making the predictions was trained on data that didn’t include this disruption, its predictions will become increasingly inaccurate.

The way to solve this issue is training the AI “online”[14], meaning it is regularly shown the latest temperature data while being used to predict temperatures.

This sounds like a great solution, but there are a few risks associated with online training. We can leave the AI system to train itself using the latest data, but it may get out of control.

Fundamentally, this can happen because of chaos theory[15], which, in simple terms, means most AI systems are sensitive to initial conditions. When we don’t know what data the system will come across, we can’t know how to tune the initial conditions to control potential instabilities in the future.

When the data isn’t right

Sometimes, the training data just isn’t fit for purpose. For example, it may not have the qualities the AI system needs to perform whatever task we are training it to do.

To use an extremely simplified example, imagine an AI tool for identifying “tall” and “short” people. In the training data, should a person who is 170cm be labelled tall or short? If tall, what will the system return when it comes across someone who is 169.5cm? (Perhaps the best solution would be to add a “medium” label.)

The above may seem trivial, but issues with data labelling or poor data sets can have problematic consequences if the AI system is involved in medical diagnosis[16], for example.

Fixing this problem is not easy, since identifying the relevant pieces of information requires a great deal of knowledge and experience. Bringing on board a subject matter expert in the data collection process can be a great solution, as it can guide the developers on what types of data to even include to begin with.

As (future) users of AI and technology, it is important for all of us to be aware of these issues to broaden our perspective on AI and its prediction outcomes concerning different aspects of our lives.

References

  1. ^ some of the countless AI tools (www.futuretools.io)
  2. ^ poisonous recipes (www.forbes.com)
  3. ^ advised people to break the law (tech.co)
  4. ^ telling people to eat rocks (theconversation.com)
  5. ^ particular system that addresses a particular problem (theconversation.com)
  6. ^ AI making biased decisions (www.mdpi.com)
  7. ^ will commit a crime (theconversation.com)
  8. ^ severe ethical consequences (theconversation.com)
  9. ^ synthetic (research.ibm.com)
  10. ^ for testing and training AI (journals.plos.org)
  11. ^ Comuzi/© BBC/Better Images of AI/Surveillance View A. (betterimagesofai.org)
  12. ^ CC BY (creativecommons.org)
  13. ^ it’s been trained “offline” (www.sciencedirect.com)
  14. ^ training the AI “online” (www.ibm.com)
  15. ^ chaos theory (fractalfoundation.org)
  16. ^ AI system is involved in medical diagnosis (theconversation.com)

Read more https://theconversation.com/ai-is-not-a-magic-wand-it-has-built-in-problems-that-are-difficult-to-fix-and-can-be-dangerous-230878

Times Magazine

Governance Models for Headless CMS in Large Organizations

Where headless CMS is adopted by large enterprises, governance is the single most crucial factor d...

Narwal Freo Z Ultra Robotic Vacuum and Mop Cleaner

Rating: ★★★★☆ (4.4/5)Category: Premium Robot Vacuum & Mop ComboBest for: Busy households, ha...

Shark launches SteamSpot - the shortcut for everyday floor mess

Shark introduces the Shark SteamSpot Steam Mop, a lightweight steam mop designed to make everyda...

Game Together, Stay Together: Logitech G Reveals Gaming Couples Enjoy Higher Relationship Satisfaction

With Valentine’s Day right around the corner, many lovebirds across Australia are planning for the m...

AI threatens to eat business software – and it could change the way we work

In recent weeks, a range of large “software-as-a-service” companies, including Salesforce[1], Se...

Worried AI means you won’t get a job when you graduate? Here’s what the research says

The head of the International Monetary Fund, Kristalina Georgieva, has warned[1] young people ...

The Times Features

How Businesses Are Generating Profits in a High-Inflation Economic Environment

Inflation in Australia and globally has surged to multi-decade highs since 2021, driven by pande...

The Effects of the War in the Middle East on Australian Small Businesses

The war in the Middle East is not a distant geopolitical event for Australia. In an interconnect...

Back at uni? How to help your wellbeing while you study

University can be a time of great opportunities, but it can also be very stressful[1]. Many stud...

Taste Port Douglas celebrates 10 years of world-class flavour in the tropics

30+ events, new sunrise and wellness experiences, 20+ chefs and a headline Michelin-star line-up...

Oztent RV tent range. Buy with caution

A review of the Oztent RV "30 second tent" range. Three years ago we bought an RV-4 from BCF Mack...

Essential Upgrades for a Smarter, Safer Australian Home

As we settle into 2026, the concept of the "dream home" has fundamentally shifted. The focus has m...

How To Modernise Your Home Without Overcapitalising

For many Australian homeowners, the dream of a "Grand Designs" transformation is often checked by ...

The Art of the Big Trip: Planning a Seamless Multi-Generational Getaway in Tropical North Queensland

There is a unique magic to the multi-generational holiday. It is a rare opportunity where gr...

Love Without Borders: ‘Second Marriage At First Sight’ Opens Casting Call for Melbourne Singles Willing to Relocate for Romance

Fans of Married At First Sight UK and Married At First Sight Australia are about to see the expe...