The Times Australia
Fisher and Paykel Appliances
The Times World News

.

How Australia’s new AI ‘guardrails’ can clean up the messy market for artificial intelligence

  • Written by Nicholas Davis, Industry Professor of Emerging Technology and Co-Director, Human Technology Institute, University of Technology Sydney

Australia’s federal government has today launched a proposed set of mandatory guardrails for high-risk AI[1] alongside a voluntary safety standard[2] for organisations using AI.

Each of these documents offer ten mutually reinforcing guardrails that set clear expectations for organisations across the AI supply chain. They are relevant for all organisations using AI, including internal systems aimed at boosting employee efficiency and externally-facing systems such as chatbots.

Most of the guardrails relate to things like accountability, transparency, record-keeping and making sure humans are overseeing AI systems in a meaningful way. They are aligned with emerging international standards such as the ISO standard for AI management[3] and the European Union’s AI Act[4].

The proposals for mandatory requirements for high-risk AI – which are open to public submissions[5] for the next month – recognise that AI systems are special in ways that limit the ability of existing laws to effectively prevent or mitigate a wide range of harms to Australians. While defining precisely what constitutes a high-risk setting is a core part of the consultation, the proposed principle-based approach would likely capture any systems that have a legal effect. Examples might include AI recruitment systems, systems that may limit human rights (including some facial recognition systems), and any systems that can cause physical harm, such as autonomous vehicles.

Well-designed guardrails will improve technology and make us all better off. On this front, the government should accelerate law reform efforts to clarify existing rules and improve both transparency and accountability in the market. At the same time, we don’t need to – nor should we – wait for the government to act.

The AI market is a mess

As it stands, the market for AI products and services is a mess. The central problem is that people don’t know how AI systems work, when they’re using them, and whether the output helps or hurts them.

Take, for example, a company that recently asked my advice on a generative AI service projected to cost hundreds of thousands of dollars each year. It was worried about falling behind competitors and having difficulty choosing between vendors.

Yet, in the first 15 minutes of discussion, the company revealed it had no reliable information around the potential benefit for the business, and no knowledge of existing generative AI use by its teams.

It’s important we get this right. If you believe even a fraction of the hype, AI represents a huge opportunity for Australia. Estimates referenced by the federal government[6] suggest the economic boost from AI and automation could be up to A$600 billion every year by 2030. This would lift our GDP to 25% above 2023 levels.

But all of this is at risk. The evidence is in the alarmingly high failure rates of AI projects (above 80% by some estimates[7]), an array of reckless rollouts, low levels of citizen trust[8] and the prospect of thousands of Robodebt-esque crises across both industry and government.

The information asymmetry problem

A lack of skills and experience among decision-makers is undoubtedly part of the problem. But the rapid pace of innovation in AI is supercharging another challenge: information asymmetry.

Information asymmetry is a simple, Nobel prize-winning economic concept[9] with serious implications for everyone. And it’s a particularly pernicious challenge when it comes to AI.

When buyers and sellers have uneven knowledge about a product or service, it doesn’t just mean one party gains at the other’s expense. It can lead to poor-quality goods dominating the market, and even the market failing entirely.

AI creates information asymmetries in spades. AI models are technical and complex, they are often embedded and hidden inside other systems, and they are increasingly being used to make important choices.

Balancing out these asymmetries should deeply concern all of us. Boards, executives and shareholders want AI investments to pay off. Consumers want systems that work in their interests. And we all want to enjoy the benefits of economic expansion while avoiding the very real harms AI systems can inflict if they fail, or if they are used maliciously or deployed inappropriately.

In the short term, at least, companies selling AI gain a real benefit from restricting information so they can do deals with naïve counterparties. Solving this problem will require more than upskilling. It means using a range of tools and incentives to gather and share accurate, timely and important information about AI systems.

What businesses can do today

Now is the time to act. Businesses across Australia can pick up the Voluntary AI Safety Standard[10] (or the International Standard Organisation’s version[11]) and start gathering and documenting the information they need to make better decisions about AI today.

This will help in two ways. First, it will help businesses to take a structured approach to understanding and governing their own use of AI systems, to ask useful questions to (and demand answers from) their technology partners, and to signal to the market that their AI use is trustworthy.

Second, as more and more businesses adopt the standard, Australian and international vendors and deployers will feel market pressure to ensure their products and services are fit for purpose. In turn, it will become cheaper and easier for all of us to know whether the AI system we’re buying, relying on or being judged by actually serves our needs.

Clearing a path

Australian consumers and businesses both want AI to be safe and responsible. But we urgently need to close the huge gap that exists between aspiration and practice.

The National AI Centre’s Responsible AI index[12] shows that while 78% of organisations believed they were developing and deploying AI systems responsibly, only 29% of organisations were applying actual practices towards this end.

Safe and responsible AI is where good governance meets good business practice and human-centred technology. In the bigger picture, it’s also about ensuring that innovation thrives in a well-functioning market. On both these fronts, standards can help us clear a path through the clutter.

References

  1. ^ mandatory guardrails for high-risk AI (consult.industry.gov.au)
  2. ^ voluntary safety standard (www.industry.gov.au)
  3. ^ ISO standard for AI management (www.iso.org)
  4. ^ European Union’s AI Act (artificialintelligenceact.eu)
  5. ^ public submissions (consult.industry.gov.au)
  6. ^ Estimates referenced by the federal government (storage.googleapis.com)
  7. ^ above 80% by some estimates (www.rand.org)
  8. ^ low levels of citizen trust (kpmg.com)
  9. ^ Nobel prize-winning economic concept (www.nobelprize.org)
  10. ^ Voluntary AI Safety Standard (www.industry.gov.au)
  11. ^ International Standard Organisation’s version (www.iso.org)
  12. ^ Responsible AI index (url.au.m.mimecastprotect.com)

Read more https://theconversation.com/how-australias-new-ai-guardrails-can-clean-up-the-messy-market-for-artificial-intelligence-238307

Times Magazine

Can bigger-is-better ‘scaling laws’ keep AI improving forever? History says we can’t be too sure

OpenAI chief executive Sam Altman – perhaps the most prominent face of the artificial intellig...

A backlash against AI imagery in ads may have begun as brands promote ‘human-made’

In a wave of new ads, brands like Heineken, Polaroid and Cadbury have started hating on artifici...

Home batteries now four times the size as new installers enter the market

Australians are investing in larger home battery set ups than ever before with data showing the ...

Q&A with Freya Alexander – the young artist transforming co-working spaces into creative galleries

As the current Artist in Residence at Hub Australia, Freya Alexander is bringing colour and creativi...

This Christmas, Give the Navman Gift That Never Stops Giving – Safety

Protect your loved one’s drives with a Navman Dash Cam.  This Christmas don’t just give – prote...

Yoto now available in Kmart and The Memo, bringing screen-free storytelling to Australian families

Yoto, the kids’ audio platform inspiring creativity and imagination around the world, has launched i...

The Times Features

Why the Mortgage Industry Needs More Women (And What We're Actually Doing About It)

I've been in fintech and the mortgage industry for about a year and a half now. My background is i...

Inflation jumps in October, adding to pressure on government to make budget savings

Annual inflation rose[1] to a 16-month high of 3.8% in October, adding to pressure on the govern...

Transforming Addiction Treatment Marketing Across Australasia & Southeast Asia

In a competitive and highly regulated space like addiction treatment, standing out online is no sm...

Aiper Scuba X1 Robotic Pool Cleaner Review: Powerful Cleaning, Smart Design

If you’re anything like me, the dream is a pool that always looks swimmable without you having to ha...

YepAI Emerges as AI Dark Horse, Launches V3 SuperAgent to Revolutionize E-commerce

November 24, 2025 – YepAI today announced the launch of its V3 SuperAgent, an enhanced AI platf...

What SMEs Should Look For When Choosing a Shared Office in 2026

Small and medium-sized enterprises remain the backbone of Australia’s economy. As of mid-2024, sma...

Anthony Albanese Probably Won’t Lead Labor Into the Next Federal Election — So Who Will?

As Australia edges closer to the next federal election, a quiet but unmistakable shift is rippli...

Top doctors tip into AI medtech capital raise a second time as Aussie start up expands globally

Medow Health AI, an Australian start up developing AI native tools for specialist doctors to  auto...

Record-breaking prize home draw offers Aussies a shot at luxury living

With home ownership slipping out of reach for many Australians, a growing number are snapping up...