The Times Australia
The Times World News

.
The Times Real Estate

.

Australia plans to regulate 'high-risk' AI. Here's how to do that successfully

  • Written by Lisa M. Given, Professor of Information Sciences & Director, Social Change Enabling Impact Platform, RMIT University
Australia plans to regulate 'high-risk' AI. Here's how to do that successfully

This week, federal Minister for Industry and Science Ed Husic announced the Australian government’s response to the Safe and Responsible AI in Australia[1] consultation.

The response addresses feedback from last year’s consultation on artificial intelligence (AI). It received more than 500 submissions, noting “excitement for the opportunities” of AI tools, but also raising concerns about potential risks and Australians’ expectations for “regulatory safeguards to prevent harms”.

Instead of enacting a single AI regulatory law like the European Union has done[2], the Australian government plans to focus on high-risk areas of AI implementation – ones with the greatest potential for harm. This could include examples such as discrimination in the workplace, the justice system, surveillance or self-driving cars.

The government also plans to create a temporary expert advisory group to support the development of these guardrails.

Read more: Australians are concerned about AI. Is the federal government doing enough to mitigate risks?[3]

How will we define ‘high-risk’ AI?

While this proportional response may be welcomed by some, focusing on high-risk areas with only a temporary advisory body raises significant questions:

  • how will high-risk areas be defined – and who makes that decision?

  • should low-risk AI applications face similar regulation, when some interventions (such as requiring watermarks for AI-generated content) could broadly combat misinformation?

  • without a permanent advisory board, how can organisations anticipate risks for new AI technologies and new applications of AI tools in the future?

Assessing “risk” in using new technologies is not new. We have many existing principles, guidelines, and regulations that can be adapted to address concerns about AI tools.

For example, many Australian sectors are already highly regulated to address safety concerns, such as vehicles[4] and medical devices[5].

In all research involving people, Australian researchers must comply with national guidelines[6] where risk assessment practices are well defined:

  • identifying the risks and who might be at risk of harm;

  • assessing the likelihood, severity and magnitude of risk;

  • considering strategies to minimise, mitigate, and/or manage risks;

  • identifying potential benefits, and who may benefit; and

  • weighing the risks and determining whether the risks are justified by potential benefits.

This risk assessment is done before research being done, with significant review and oversight by Human Research Ethics Committees. A similar approach could be used for AI risk assessment.

AI is already in our lives

One significant problem with AI regulation is that many tools are already used in Australian homes and workplaces, but without regulatory guardrails to manage risks.

A recent YouGov report found 90% of Australian workers[7] used AI tools for daily tasks, despite serious limitations and flaws. AI tools can “hallucinate” and present fake information[8] to users. The lack of transparency[9] about training data raises concerns about bias and copyright infringement[10].

Consumers and organisations need guidance on appropriate adoption of AI tools to manage risks, but many uses are outside “high risk” areas.

Defining “high risk” settings is challenging. The concept of “risk” sits on a spectrum and is not absolute. Risk is not determined by a tool itself, or the setting where it is used. Risk arises from contextual factors that create potential for harm.

For example, while knitting needles pose little risk in everyday life, knitters are cautioned[11] against carrying metal needles on airplanes. Airport security views these as “dangerous” tools and restricts their use in this setting to prevent harm.

To identify “high risk” settings we must understand how AI tools work. Knowing AI tools can lead to gender discrimination in hiring practices[12] means all organisations must manage risk in recruitment. Not understanding the limitations of AI, like the American lawyer who trusted fake case law[13] generated by ChatGPT, highlights the risk of human error in AI tool use.

Risks posed by people and organisations in using AI tools must be managed alongside risks posed by the technology itself.

Read more: How a New York Times copyright lawsuit against OpenAI could potentially transform how AI and copyright work[14]

Who will advise the government?

The government notes in its response that the expert advisory body on AI risks will need “diverse membership and expertise from across industry, academia, civil society and the legal profession”.

Within industry, membership should include various sectors (such as healthcare, banking, law enforcement) with representation from large organisations and small-to-medium enterprises.

Within academia, membership should include not just AI computing experts, but also social scientists with expertise in consumer and organisational behaviour. They can advise on risk analysis, ethics, and what people worry about when it comes to adopting new technology, including misinformation, trust and privacy concerns.

The government must also decide how to manage potential future AI risks. A permanent advisory body could manage risks for future technologies and for new uses of existing tools.

Such a body could also advise consumers and workplaces on AI applications at lower levels of risk, particularly where limited or no regulations are in place.

Misinformation is one key area where the limitations of AI tools are known, requiring people to have strong critical thinking and information literacy skills. For example, requiring transparency in the use of AI-generated images can ensure consumers are not misled.

Yet the government’s current focus for transparency is limited to “high-risk” settings. This is a start, but more advice – and more regulation – will be needed.

References

  1. ^ Safe and Responsible AI in Australia (www.industry.gov.au)
  2. ^ European Union has done (www.eeas.europa.eu)
  3. ^ Australians are concerned about AI. Is the federal government doing enough to mitigate risks? (theconversation.com)
  4. ^ vehicles (www.infrastructure.gov.au)
  5. ^ medical devices (www.tga.gov.au)
  6. ^ national guidelines (www.nhmrc.gov.au)
  7. ^ 90% of Australian workers (itbrief.com.au)
  8. ^ hallucinate” and present fake information (theconversation.com)
  9. ^ lack of transparency (dl.acm.org)
  10. ^ copyright infringement (www.nytimes.com)
  11. ^ knitters are cautioned (www.knittingroomhk.com)
  12. ^ gender discrimination in hiring practices (www.nature.com)
  13. ^ American lawyer who trusted fake case law (www.abc.net.au)
  14. ^ How a New York Times copyright lawsuit against OpenAI could potentially transform how AI and copyright work (theconversation.com)

Read more https://theconversation.com/australia-plans-to-regulate-high-risk-ai-heres-how-to-do-that-successfully-221321

The Times Features

From Classic to Contemporary: 5 Timeless Costumes for Any Party

When it comes to dressing up for a costume party, you want to choose something that is not only fun but also memorable. Whether you're attending a Halloween event, a themed gathe...

Action Figures as Art: The Growing Trend of Custom Figures and Modding

Action figures have long been regarded as collectible items, valued by enthusiasts and fans for their connection to popular culture. However, in recent years, a growing trend has...

The Ultimate Guide to Securing Grants for Your Small Business in Australia

Running a small business in Australia comes with both opportunities and challenges. While it can be rewarding, funding your business through the early stages or periods of growth...

Men’s Guide to Styling Suits with Sweaters for the Ultimate Winter Look

Winter brings a unique set of challenges when it comes to maintaining a sharp and stylish wardrobe. While suits are a staple of a well-dressed man’s closet, staying warm during t...

Discover the Charm of a Girl's White Dress

Key Highlights Timeless elegance: White dresses feel pure and have classic charm. They are perfect for many types of events. Versatile styles: You can pick from lace, embroid...

Top Medical Grade Red Light Therapy Devices in Australia

Key Highlights Lumitter™ has advanced red light therapy devices that combine new ideas with wellness. Red light therapy helps lower pain and swelling and boosts skin health. ...

Times Magazine

The Essential Guide to Transforming Office Spaces for Maximum Efficiency

Why Office Fitouts MatterA well-designed office can make all the difference in productivity, employee satisfaction, and client impressions. Businesses of all sizes are investing in updated office spaces to create environments that foster collaborat...

The A/B Testing Revolution: How AI Optimized Landing Pages Without Human Input

A/B testing was always integral to the web-based marketing world. Was there a button that converted better? Marketing could pit one against the other and see which option worked better. This was always through human observation, and over time, as d...

Using Countdown Timers in Email: Do They Really Increase Conversions?

In a world that's always on, where marketers are attempting to entice a subscriber and get them to convert on the same screen with one email, the power of urgency is sometimes the essential element needed. One of the most popular ways to create urg...

Types of Software Consultants

In today's technology-driven world, businesses often seek the expertise of software consultants to navigate complex software needs. There are several types of software consultants, including solution architects, project managers, and user experienc...

CWU Assistive Tech Hub is Changing Lives: Win a Free Rollator Walker This Easter!

🌟 Mobility. Independence. Community. All in One. This Easter, the CWU Assistive Tech Hub is pleased to support the Banyule community by giving away a rollator walker. The giveaway will take place during the Macleod Village Easter Egg Hunt & Ma...

"Eternal Nurture" by Cara Barilla: A Timeless Collection of Wisdom and Healing

Renowned Sydney-born author and educator Cara Barilla has released her latest book, Eternal Nurture, a profound collection of inspirational quotes designed to support mindfulness, emotional healing, and personal growth. With a deep commitment to ...

LayBy Shopping