Google AI
The Times Australia
The Times World News

.

The best way to regulate AI might be not to specifically regulate AI. This is why

  • Written by Stephen King, Professor of Economics, Monash University



The new wave of artificial intelligence – so-called AI – is bringing with it promises as well as threats.

By assisting workers, it can raise productivity and boost real wages. By making use of large, underutilised data, it can improve outcomes in services including retailing, health and education.

The risks include deepfakes, privacy abuse, unappealable algorithmic decisions, intellectual property infringement and wholesale job losses.

Both the risks and the potential benefits seem to grow by the day. On Thursday Open AI released new models it said could reason[1], performing complex calculations and drawing conclusions.

But, as a specialist in competition and consumer protection, I have formed the view that calls for new AI-specific regulations[2] are largely misguided.

Most uses of AI are already regulated

A Senate committee[3] is about to report on the opportunities and impacts of the uptake of AI, and I helped draft the Productivity Commission’s submission[4].

Separately, the government is consulting about mandatory guardrails[5] for AI in high-risk settings, which would function as a sort of checklist for what developers should consider alongside a voluntary safety standard[6].

Here’s my thinking: most of the potential uses of AI are already covered by existing rules and regulations designed to do things such as protect consumers, protect privacy and outlaw discrimination.

These laws are far from perfect, but where they are not perfect the best approach is to fix or extend them rather than introduce special extra rules for AI.

AI can certainly raise challenges for the laws we have – for example, by making it easier to mislead consumers or to apply algorithms that help businesses to collude on prices.

But the key point is that laws to control these things exist, as do the regulators experienced in enforcing them.

The best approach is to make existing rules work

One of Australia’s great advantages is the strength and expertise of its regulators, among them the Competition and Consumer Commission, the Communications and Media Authority, the Australian Information Commissioner, the Australian Securities and Investments Commission, and the Australian Energy Regulator.

Their job ought to be to show where AI is covered by the existing rules, to evaluate the ways in which AI might fall foul of those rules, and to run test cases that make the applicability of the rules clear.

It is an approach that will help build trust in AI, as consumers see they are already protected, as well as providing clarity for businesses.

AI might be new, but the established consensus about what is and is not acceptable behaviour hasn’t much changed.

Some rules will need to be tweaked

In some situations, existing regulations will need to be amended or extended to ensure behaviours facilitated by AI are covered. Approval processes for vehicles, machinery and medical equipment are among those that will increasingly need to take account of AI.

Humans make mistakes too. DenisStarostin/Shutterstock[7]

And in some cases, new regulations will be needed. But this should be where we end up, not where we begin. Trying to regulate AI because it is AI will, at best, be ineffective. At worst, it will stifle the development of socially desirable uses of AI.

Many uses of AI will create little if any risk. Where potential harm exists, it will need to be weighed against the potential benefits of the use. The risks and benefits ought to be judged against real-world, human-based alternatives, which are themselves far from risk-free.

New regulations will only be needed where existing regulations – even when clarified, amended or extended – are inadequate.

Where they are needed, they ought to be technology-neutral wherever possible. Rules written for specific technologies are likely to quickly become obsolete.

Last mover advantage

Finally, there’s a lot to be said for becoming an international “regulation taker”. Other jurisdictions such as the European Union are leading the way in designing AI-specific regulations.

Product developers worldwide, including those in Australia, will need to meet those new rules if they want to access the EU and those other big markets.

If Australia developed its own idiosyncratic AI-specific rules, developers might ignore our relatively small market and go elsewhere.

This means that, in those limited situations where AI-specific regulation is needed, the starting point ought to be the overseas rules that already exist.

There’s an advantage in being a late or last mover. This doesn’t mean Australia shouldn’t be in the forefront of developing international standards. It merely means it should help design those standards with other countries in international forums rather than striking out on its own.

The landscape is still developing. Our aim ought to be to give ourselves the best chance of maximising the gains from AI while providing safety nets to protect ourselves from adverse consequences. Our existing rules, rather than new AI-specific ones, ought to be where we start.

References

  1. ^ reason (azure.microsoft.com)
  2. ^ AI-specific regulations (humanrights.gov.au)
  3. ^ Senate committee (www.aph.gov.au)
  4. ^ Productivity Commission’s submission (www.pc.gov.au)
  5. ^ mandatory guardrails (www.minister.industry.gov.au)
  6. ^ safety standard (www.industry.gov.au)
  7. ^ DenisStarostin/Shutterstock (www.shutterstock.com)

Read more https://theconversation.com/the-best-way-to-regulate-ai-might-be-not-to-specifically-regulate-ai-this-is-why-238788

Times Magazine

The Voltx Topband V1200 Portable Power Station Review

When we received a Voltx Topband V1200 portable power station for review, a staff member at The Time...

Is E10 fuel bad for my car? And could it save me money?

Fuel has become a precious, and increasingly expensive, commodity. The ongoing Middle East co...

Efficient Water Carts for Dust Control

Managing dust effectively is a critical challenge across numerous industries in Australia. From sp...

How new rules could stop AI scrapers destroying the internet

Australians are among the most anxious in the world[1] about artificial intelligence (AI). This...

Why Car Enthusiasts Are Turning to Container Shipping for Interstate Moves

Moving across the country requires careful planning and plenty of patience. The scale of domestic ...

What to know if you’re considering an EV

Soaring petrol prices are once again making many Australians think seriously[1] about switching ...

The Times Features

Hearing Australia first in the world to provide innovat…

Australians with hearing loss will benefit from a new generation hearing aid fitting prescription...

Running Run Army this month? Here's how to prep for rac…

With Run Army Brisbane this Sunday and Townsville to follow on 19 April, GO2 Health’s Kate Boucher...

As the Iran war disrupts supplies, will it affect acces…

As the conflict in the Middle East disrupts fuel, shipping and food supplies, many are starting ...

Finding the Right Disability Housing in Perth: A Practi…

Where you live shapes everything. It shapes the relationships you build, the community you belong ...

Housing construction costs are already rising, increasi…

For Australia’s building industry, higher fuel costs since the start of the Middle East war have...

Shou Sugi Ban: The Ancient Japanese Timber Technique Tr…

There is something quietly extraordinary about a building material that has been refined over cent...

The Complete Guide to LED Installation: What Homeowners…

Electricity bills in Australia are among the highest in the developed world, and lighting accounts...

I’m close to retirement age. What are my options for dr…

Retiring well means making a series of decisions to ensure a financially secure post-work life. ...

Samsung expands B2B Mobile eXperience distribution wit…

The channel diversification reinforcers the Australian B2B division’s positive trajectory SYDNE...