The Times Australia
Google AI
The Times World News

.

Will AI decide if you get your next job? Without legal regulation, you may never even know

  • Written by Natalie Sheard, Lawyer and PhD Candidate, La Trobe University
Will AI decide if you get your next job? Without legal regulation, you may never even know

The use of artificial intelligence (AI) and other automated decision-making tools in recruitment is on the rise among Australian organisations. However, research shows these tools may be unreliable and discriminatory, and in some cases rely on discredited science.

At present, Australia has no specific laws to regulate how these tools operate or how organisations may use them.

The closest thing we have is new guidance[1] for employers in the public sector, issued by the Merit Protection Commissioner after overturning several automated promotion decisions.

A first step

The commissioner reviews promotion decisions[2] in the Australian public sector to make sure they are lawful, fair and reasonable. In the 2021-22 financial year, Commissioner Linda Waugh overturned 11 promotion decisions[3] made by government agency Services Australia in a single recruitment round.

These decisions were made using a new automated process that required applicants to pass through a sequence of AI assessments, including psychometric testing, questionnaires and self-recorded video responses. The commissioner found this process, which involved no human decision-making or review, led to meritorious applicants missing out on promotions.

Read more: Algorithms can decide your marks, your work prospects and your financial security. How do you know they're fair?[4]

The commissioner has now issued guidance material[5] for Australian government departments on how to choose and use AI recruitment tools.

This is the first official guidance given to employers in Australia. It warns that not all AI recruitment tools on the market here have been thoroughly tested, nor are they guaranteed to be completely unbiased.

AI recruitment tools risky and unregulated

AI tools are used to automate or assist recruiters with sourcing, screening and onboarding job applicants. By one estimate[6], more than 250 commercial AI recruitment tools are available in Australia, including CV screening and video assessment.

A recent survey[7] by researchers at Monash University and the Diversity Council of Australia found one in three Australian organisations have used AI in recruitment recently.

The use of AI recruitment tools is a “high risk[8]” activity. By affecting decisions related to employment, these tools may impact the human rights of job seekers and risk locking disadvantaged groups out of employment opportunities.

Australia has no specific legislation regulating the use of these tools. Australia’s Department of Industry has published AI Ethics Principles[9], but these are not legally binding. Existing laws, such as the Privacy Act and anti-discrimination legislation, are in urgent need of reform.

Unreliable and discriminatory?

AI recruitment tools involve new and developing technologies. They may be unreliable[10] and there are well-publicised examples[11] of discrimination against historically disadvantaged groups.

AI recruitment tools may discriminate against these groups[12] when their members are missing from the datasets on which AI is trained, or when discriminatory structures, practices or attitudes are transmitted to these tools in their development or deployment.

There is currently no standard test that identifies when an AI recruitment tool is discriminatory. Further, as these tools are often made outside Australia, they are not attuned to Australian law or demographics. For example, it is very likely training datasets do not include Australia’s First Nations peoples.

Lack of safeguards

AI recruitment tools used by and on behalf of employers in Australia lack adequate safeguards.

Human rights risk and impact assessments are not required prior to deployment. Monitoring and evaluation once they are in use may not occur. Job seekers lack meaningful opportunities to provide input on their use.

While the vendors of these tools may conduct internal testing and auditing, the results are often not publicly available. Independent external auditing is rare.

Power imbalance

Job seekers are at a considerable disadvantage when employers use these tools. They may be invisible and inscrutable, and they are changing hiring practices in ways that are not well understood.

Job seekers have no legal right to be told[13] when AI is used to assess them in the hiring process. Nor are they required to be given an explanation of how an AI recruitment tool will assess them.

Read more: Artificial intelligence can deepen social inequality. Here are 5 ways to help prevent this[14]

My research has found this is particularly problematic for job seekers with disabilities. For example, job seekers with low vision or limited manual dexterity may not know they will be assessed on the speed of their responses until it is too late.

Job seekers in Australia also lack the protection available to their counterparts in the European Union, who have the right not to be subjected to a fully automated recruitment decision[15].

Facial analysis

The use of video assessment tools, like those used by Services Australia, is particularly concerning. Many of these AI tools rely on facial analysis, which uses facial features and movements to infer behavioural, emotional and character traits.

This type of analysis has been scientifically discredited[16]. One prominent vendor, HireVue, was forced to cease the use of facial analysis in its AI tool as a result of a formal complaint in the United States[17].

What’s next?

The Services Australia example highlights the urgent need for a regulatory response. The Australian government is currently consulting on the regulation of AI and automated decision-making[18].

We can hope that new regulations will address the many issues with the use of AI tools in recruitment. Until legal protections are in place, it might be best to hold off on the use of these tools to screen job seekers.

References

  1. ^ new guidance (www.mpc.gov.au)
  2. ^ reviews promotion decisions (www.mpc.gov.au)
  3. ^ overturned 11 promotion decisions (www.transparency.gov.au)
  4. ^ Algorithms can decide your marks, your work prospects and your financial security. How do you know they're fair? (theconversation.com)
  5. ^ guidance material (theconversation.com)
  6. ^ one estimate (www.dca.org.au)
  7. ^ recent survey (www.dca.org.au)
  8. ^ high risk (artificialintelligenceact.eu)
  9. ^ AI Ethics Principles (www.industry.gov.au)
  10. ^ may be unreliable (www.youtube.com)
  11. ^ well-publicised examples (www.reuters.com)
  12. ^ discriminate against these groups (papers.ssrn.com)
  13. ^ no legal right to be told (search.informit.org)
  14. ^ Artificial intelligence can deepen social inequality. Here are 5 ways to help prevent this (theconversation.com)
  15. ^ the right not to be subjected to a fully automated recruitment decision (gdpr-info.eu)
  16. ^ scientifically discredited (journals.sagepub.com)
  17. ^ formal complaint in the United States (www.wired.com)
  18. ^ regulation of AI and automated decision-making (consult.industry.gov.au)

Read more https://theconversation.com/will-ai-decide-if-you-get-your-next-job-without-legal-regulation-you-may-never-even-know-196282

Times Magazine

Epson launches ELPCS01 mobile projector cart

Designed for the EB-810E[1] projector and provides easy setup for portable displays in flexible ...

Governance Models for Headless CMS in Large Organizations

Where headless CMS is adopted by large enterprises, governance is the single most crucial factor d...

Narwal Freo Z10 Robotic Vacuum and Mop Cleaner

Narwal Freo Z10 Robotic Vacuum and Mop Cleaner  Rating: ★★★★☆ (4.4/5) Category: Premium Robot ...

Shark launches SteamSpot - the shortcut for everyday floor mess

Shark introduces the Shark SteamSpot Steam Mop, a lightweight steam mop designed to make everyda...

Game Together, Stay Together: Logitech G Reveals Gaming Couples Enjoy Higher Relationship Satisfaction

With Valentine’s Day right around the corner, many lovebirds across Australia are planning for the m...

AI threatens to eat business software – and it could change the way we work

In recent weeks, a range of large “software-as-a-service” companies, including Salesforce[1], Se...

The Times Features

To Make Your Home & Garden Stand Out In Moorabbin – Try These Excellent Ideas.

We shouldn’t always be ‘trying to keep up with the Joneses’, but it is a common human trait to wan...

Travel Trends: Where Are Australians Going in 2026?

For Australians, travel has always been more than just a holiday. It is a cultural habit, a reward...

Applications Open for TasPorts Industry Support Program

TasPorts has opened applications for its 2026 Industry Support Program, offering $100,000 in f...

STATEMENT FROM DEPUTY LEADER OF THE NATIONALS DARREN CHESTER

I'm incredibly honoured to have been elected Deputy Leader of The Nationals Federal Parliamentary ...

Grill'd Oscar Piastri's burger just landed at Coles

Grill’d is putting the pedal down with the launch of an all-new Oscar Piastri Burger on 10 Febru...

Tasmanian MP Andrew Wilkie has issued a statement regard Robodebt

 A STATEMENT ON NACC ROBODEBT FINDINGS - Andrew Wilkie The National Anti-Corruption Commission h...

Can exercise reduce period pain? And what kind is best?

Having your period can be a painful experience. Period pain, also known as dysmenorrhea, is a...

Tasmania in 2026: Opportunity, Pressure and the Island State’s Defining Moment

Tasmania has long held a unique place in the Australian story. It is a state known for natural b...

Middle East war set to push inflation higher than forecast, warns RBA deputy governor

The Reserve Bank’s Deputy Governor Andrew Hauser says inflation in Australia looks likely to be ...