The Times Australia
Fisher and Paykel Appliances
The Times World News

.

Feds are increasing use of facial recognition systems – despite calls for a moratorium

  • Written by James Hendler, Professor of Computer, Web and Cognitive Sciences, Rensselaer Polytechnic Institute
Feds are increasing use of facial recognition systems – despite calls for a moratorium

Despite growing opposition, the U.S. government is on track to increase its use of controversial facial recognition technology.

The U.S. Government Accountability Office released a report[1] on Aug. 24, 2021, detailing current and planned use of facial recognition technology by federal agencies. The GAO surveyed 24 departments and agencies[2] – from the Department of Defense to the Small Business Administration – and found that 18 reported using the technology and 10 reported plans to expand their use of it[3].

The report comes more than a year after the U.S. Technology Policy Committee[4] of the Association for Computing Machinery, the world’s largest educational and scientific computing society, called for an immediate halt[5] to virtually all government use of facial recognition technology.

The U.S. Technology Policy Committee is one of numerous groups and prominent figures, including the ACLU[6], the American Library Association[7] and the United Nations Special Rapporteur on Freedom of Opinion and Expression[8], to call for curbs on use of the technology. A common theme of this opposition is the lack of standards and regulations for facial recognition technology.

A year ago, Amazon, IBM and Microsoft also announced that they would stop selling facial recognition technology[9] to police departments pending federal regulation of the technology. Congress is weighing a moratorium[10] on government use of the technology. Some cities and states, notably Maine[11], have introduced restrictions.

Why computing experts say no

The Association for Computing Machinery’s U.S. Technology Policy Committee, which issued the call for a moratorium, includes computing professionals from academia, industry and government, a number of whom were actively involved in the development or analysis of the technology. As chair of the committee at the time the statement was issued and as a computer science researcher[12], I can explain what prompted our committee to recommend this ban and, perhaps more significantly, what it would take for the committee to rescind its call.

If your cellphone doesn’t recognize your face and makes you type in your passcode, or if the photo-sorting software you’re using misidentifies a family member, no real harm is done. On the other hand, if you become liable for arrest or denied entrance to a facility because the recognition algorithms are imperfect, the impact can be drastic.

The statement we wrote outlines principles for the use of facial recognition technologies in these consequential applications. The first and most critical of these is the need to understand the accuracy of these systems. One of the key problems with these algorithms is that they perform differently for different ethnic groups[13].

An evaluation of facial recognition vendors[14] by the U.S. National Institute of Standards and Technology found that the majority of the systems tested had clear differences in their ability to match two images of the same person when one ethnic group was compared with another. Another study found the algorithms are more accurate for lighter-skinned males[15] than for darker-skinned females. Researchers are also exploring how other features, such as age, disease and disability status[16], affect these systems. These studies are also turning up disparities[17].

MIT’s Joy Buolamwini explains her study finding racial and gender bias in facial recognition technology.

A number of other features affect the performance of these algorithms. Consider the difference between how you might look in a nice family photo you have shared on social media versus a picture of you taken by a grainy security camera, or a moving police car, late on a misty night. Would a system trained on the former perform well in the latter context? How lighting, weather, camera angle and other factors affect these algorithms[18] is still an open question.

In the past, systems that matched fingerprints[19] or DNA traces[20] had to be formally evaluated, and standards set, before they were trusted for use by the police and others. Until facial recognition algorithms can meet similar standards – and researchers and regulators truly understand how the context in which the technology is used affects its accuracy – the systems shouldn’t be used in applications that can have serious consequences for people’s lives.

Transparency and accountability

It’s also important that organizations using facial recognition provide some form of meaningful advanced and ongoing public notice. If a system can result in your losing your liberty or your life, you should know it is being used. In the U.S., this has been a principle for the use of many potentially harmful technologies, from speed cameras to video surveillance[21], and the USTPC’s position is that facial recognition systems should be held to the same standard.

To get transparency, there also must be rules that govern the collection and use of the personal information that underlies the training of facial recognition systems. The company Clearview AI, which now has software in use by police agencies around the world[22], is a case in point[23]. The company collected its data – photos of individuals’ faces – with no notification.

PBS Nova explains Clearview AI’s massive database of images of people.

Clearview AI collected data from many different applications, vendors and systems, taking advantage of the lax laws controlling such collection[24]. Kids who post videos of themselves on TikTok, users who tag friends in photos on Facebook, consumers who make purchases with Venmo, people who upload videos to YouTube and many others all create images that can be linked to their names and scraped from these applications by companies like Clearview AI.

Are you in the dataset Clearview uses? You have no way to know. The ACM’s position is that you should have a right to know, and that governments should put limits on how this data is collected, stored and used.

In 2017, the Association for Computing Machinery U.S. Technology Policy Committee and its European counterpart released a joint statement[25] on algorithms for automated decision-making about individuals that can result in harmful discrimination. In short, we called for policymakers to hold institutions using analytics to the same standards as for institutions where humans have traditionally made decisions, whether it be traffic enforcement or criminal prosecution.

This includes understanding the trade-offs between the risks and benefits of powerful computational technologies when they are put into practice and having clear principles about who is liable when harms occur. Facial recognition technologies are in this category, and it’s important to understand how to measure their risks and benefits and who is responsible when they fail.

Protecting the public

One of the primary roles of governments is to manage technology risks and protect their populations. The principles the Association for Computing Machinery’s USTPC has outlined have been used in regulating transportation systems, medical and pharmaceutical products, food safety practices and many other aspects of society. The Association for Computing Machinery’s USTPC is, in short, asking that governments recognize the potential for facial recognition systems to cause significant harm to many people, through errors and bias.

These systems are still in an early stage of maturity, and there is much that researchers, government and industry don’t understand about them. Until facial recognition technologies are better understood, their use in consequential applications should be halted until they can be properly regulated.

[Get the best of The Conversation, every weekend. Sign up for our weekly newsletter[26].]

References

  1. ^ a report (www.gao.gov)
  2. ^ 24 departments and agencies (www.cfo.gov)
  3. ^ expand their use of it (www.technologyreview.com)
  4. ^ U.S. Technology Policy Committee (www.acm.org)
  5. ^ an immediate halt (www.acm.org)
  6. ^ ACLU (www.aclu.org)
  7. ^ American Library Association (www.ala.org)
  8. ^ Special Rapporteur on Freedom of Opinion and Expression (www.ohchr.org)
  9. ^ stop selling facial recognition technology (www.vox.com)
  10. ^ weighing a moratorium (www.markey.senate.gov)
  11. ^ Maine (slate.com)
  12. ^ computer science researcher (scholar.google.com)
  13. ^ perform differently for different ethnic groups (theconversation.com)
  14. ^ evaluation of facial recognition vendors (www.nist.gov)
  15. ^ more accurate for lighter-skinned males (proceedings.mlr.press)
  16. ^ disability status (sheribyrnehaber.medium.com)
  17. ^ turning up disparities (www.csis.org)
  18. ^ lighting, weather, camera angle and other factors affect these algorithms (www.csis.org)
  19. ^ fingerprints (onin.com)
  20. ^ DNA traces (www.ncbi.nlm.nih.gov)
  21. ^ video surveillance (www.law.berkeley.edu)
  22. ^ in use by police agencies around the world (www.theverge.com)
  23. ^ case in point (www.nytimes.com)
  24. ^ lax laws controlling such collection (www.nytimes.com)
  25. ^ joint statement (www.acm.org)
  26. ^ Sign up for our weekly newsletter (theconversation.com)

Read more https://theconversation.com/feds-are-increasing-use-of-facial-recognition-systems-despite-calls-for-a-moratorium-145913

Times Magazine

Can bigger-is-better ‘scaling laws’ keep AI improving forever? History says we can’t be too sure

OpenAI chief executive Sam Altman – perhaps the most prominent face of the artificial intellig...

A backlash against AI imagery in ads may have begun as brands promote ‘human-made’

In a wave of new ads, brands like Heineken, Polaroid and Cadbury have started hating on artifici...

Home batteries now four times the size as new installers enter the market

Australians are investing in larger home battery set ups than ever before with data showing the ...

Q&A with Freya Alexander – the young artist transforming co-working spaces into creative galleries

As the current Artist in Residence at Hub Australia, Freya Alexander is bringing colour and creativi...

This Christmas, Give the Navman Gift That Never Stops Giving – Safety

Protect your loved one’s drives with a Navman Dash Cam.  This Christmas don’t just give – prote...

Yoto now available in Kmart and The Memo, bringing screen-free storytelling to Australian families

Yoto, the kids’ audio platform inspiring creativity and imagination around the world, has launched i...

The Times Features

The rise of chatbot therapists: Why AI cannot replace human care

Some are dubbing AI as the fourth industrial revolution, with the sweeping changes it is propellin...

Australians Can Now Experience The World of Wicked Across Universal Studios Singapore and Resorts World Sentosa

This holiday season, Resorts World Sentosa (RWS), in partnership with Universal Pictures, Sentosa ...

Mineral vs chemical sunscreens? Science shows the difference is smaller than you think

“Mineral-only” sunscreens are making huge inroads[1] into the sunscreen market, driven by fears of “...

Here’s what new debt-to-income home loan caps mean for banks and borrowers

For the first time ever, the Australian banking regulator has announced it will impose new debt-...

Why the Mortgage Industry Needs More Women (And What We're Actually Doing About It)

I've been in fintech and the mortgage industry for about a year and a half now. My background is i...

Inflation jumps in October, adding to pressure on government to make budget savings

Annual inflation rose[1] to a 16-month high of 3.8% in October, adding to pressure on the govern...

Transforming Addiction Treatment Marketing Across Australasia & Southeast Asia

In a competitive and highly regulated space like addiction treatment, standing out online is no sm...

Aiper Scuba X1 Robotic Pool Cleaner Review: Powerful Cleaning, Smart Design

If you’re anything like me, the dream is a pool that always looks swimmable without you having to ha...

YepAI Emerges as AI Dark Horse, Launches V3 SuperAgent to Revolutionize E-commerce

November 24, 2025 – YepAI today announced the launch of its V3 SuperAgent, an enhanced AI platf...