The Times Australia
Mirvac Harbourside
The Times World News

.

OpenAI’s data hunger raises privacy concerns

  • Written by Uri Gal, Professor in Business Information Systems, University of Sydney



Last month, OpenAI came out against[1] a yet-to-be enacted Californian law that aims to set basic safety standards for developers of large artificial intelligence (AI) models. This was a change of posture for the company, whose chief executive Sam Altman has previously spoken in support[2] of AI regulation.

The former nonprofit organisation, which shot to prominence in 2022 with the release of ChatGPT, is now valued at up to US$150 billion[3]. It remains at the forefront of AI development, with the release last week of a new “reasoning” model[4] designed to tackle more complex tasks.

The company has made several moves in recent months suggesting a growing appetite for data acquisition. This isn’t just the text or images used for training current generative AI tools, but may also include intimate data related to online behaviour, personal interactions and health.

There is no evidence OpenAI plans to bring these different streams of data together, but doing so would offer strong commercial benefits. Even the possibility of access to such wide-ranging information raises significant questions about privacy and the ethical implications of centralised data control.

Media deals

This year, OpenAI has signed multiple partnerships[5] with media companies including Time magazine, the Financial Times, Axel Springer, Le Monde, Prisa Media, and most recently Condé Nast, owner of the likes of Vogue, The New Yorker, Vanity Fair and Wired.

The partnerships grant OpenAI access to large amounts of content. OpenAI’s products may also be used to analyse user behaviour and interaction metrics such as reading habits, preferences, and engagement patterns across platforms.

If OpenAI gained access to this data, the company could gain a comprehensive understanding of how users engage with various types of content, which could be used for in-depth user profiling and tracking.

Video, biometrics and health

OpenAI has also invested in a webcam startup called Opal[6]. The aim is to enhance the cameras with advanced AI capabilities.

Video footage collected by AI-powered webcams could translate to more sensitive biometric data, such as facial expressions and inferred psychological states.

In July, OpenAI and Thrive Global launched Thrive AI Health. The company says it will use AI to “hyper-personalise and scale behaviour change[7]” in health.

While Thrive AI Health says it will have “robust privacy and security guardrails”, it is unclear what these will look like.

Previous AI health projects have involved extensive sharing of personal data, such as a partnership between Microsoft and Providence Health in the United States and another between Google DeepMind and the Royal Free London NHS Foundation Trust in the United Kingdom. In the latter case, DeepMind faced legal action[8] for its use of private health data.

Sam Altman’s eyeball-scanning side project

Altman also has investments in other data-hungry ventures, most notably a controversial cryptocurrency project called WorldCoin (which he cofounded). WorldCoin aims to create a global financial network and identification system using biometric identification, specifically iris scans.

The company claims it has already scanned the eyeballs of more than 6.5 million people[9] across almost 40 countries. Meanwhile, more than a dozen jurisdictions have either suspended its operations or scrutinised its data processing.

Bavarian authorities are currently deliberating on whether Worldcoin complies with European data privacy regulations[10]. A negative ruling could see the company barred from operating in Europe.

The main concerns being investigated include the collection and storage of sensitive biometric data.

Why does this matter?

Existing AI models such as OpenAI’s flagship GPT-4o have largely been trained on publicly available data[11] from the internet. However, future models will need more data – and it’s getting harder to come by[12].

Last year, the company said[13] it wanted AI models “to deeply understand all subject matters, industries, cultures, and languages”, which would require “as broad a training dataset as possible”.

In this context, OpenAI’s pursuit of media partnerships, investments in biometric and health data collection technologies, and the CEO’s links to controversial projects such as Worldcoin, begin to paint a concerning picture.

By gaining access to vast amounts of user data, OpenAI is positioning itself to build the next wave of AI models – but privacy may be a casualty.

The risks are multifaceted. Large collections of personal data are vulnerable to breaches and misuse, such as the Medisecure data breach[14] in which almost half of Australians had their personal and medical data stolen.

The potential for large-scale data consolidation also raises concerns about profiling and surveillance. Again, there is no evidence that OpenAI currently plans to engage in such practices.

However, OpenAI’s privacy policies have been less than perfect in the past[15]. Tech companies more broadly also have a long history of questionable data practices[16].

It is not difficult to imagine a scenario in which centralised control over many kinds of data would let OpenAI exert significant influence over people, in both personal and public domains.

Will safety take a back seat?

OpenAI’s recent history does little to assuage safety and privacy concerns. In November 2023, Altman was temporarily ousted[17] as chief executive, reportedly due to internal conflicts over the company’s strategic direction.

Altman has been a strong advocate for the rapid commercialisation and deployment of AI technologies. He has reportedly often prioritised growth and market penetration over safety measures[18].

Altman’s removal from the role was brief, followed by a swift reinstatement[19] and a significant shakeup of OpenAI’s board. This suggests the company’s leadership now endorses his aggressive approach to AI deployment, despite potential risks.

Against this backdrop, the implications of OpenAI’s recent opposition to the California bill extend beyond a single policy disagreement. The anti-regulation stance suggests a troubling trend.

OpenAI did not respond to The Conversation’s request for comment before deadline.

References

  1. ^ came out against (techcrunch.com)
  2. ^ spoken in support (time.com)
  3. ^ up to US$150 billion (www.nytimes.com)
  4. ^ a new “reasoning” model (openai.com)
  5. ^ multiple partnerships (www.theguardian.com)
  6. ^ a webcam startup called Opal (www.theinformation.com)
  7. ^ hyper-personalise and scale behaviour change (www.prnewswire.com)
  8. ^ faced legal action (www.bbc.com)
  9. ^ scanned the eyeballs of more than 6.5 million people (worldcoin.org)
  10. ^ whether Worldcoin complies with European data privacy regulations (techcrunch.com)
  11. ^ publicly available data (www.ncbi.nlm.nih.gov)
  12. ^ getting harder to come by (www.dataprovenance.org)
  13. ^ said (openai.com)
  14. ^ Medisecure data breach (www.smh.com.au)
  15. ^ less than perfect in the past (www.forbes.com)
  16. ^ questionable data practices (www.taylorfrancis.com)
  17. ^ temporarily ousted (arstechnica.com)
  18. ^ over safety measures (www.washingtonpost.com)
  19. ^ followed by a swift reinstatement (www.reuters.com)

Read more https://theconversation.com/openais-data-hunger-raises-privacy-concerns-237448

Mirvac Harbourside

Times Magazine

YepAI Joins Victoria's AI Trade Mission to Singapore for Big Data & AI World Asia 2025

YepAI, a Melbourne-based leader in enterprise artificial intelligence solutions, announced today...

Building a Strong Online Presence with Katoomba Web Design

Katoomba web design is more than just creating a website that looks good—it’s about building an onli...

September Sunset Polo

International Polo Tour To Bridge Historic Sport, Life-Changing Philanthropy, and Breath-Taking Beau...

5 Ways Microsoft Fabric Simplifies Your Data Analytics Workflow

In today's data-driven world, businesses are constantly seeking ways to streamline their data anal...

7 Questions to Ask Before You Sign IT Support Companies in Sydney

Choosing an IT partner can feel like buying an insurance policy you hope you never need. The right c...

Choosing the Right Legal Aid Lawyer in Sutherland Shire: Key Considerations

Legal aid services play an essential role in ensuring access to justice for all. For people in t...

The Times Features

Macquarie Bank Democratises Agentic AI, Scaling Customer Innovation with Gemini Enterprise

Macquarie’s Banking and Financial Services group (Macquarie Bank), in collaboration with Google ...

Do kids really need vitamin supplements?

Walk down the health aisle of any supermarket and you’ll see shelves lined with brightly packa...

Why is it so shameful to have missing or damaged teeth?

When your teeth and gums are in good condition, you might not even notice their impact on your...

Australian travellers at risk of ATM fee rip-offs according to new data from Wise

Wise, the global technology company building the smartest way to spend and manage money internat...

Does ‘fasted’ cardio help you lose weight? Here’s the science

Every few years, the concept of fasted exercise training pops up all over social media. Faste...

How Music and Culture Are Shaping Family Road Trips in Australia

School holiday season is here, and Aussies aren’t just hitting the road - they’re following the musi...

The Role of Spinal Physiotherapy in Recovery and Long-Term Wellbeing

Back pain and spinal conditions are among the most common reasons people seek medical support, oft...

Italian Lamb Ragu Recipe: The Best Ragù di Agnello for Pasta

Ciao! It’s Friday night, and the weekend is calling for a little Italian magic. What’s better than t...

It’s OK to use paracetamol in pregnancy. Here’s what the science says about the link with autism

United States President Donald Trump has urged pregnant women[1] to avoid paracetamol except in ...