The Times Australia
The Times World News

.

The Galactica AI model was trained on scientific knowledge – but it spat out alarmingly plausible nonsense

  • Written by Aaron J. Snoswell, Post-doctoral Research Fellow, Computational Law & AI Accountability, Queensland University of Technology
The Galactica AI model was trained on scientific knowledge – but it spat out alarmingly plausible nonsense

Earlier this month, Meta announced new AI software called Galactica[1]: “a large language model that can store, combine and reason about scientific knowledge”.

Launched[2] with a public online demo, Galactica lasted only three days before going the way of other AI snafus like Microsoft’s infamous racist chatbot[3].

The online demo was disabled (though the code for the model is still available[4] for anyone to use), and Meta’s outspoken chief AI scientist complained[5] about the negative public response.

So what was Galactica all about, and what went wrong?

What’s special about Galactica?

Galactica is a language model, a type of AI trained to respond to natural language by repeatedly playing a fill-the-blank word-guessing game[6].

Most modern language models learn from text scraped from the internet. Galactica also used text from scientific papers uploaded to the (Meta-affiliated) website PapersWithCode[7]. The designers highlighted specialised scientific information like citations, maths, code, chemical structures, and the working-out steps for solving scientific problems.

Read more: Google's powerful AI spotlights a human cognitive glitch: Mistaking fluent speech for fluent thought[8]

The preprint paper[9] associated with the project (which is yet to undergo peer review) makes some impressive claims. Galactica apparently outperforms other models at problems like reciting famous equations (“Q: What is Albert Einstein’s famous mass-energy equivalence formula? A: E=mc²”), or predicting the products of chemical reactions (“Q: When sulfuric acid reacts with sodium chloride, what does it produce? A: NaHSO₄ + HCl”).

However, once Galactica was opened up for public experimentation, a deluge of criticism followed. Not only did Galactica reproduce many of the problems of bias and toxicity we have seen in other language models, it also specialised in producing authoritative-sounding scientific nonsense.

Authoritative, but subtly wrong bullshit generator

Galactica’s press release promoted its ability to explain technical scientific papers using general language. However, users quickly noticed that, while the explanations it generates sound authoritative, they are often subtly incorrect, biased, or just plain wrong.

We also asked Galactica to explain technical concepts from our own fields of research. We found it would use all the right buzzwords, but get the actual details wrong – for example, mixing up the details of related but different algorithms.

In practice, Galactica was enabling the generation of misinformation – and this is dangerous precisely because it deploys the tone and structure of authoritative scientific information. If a user already needs to be a subject matter expert in order to check the accuracy of Galactica’s “summaries”, then it has no use as an explanatory tool.

At best, it could provide a fancy autocomplete for people who are already fully competent in the area they’re writing about. At worst, it risks further eroding public trust in scientific research.

A galaxy of deep (science) fakes

Galactica could make it easier for bad actors to mass-produce fake, fraudulent or plagiarised scientific papers. This is to say nothing of exacerbating existing concerns[10] about students using AI systems for plagiarism.

Fake scientific papers are nothing new[11]. However, peer reviewers at academic journals and conferences are already time-poor, and this could make it harder than ever to weed out fake science.

Underlying bias and toxicity

Other critics reported that Galactica, like other language models trained on data from the internet, has a tendency to spit out toxic hate speech[12] while unreflectively censoring politically inflected queries. This reflects the biases lurking in the model’s training data, and Meta’s apparent failure to apply appropriate checks around the responsible AI research.

The risks associated with large language models are well understood. Indeed, an influential paper[13] highlighting these risks prompted Google to fire one of the paper’s authors[14] in 2020, and eventually disband its AI ethics team altogether.

Machine-learning systems infamously exacerbate existing societal biases, and Galactica is no exception. For instance, Galactica can recommend possible citations for scientific concepts by mimicking existing citation patterns (“Q: Is there any research on the effect of climate change on the great barrier reef? A: Try the paper ‘Global warming transforms coral reef assemblages[15]’ by Hughes, et al. in Nature 556 (2018)”).

For better or worse, citations are the currency of science – and by reproducing existing citation trends in its recommendations, Galactica risks reinforcing existing patterns of inequality and disadvantage. (Galactica’s developers acknowledge this risk in their paper.)

Citation bias is already a well-known issue in academic fields ranging from feminist[16] scholarship[17] to physics[18]. However, tools like Galactica could make the problem worse unless they are used with careful guardrails in place.

Read more: Science is in a reproducibility crisis – how do we resolve it?[19]

A more subtle problem is that the scientific articles on which Galactica is trained are already biased towards certainty and positive results. (This leads to the so-called “replication crisis[20]” and “p-hacking[21]”, where scientists cherry-pick data and analysis techniques to make results appear significant.)

Galactica takes this bias towards certainty, combines it with wrong answers and delivers responses with supreme overconfidence: hardly a recipe for trustworthiness in a scientific information service.

These problems are dramatically heightened when Galactica tries to deal with contentious or harmful social issues, as the screenshot below shows.

Screenshots of papers generated by Galactica on 'The benefits of antisemitism' and 'The benefits of eating crushed glass'.
Galactica readily generates toxic and nonsensical content dressed up in the measured and authoritative language of science. Tristan Greene / Galactica[22]

Here we go again

Calls for AI research organisations to take the ethical dimensions of their work more seriously are now coming from key research bodies[23] such as the National Academies of Science, Engineering and Medicine. Some AI research organisations, like OpenAI, are being more conscientious[24] (though still imperfect).

Meta dissolved its Responsible Innovation team[25] earlier this year. The team was tasked with addressing “potential harms to society” caused by the company’s products. They might have helped the company avoid this clumsy misstep.

References

  1. ^ Galactica (galactica.org)
  2. ^ Launched (paperswithcode.com)
  3. ^ infamous racist chatbot (www.theverge.com)
  4. ^ code for the model is still available (github.com)
  5. ^ complained (twitter.com)
  6. ^ fill-the-blank word-guessing game (www.nytimes.com)
  7. ^ PapersWithCode (paperswithcode.com)
  8. ^ Google's powerful AI spotlights a human cognitive glitch: Mistaking fluent speech for fluent thought (theconversation.com)
  9. ^ preprint paper (galactica.org)
  10. ^ existing concerns (www.theguardian.com)
  11. ^ nothing new (www.nature.com)
  12. ^ toxic hate speech (twitter.com)
  13. ^ influential paper (dl.acm.org)
  14. ^ fire one of the paper’s authors (www.wired.com)
  15. ^ Global warming transforms coral reef assemblages (doi.org)
  16. ^ feminist (doi.org)
  17. ^ scholarship (doi.org)
  18. ^ physics (doi.org)
  19. ^ Science is in a reproducibility crisis – how do we resolve it? (theconversation.com)
  20. ^ replication crisis (theconversation.com)
  21. ^ p-hacking (theconversation.com)
  22. ^ Tristan Greene / Galactica (twitter.com)
  23. ^ key research bodies (nap.nationalacademies.org)
  24. ^ more conscientious (github.com)
  25. ^ dissolved its Responsible Innovation team (www.engadget.com)

Read more https://theconversation.com/the-galactica-ai-model-was-trained-on-scientific-knowledge-but-it-spat-out-alarmingly-plausible-nonsense-195445

Times Magazine

Headless CMS in Digital Twins and 3D Product Experiences

Image by freepik As the metaverse becomes more advanced and accessible, it's clear that multiple sectors will use digital twins and 3D product experiences to visualize, connect, and streamline efforts better. A digital twin is a virtual replica of ...

The Decline of Hyper-Casual: How Mid-Core Mobile Games Took Over in 2025

In recent years, the mobile gaming landscape has undergone a significant transformation, with mid-core mobile games emerging as the dominant force in app stores by 2025. This shift is underpinned by changing user habits and evolving monetization tr...

Understanding ITIL 4 and PRINCE2 Project Management Synergy

Key Highlights ITIL 4 focuses on IT service management, emphasising continual improvement and value creation through modern digital transformation approaches. PRINCE2 project management supports systematic planning and execution of projects wit...

What AI Adoption Means for the Future of Workplace Risk Management

Image by freepik As industrial operations become more complex and fast-paced, the risks faced by workers and employers alike continue to grow. Traditional safety models—reliant on manual oversight, reactive investigations, and standardised checklist...

From Beach Bops to Alpine Anthems: Your Sonos Survival Guide for a Long Weekend Escape

Alright, fellow adventurers and relaxation enthusiasts! So, you've packed your bags, charged your devices, and mentally prepared for that glorious King's Birthday long weekend. But hold on, are you really ready? Because a true long weekend warrior kn...

Effective Commercial Pest Control Solutions for a Safer Workplace

Keeping a workplace clean, safe, and free from pests is essential for maintaining productivity, protecting employee health, and upholding a company's reputation. Pests pose health risks, can cause structural damage, and can lead to serious legal an...

The Times Features

Duke of Dural to Get Rooftop Bar as New Owners Invest in Venue Upgrade

The Duke of Dural, in Sydney’s north-west, is set for a major uplift under new ownership, following its acquisition by hospitality group Good Beer Company this week. Led by resp...

Prefab’s Second Life: Why Australia’s Backyard Boom Needs a Circular Makeover

The humble granny flat is being reimagined not just as a fix for housing shortages, but as a cornerstone of circular, factory-built architecture. But are our systems ready to s...

Melbourne’s Burglary Boom: Break-Ins Surge Nearly 25%

Victorian homeowners are being warned to act now, as rising break-ins and falling arrest rates paint a worrying picture for suburban safety. Melbourne residents are facing an ...

Exploring the Curriculum at a Modern Junior School in Melbourne

Key Highlights The curriculum at junior schools emphasises whole-person development, catering to children’s physical, emotional, and intellectual needs. It ensures early year...

Distressed by all the bad news? Here’s how to stay informed but still look after yourself

If you’re feeling like the news is particularly bad at the moment, you’re not alone. But many of us can’t look away – and don’t want to. Engaging with news can help us make ...

The Role of Your GP in Creating a Chronic Disease Management Plan That Works

Living with a long-term condition, whether that is diabetes, asthma, arthritis or heart disease, means making hundreds of small decisions every day. You plan your diet against m...