The Times Australia
Fisher and Paykel Appliances
The Times World News

.

A new ‘AI scientist’ can write science papers without any human input. Here’s why that’s a problem

  • Written by Karin Verspoor, Dean, School of Computing Technologies, RMIT University, RMIT University

Scientific discovery is one of the most sophisticated human activities. First, scientists must understand the existing knowledge and identify a significant gap. Next, they must formulate a research question and design and conduct an experiment in pursuit of an answer. Then, they must analyse and interpret the results of the experiment, which may raise yet another research question.

Can a process this complex be automated? Last week, Sakana AI Labs announced[1] the creation of an “AI scientist” – an artificial intelligence system they claim can make scientific discoveries in the area of machine learning in a fully automated way.

Using generative large language models (LLMs) like those behind ChatGPT and other AI chatbots, the system can brainstorm, select a promising idea, code new algorithms, plot results, and write a paper summarising the experiment and its findings, complete with references. Sakana claims the AI tool can undertake the complete lifecycle of a scientific experiment at a cost of just US$15 per paper – less than the cost of a scientist’s lunch.

These are some big claims. Do they stack up? And even if they do, would an army of AI scientists churning out research papers with inhuman speed really be good news for science?

How a computer can ‘do science’

A lot of science is done in the open, and almost all scientific knowledge has been written down somewhere (or we wouldn’t have a way to “know” it). Millions of scientific papers are freely available online in repositories such as arXiv[2] and PubMed[3].

LLMs trained with this data capture the language of science and its patterns. It is therefore perhaps not at all surprising that a generative LLM can produce something that looks like a good scientific paper – it has ingested many examples that it can copy.

What is less clear is whether an AI system can produce an interesting scientific paper. Crucially, good science requires novelty.

But is it interesting?

Scientists don’t want to be told about things that are already known. Rather, they want to learn new things, especially new things that are significantly different from what is already known. This requires judgement about the scope and value of a contribution.

The Sakana system tries to address interestingness in two ways. First, it “scores” new paper ideas for similarity to existing research (indexed in the Semantic Scholar[4] repository). Anything too similar is discarded.

Second, Sakana’s system introduces a “peer review” step – using another LLM to judge the quality and novelty of the generated paper. Here again, there are plenty of examples of peer review online on sites such as openreview.net[5] that can guide how to critique a paper. LLMs have ingested these, too.

AI may be a poor judge of AI output

Feedback is mixed on Sakana AI’s output. Some have described it as producing “endless scientific slop[6]”.

Even the system’s own review of its outputs judges the papers weak at best. This is likely to improve as the technology evolves, but the question of whether automated scientific papers are valuable remains.

The ability of LLMs to judge the quality of research is also an open question. My own work (soon to be published in Research Synthesis Methods[7]) shows LLMs are not great at judging the risk of bias in medical research studies, though this too may improve over time.

Sakana’s system automates discoveries in computational research, which is much easier than in other types of science that require physical experiments. Sakana’s experiments are done with code, which is also structured text that LLMs can be trained to generate.

AI tools to support scientists, not replace them

AI researchers have been developing systems to support science for decades. Given the huge volumes of published research, even finding publications relevant to a specific scientific question can be challenging.

Specialised search tools make use of AI to help scientists find and synthesise existing work. These include the above-mentioned Semantic Scholar, but also newer systems such as Elicit[8], Research Rabbit[9], scite[10] and Consensus[11].

Text mining tools such as PubTator[12] dig deeper into papers to identify key points of focus, such as specific genetic mutations and diseases, and their established relationships. This is especially useful for curating and organising scientific information.

Machine learning has also been used to support the synthesis and analysis of medical evidence, in tools such as Robot Reviewer[13]. Summaries that compare and contrast claims in papers from Scholarcy[14] help to perform literature reviews.

All these tools aim to help scientists do their jobs more effectively, not to replace them.

AI research may exacerbate existing problems

While Sakana AI states[15] it doesn’t see the role of human scientists diminishing, the company’s vision of “a fully AI-driven scientific ecosystem” would have major implications for science.

One concern is that, if AI-generated papers flood the scientific literature, future AI systems may be trained on AI output and undergo model collapse[16]. This means they may become increasingly ineffectual at innovating.

However, the implications for science go well beyond impacts on AI science systems themselves.

There are already bad actors in science, including “paper mills” churning out fake papers[17]. This problem will only get worse[18] when a scientific paper can be produced with US$15 and a vague initial prompt.

The need to check for errors in a mountain of automatically generated research could rapidly overwhelm the capacity of actual scientists. The peer review system is arguably already broken[19], and dumping more research of questionable quality into the system won’t fix it.

Science is fundamentally based on trust. Scientists emphasise the integrity of the scientific process so we can be confident our understanding of the world (and now, the world’s machines) is valid and improving.

A scientific ecosystem where AI systems are key players raises fundamental questions about the meaning and value of this process, and what level of trust we should have in AI scientists. Is this the kind of scientific ecosystem we want?

References

  1. ^ Sakana AI Labs announced (sakana.ai)
  2. ^ arXiv (arxiv.org)
  3. ^ PubMed (pubmed.ncbi.nlm.nih.gov)
  4. ^ Semantic Scholar (www.semanticscholar.org)
  5. ^ openreview.net (openreview.net)
  6. ^ endless scientific slop (arstechnica.com)
  7. ^ Research Synthesis Methods (onlinelibrary.wiley.com)
  8. ^ Elicit (elicit.com)
  9. ^ Research Rabbit (www.researchrabbit.ai)
  10. ^ scite (scite.ai)
  11. ^ Consensus (consensus.app)
  12. ^ PubTator (www.ncbi.nlm.nih.gov)
  13. ^ Robot Reviewer (www.robotreviewer.net)
  14. ^ Scholarcy (www.scholarcy.com)
  15. ^ states (sakana.ai)
  16. ^ model collapse (www.nature.com)
  17. ^ fake papers (www.nature.com)
  18. ^ get worse (www.nature.com)
  19. ^ already broken (theconversation.com)

Read more https://theconversation.com/a-new-ai-scientist-can-write-science-papers-without-any-human-input-heres-why-thats-a-problem-237029

Times Magazine

Can bigger-is-better ‘scaling laws’ keep AI improving forever? History says we can’t be too sure

OpenAI chief executive Sam Altman – perhaps the most prominent face of the artificial intellig...

A backlash against AI imagery in ads may have begun as brands promote ‘human-made’

In a wave of new ads, brands like Heineken, Polaroid and Cadbury have started hating on artifici...

Home batteries now four times the size as new installers enter the market

Australians are investing in larger home battery set ups than ever before with data showing the ...

Q&A with Freya Alexander – the young artist transforming co-working spaces into creative galleries

As the current Artist in Residence at Hub Australia, Freya Alexander is bringing colour and creativi...

This Christmas, Give the Navman Gift That Never Stops Giving – Safety

Protect your loved one’s drives with a Navman Dash Cam.  This Christmas don’t just give – prote...

Yoto now available in Kmart and The Memo, bringing screen-free storytelling to Australian families

Yoto, the kids’ audio platform inspiring creativity and imagination around the world, has launched i...

The Times Features

Why the Mortgage Industry Needs More Women (And What We're Actually Doing About It)

I've been in fintech and the mortgage industry for about a year and a half now. My background is i...

Inflation jumps in October, adding to pressure on government to make budget savings

Annual inflation rose[1] to a 16-month high of 3.8% in October, adding to pressure on the govern...

Transforming Addiction Treatment Marketing Across Australasia & Southeast Asia

In a competitive and highly regulated space like addiction treatment, standing out online is no sm...

Aiper Scuba X1 Robotic Pool Cleaner Review: Powerful Cleaning, Smart Design

If you’re anything like me, the dream is a pool that always looks swimmable without you having to ha...

YepAI Emerges as AI Dark Horse, Launches V3 SuperAgent to Revolutionize E-commerce

November 24, 2025 – YepAI today announced the launch of its V3 SuperAgent, an enhanced AI platf...

What SMEs Should Look For When Choosing a Shared Office in 2026

Small and medium-sized enterprises remain the backbone of Australia’s economy. As of mid-2024, sma...

Anthony Albanese Probably Won’t Lead Labor Into the Next Federal Election — So Who Will?

As Australia edges closer to the next federal election, a quiet but unmistakable shift is rippli...

Top doctors tip into AI medtech capital raise a second time as Aussie start up expands globally

Medow Health AI, an Australian start up developing AI native tools for specialist doctors to  auto...

Record-breaking prize home draw offers Aussies a shot at luxury living

With home ownership slipping out of reach for many Australians, a growing number are snapping up...