The Times Australia
Google AI
The Times World News

.

AI-assisted writing is quietly booming in academic journals. Here’s why that’s OK

  • Written by Julian Koplin, Lecturer in Bioethics, Monash University & Honorary fellow, Melbourne Law School, Monash University
AI-assisted writing is quietly booming in academic journals. Here’s why that’s OK

If you search Google Scholar for the phrase “as an AI language model[1]”, you’ll find plenty of AI research literature and also some rather suspicious results. For example, one paper[2] on agricultural technology says:

As an AI language model, I don’t have direct access to current research articles or studies. However, I can provide you with an overview of some recent trends and advancements …

Obvious gaffes like this aren’t the only signs that researchers are increasingly turning to generative AI tools when writing up their research. A recent study[3] examined the frequency of certain words in academic writing (such as “commendable”, “meticulously” and “intricate”), and found they became far more common after the launch of ChatGPT – so much so that 1% of all journal articles published in 2023 may have contained AI-generated text.

(Why do AI models overuse these words? There is speculation[4] it’s because they are more common in English as spoken in Nigeria, where key elements of model training often occur.)

The aforementioned study also looks at preliminary data from 2024, which indicates that AI writing assistance is only becoming more common. Is this a crisis for modern scholarship, or a boon for academic productivity?

Who should take credit for AI writing?

Many people are worried by the use of AI in academic papers. Indeed, the practice has been described as “contaminating[5]” scholarly literature.

Some argue that using AI output amounts to plagiarism. If your ideas are copy-pasted from ChatGPT, it is questionable whether you really deserve credit for them.

But there are important differences between “plagiarising” text authored by humans and text authored by AI. Those who plagiarise humans’ work receive credit for ideas that ought to have gone to the original author.

By contrast, it is debatable whether AI systems like ChatGPT can have ideas, let alone deserve credit for them. An AI tool is more like[6] your phone’s autocomplete function than a human researcher.

The question of bias

Another worry is that AI outputs might be biased in ways that could seep into the scholarly record. Infamously, older language models tended to portray[7] people who are female, black and/or gay in distinctly unflattering ways, compared with people who are male, white and/or straight.

This kind of bias is less pronounced[8] in the current version of ChatGPT.

However, other studies have found a different kind[9] of bias[10] in ChatGPT and other large language models[11]: a tendency to reflect a left-liberal political ideology.

Any such bias could subtly distort scholarly writing produced using these tools.

The hallucination problem

The most serious worry relates to a well-known limitation of generative AI systems: that they often make serious mistakes.

For example, when I asked ChatGPT-4 to generate an ASCII image of a mushroom, it provided me with the following output.

   .--'|
   /___^ |     .--.
       ) |    /    \
      / |   |      |
     |   `-._\    /
     \        `~~`
      `-..._____.-`

It then confidently told me I could use this image of a “mushroom” for my own purposes.

These kinds of overconfident mistakes have been referred to as “AI hallucinations[12]” and “AI bullshit[13]”. While it is easy to spot that the above ASCII image looks nothing like a mushroom (and quite a bit like a snail), it may be much harder to identify any mistakes ChatGPT makes when surveying scientific literature[14] or describing the state of a philosophical debate.

Unlike (most) humans, AI systems are fundamentally unconcerned with the truth of what they say. If used carelessly, their hallucinations could corrupt the scholarly record.

Should AI-produced text be banned?

One response to the rise of text generators has been to ban them outright. For example, Science – one of the world’s most influential academic journals – disallows any use of AI-generated text[15].

I see two problems with this approach.

The first problem is a practical one: current tools for detecting AI-generated text are highly unreliable. This includes the detector created by ChatGPT’s own developers, which was taken offline[16] after it was found to have only a 26% accuracy rate (and a 9% false positive rate[17]). Humans also make mistakes[18] when assessing whether something was written by AI.

It is also possible to circumvent AI text detectors. Online communities are actively exploring[19] how to prompt ChatGPT in ways that allow the user to evade detection. Human users can also superficially rewrite AI outputs, effectively scrubbing away the traces of AI (like its overuse of the words “commendable”, “meticulously” and “intricate”).

The second problem is that banning generative AI outright prevents us from realising these technologies’ benefits. Used well, generative AI can boost academic productivity[20] by streamlining the writing process. In this way, it could help further human knowledge. Ideally, we should try to reap these benefits while avoiding the problems.

The problem is poor quality control, not AI

The most serious problem with AI is the risk of introducing unnoticed errors, leading to sloppy scholarship. Instead of banning AI, we should try to ensure that mistaken, implausible or biased claims cannot make it onto the academic record.

After all, humans can also produce writing with serious errors, and mechanisms such as peer review often fail[21] to prevent its publication.

We need to get better at ensuring academic papers are free from serious mistakes, regardless of whether these mistakes are caused by careless use of AI or sloppy human scholarship. Not only is this more achievable than policing AI usage, it will improve the standards of academic research as a whole.

This would be (as ChatGPT might say) a commendable and meticulously intricate solution.

References

  1. ^ as an AI language model (twitter.com)
  2. ^ paper (journals.ekb.eg)
  3. ^ recent study (arxiv.org)
  4. ^ speculation (www.theguardian.com)
  5. ^ contaminating (arxiv.org)
  6. ^ more like (theconversation.com)
  7. ^ tended to portray (arxiv.org)
  8. ^ less pronounced (www.nature.com)
  9. ^ kind (arxiv.org)
  10. ^ bias (www.mdpi.com)
  11. ^ other large language models (www.maximumtruth.org)
  12. ^ AI hallucinations (theconversation.com)
  13. ^ AI bullshit (blog.practicalethics.ox.ac.uk)
  14. ^ surveying scientific literature (time.com)
  15. ^ any use of AI-generated text (www.science.org)
  16. ^ taken offline (arstechnica.com)
  17. ^ 9% false positive rate (openai.com)
  18. ^ make mistakes (www.nature.com)
  19. ^ actively exploring (www.youtube.com)
  20. ^ boost academic productivity (www.nature.com)
  21. ^ often fail (www.routledge.com)

Read more https://theconversation.com/ai-assisted-writing-is-quietly-booming-in-academic-journals-heres-why-thats-ok-229416

Times Magazine

Narwal Freo Z Ultra Robotic Vacuum and Mop Cleaner

Rating: ★★★★☆ (4.4/5)Category: Premium Robot Vacuum & Mop ComboBest for: Busy households, ha...

Shark launches SteamSpot - the shortcut for everyday floor mess

Shark introduces the Shark SteamSpot Steam Mop, a lightweight steam mop designed to make everyda...

Game Together, Stay Together: Logitech G Reveals Gaming Couples Enjoy Higher Relationship Satisfaction

With Valentine’s Day right around the corner, many lovebirds across Australia are planning for the m...

AI threatens to eat business software – and it could change the way we work

In recent weeks, a range of large “software-as-a-service” companies, including Salesforce[1], Se...

Worried AI means you won’t get a job when you graduate? Here’s what the research says

The head of the International Monetary Fund, Kristalina Georgieva, has warned[1] young people ...

How Managed IT Support Improves Security, Uptime, And Productivity

Managed IT support is a comprehensive, subscription model approach to running and protecting your ...

The Times Features

The Art of the Big Trip: Planning a Seamless Multi-Generational Getaway in Tropical North Queensland

There is a unique magic to the multi-generational holiday. It is a rare opportunity where gr...

Love Without Borders: ‘Second Marriage At First Sight’ Opens Casting Call for Melbourne Singles Willing to Relocate for Romance

Fans of Married At First Sight UK and Married At First Sight Australia are about to see the expe...

Macca’s is bringing pub-style vibes to the menu with the new Bistro Béarnaise Angus range

Two indulgent Aussie Angus burgers – plus the arrival of Kirks Lemon, Lime & Bitters – the  ...

What are your options if you can’t afford to repay your mortgage?

After just three rate cuts in 2025, interest rates have risen again[1] in Australia this year. I...

Small, realistic increases in physical activity shown to significantly reduce risk of early death

Just Five Minutes More a Day Could Prevent Thousands of Deaths, Landmark Study Finds Small, rea...

Inside One Global resorts: The Sydney Stay Hosting This Season of MAFS Australia

As Married At First Sight returns to Australian screens in 2026, viewers are once again getting a ...

Migraine is more than just a headache. A neurologist explains the 4 stages

A migraine attack[1] is not just a “bad headache”. Migraine is a debilitating neurological co...

Marketers: Forget the Black Box. If You Aren't Moving the Needle, What Are You Doing?

Two years ago, I entered the digital marketing space with the mindset of an engineering student ...

Extreme weather growing threat to Australian businesses in storm and fire season

  Australian small businesses are being hit harder than ever by costly disruptions...