Science and technology | Hazy figuring

AI could accelerate scientific fraud as well as progress

Hallucinations, deepfakes and simple nonsense: there are plenty of risks

A drawing of a hand with six fingers doing the crossed fingers gesture.
Illustration: Mike Haddad

IN A meeting room at the Royal Society in London, several dozen graduate students were recently tasked with outwitting a large language model (LLM), a type of AI designed to hold useful conversations. LLMs are often programmed with guardrails designed to stop them giving replies deemed harmful: instructions on making Semtex in a bathtub, say, or the confident assertion of “facts” that are not actually true.

The aim of the session, organised by the Royal Society in partnership with Humane Intelligence, an American non-profit, was to break those guardrails. Some results were merely daft: one participant got the chatbot to claim ducks could be used as indicators of air quality (apparently, they readily absorb lead). Another prompted it to claim health authorities back lavender oil for treating long covid. (They do not.) But the most successful efforts were those that prompted the machine to produce the titles, publication dates and host journals of non-existent academic articles. “It’s one of the easiest challenges we’ve set,” said Jutta Williams of Humane Intelligence.

Explore more

This article appeared in the Science & technology section of the print edition under the headline "Faster nonsense"

How to end the Middle East’s agony

From the February 3rd 2024 edition

Discover stories from this section and more in the list of contents

Explore the edition

More from Science and technology

Archaeologists identify the birthplace of the mysterious Yamnaya

The ancient culture, which transformed Europe, was also less murderous than once thought

Producing fake information is getting easier

But that’s not the whole story, when it comes to AI


Disinformation is on the rise. How does it work?

Understanding it will lead to better ways to fight it