voices
Photo credit: Gerd Altmann/PDP

Large language models, including ChatGPT-4 and Claude Sonnet 3.5, cannot convincingly imitate the way people speak, with exaggerated imitation and incorrect use of filler words exposing them as non-human.

Research published in Cognitive Science tested four large language models by comparing transcripts of human phone conversations with those of simulated conversations. The study then checked whether people could distinguish between human phone conversations and those generated by language models.

“Large language models speak differently than people do,” said Lucas Bietti, associate professor from the Department of Psychology at the Norwegian University of Science and Technology.

The researchers tested ChatGPT-4, Claude Sonnet 3.5, Vicuna and Wayfarer. For the most part, people were not fooled by the language models.

The study identified three main problems with how language models imitate human speech. First, large language models demonstrate exaggerated alignment by imitating conversation partners too eagerly. While people slightly adapt their words and conversation according to the other person, this imitation is usually subtle.

“Large language models are a bit too eager to imitate, and this exaggerated imitation is something that humans can pick up on,” said Bietti.

Second, the models use discourse markers incorrectly. These small words including “so”, “well”, “like” and “anyway” have social functions that signal interest, belonging, attitude or meaning, and can structure conversation.

“The large language models use these small words differently, and often incorrectly,” said Bietti.

Third, language models struggle with opening and closing features of conversation. People typically engage in small talk before moving to main topics, with the shift from introduction to business taking place automatically without being explicitly stated. Similarly, people usually end conversations with phrases like “alright, then” or “talk to you later” rather than abruptly stopping once information has been conveyed.

“This introduction, and the shift to a new phase of the conversation, are also difficult for large language models to imitate,” said Bietti.

The researchers concluded that improvements in large language models will likely narrow the gap between human conversations and artificial ones, but key differences will probably remain.

The research was led by Eric Mayor from the University of Basel, with Adrian Bangerter from the University of Neuchâtel as final author.

Leave a Reply

Your email address will not be published. Required fields are marked *

You May Also Like

Super-intelligent AI could ‘play dumb’ to trick evaluators and evade controls

The dream of an AI-integrated society could turn into a nightmare if…

Satellite dataset uses deep learning to map 9.2 million kilometres of roads

Researchers have combined deep-learning models with high-resolution satellite imagery to classify 9.2…

Universities quietly deploying GenAI to ‘game’ £2bn research funding system

UK universities are widely using generative AI to prepare submissions for the…

AI guardrails defeated by poetry as ‘smarter’ models prove most gullible

The world’s most advanced artificial intelligence systems are being easily manipulated into…

Researchers hijack X feed with ad blocker tech to cool political tempers

Scientists have successfully intercepted and reshaped live social media feeds using ad-blocker-style…

Doing good buys forgiveness as CSR becomes ‘insurance’ against layoffs

Companies planning to slash jobs or freeze pay should start saving the…