AI Chatbots have telltale quirks. Researchers can spot them with 97% accuracy

There’s a cat-and-mouse game between those using generative AI chatbots to produce text undetected and those trying to catch them. Many believe they know the telltale signs—though as a journalist fond of the word “delve” and prone to em-dashes, I’m not so sure.

Researchers at four U.S. universities, however, have taken a more rigorous approach, identifying linguistic fingerprints that reveal which large language model (LLM) produced a given text.

“All these chatbots are coming out every day, and we interact with them, but we don’t really understand the differences between them,” says Mingjie Sun, a researcher at Carnegie Mellon University and lead author of the study, which was published in Cornell University’s preprint server arXiv. “By training a machine learning classifier to do this task, and by looking at the performance of that classifier, we can then assess the difference between different LLMs.”

Sun and his colleagues developed a machine learning model that analyzed the outputs of five popular LLMs, and was able to distinguish between them with 97.1% accuracy. Their machine learning model uncovered distinct verbal quirks unique to each LLM.

ChatGPT’s GPT-4o model, for instance, tends to use “utilize” more than other models. DeepSeek is partial to saying “certainly.” Google’s Gemini often prefaces its conclusions with the word “essentially,” while Anthropic’s Claude overuses phrases like “according to” and “according to the text” when citing its sources.

xAI’s Grok stands out as more discursive and didactic, often reminding users to “remember” key points while guiding them through arguments with “not only” and “but also.”

“The writing, the word choices, the formatting are all different,” says Yida Yin, a researcher at the University of California, Berkeley, and a coauthor of the paper.

These insights can help users select the best model for specific writing tasks—or aid those trying to catch AI-generated text masquerading as human work. So, remember: according to this study, if a model utilizes certain words, it’s certainly possible to identify it.

https://www.fastcompany.com/91286162/ai-chatbots-have-telltale-quirks-researchers-can-spot-them-with-97-accuracy?partner=rss&utm_source=rss&utm_medium=feed&utm_campaign=rss+fastcompany&utm_content=rss

Created 6mo | Mar 3, 2025, 2:30:07 PM


Login to add comment

Other posts in this group

Philips CEO Jeff DiLullo on how AI is changing healthcare today

AI is quietly reshaping the efficiency, power, and potential of U.S. h

Aug 18, 2025, 9:10:07 PM | Fast company - tech
How satellites and orbiting weapons make space the new battlefield

As Russia held its Victory Day parade this year, hackers backing the Kremlin hijacked an orbiting satel

Aug 18, 2025, 9:10:06 PM | Fast company - tech
Meta spent $27 million protecting Mark Zuckerberg last year, more than any other CEO

The targeted murder of United Healthcare CEO Brian Thompson last December put the business w

Aug 18, 2025, 9:10:05 PM | Fast company - tech
Tesla lowers monthly lease fee due to UK sales slump

British motorists can now lease a Tesla

Aug 18, 2025, 9:10:05 PM | Fast company - tech
Google fined $36 million for anticompetitive deals with Australia’s largest telcos

Google has agreed to pay a 55 million Australian dollar ($36 million) fine for signing anticompetitive deals with Australia’s two largest telecommun

Aug 18, 2025, 6:50:02 PM | Fast company - tech
‘Pips,’ a new logic puzzle from New York Times Games, might just be your next ‘Wordle’

On an average day, tens of millions of people visit The New York Times Games section to solve the latest crossword puzzle, keep their

Aug 18, 2025, 4:30:05 PM | Fast company - tech
Crowdfunded companies are ‘ghosting’ investors. Changing the rules could restore trust

Imagine you invest $500 to help a startup get off the ground through investment crowdfunding. The pitch is slick, the platform feels

Aug 18, 2025, 9:30:05 AM | Fast company - tech