AI Chatbots have telltale quirks. Researchers can spot them with 97% accuracy

There’s a cat-and-mouse game between those using generative AI chatbots to produce text undetected and those trying to catch them. Many believe they know the telltale signs—though as a journalist fond of the word “delve” and prone to em-dashes, I’m not so sure.

Researchers at four U.S. universities, however, have taken a more rigorous approach, identifying linguistic fingerprints that reveal which large language model (LLM) produced a given text.

“All these chatbots are coming out every day, and we interact with them, but we don’t really understand the differences between them,” says Mingjie Sun, a researcher at Carnegie Mellon University and lead author of the study, which was published in Cornell University’s preprint server arXiv. “By training a machine learning classifier to do this task, and by looking at the performance of that classifier, we can then assess the difference between different LLMs.”

Sun and his colleagues developed a machine learning model that analyzed the outputs of five popular LLMs, and was able to distinguish between them with 97.1% accuracy. Their machine learning model uncovered distinct verbal quirks unique to each LLM.

ChatGPT’s GPT-4o model, for instance, tends to use “utilize” more than other models. DeepSeek is partial to saying “certainly.” Google’s Gemini often prefaces its conclusions with the word “essentially,” while Anthropic’s Claude overuses phrases like “according to” and “according to the text” when citing its sources.

xAI’s Grok stands out as more discursive and didactic, often reminding users to “remember” key points while guiding them through arguments with “not only” and “but also.”

“The writing, the word choices, the formatting are all different,” says Yida Yin, a researcher at the University of California, Berkeley, and a coauthor of the paper.

These insights can help users select the best model for specific writing tasks—or aid those trying to catch AI-generated text masquerading as human work. So, remember: according to this study, if a model utilizes certain words, it’s certainly possible to identify it.

https://www.fastcompany.com/91286162/ai-chatbots-have-telltale-quirks-researchers-can-spot-them-with-97-accuracy?partner=rss&utm_source=rss&utm_medium=feed&utm_campaign=rss+fastcompany&utm_content=rss

Vytvořeno 3mo | 3. 3. 2025 14:30:07


Chcete-li přidat komentář, přihlaste se

Ostatní příspěvky v této skupině

Vibe coding lets anyone write software—but comes with risks

Whether you’re streaming a show, paying bills online or sending an email, each of these actions relies on computer programs that run behind the scenes. The process of writing computer programs is

8. 6. 2025 9:40:04 | Fast company - tech
This free AI supersite is like Gemini Deep Research on steroids

Everywhere you look these days, there it is—some manner of breathlessly hyped new “AI” service that’s, like, totally gonna change your life forever. (Like, totally. For realsies.)

7. 6. 2025 12:50:02 | Fast company - tech
WWDC25: Here’s everything Apple is likely to announce

Apple’s annual Worldwide Developers Conference begins this Monday, June 9. Although the five-day event has historically been aimed at developers, Apple’s consumer fans generally can’t wait to tune

7. 6. 2025 10:30:05 | Fast company - tech
Why vibecoding your own apps is so amazing—and exasperating

“The truth is, I cannot explain exactly where your 1,216 image files went or when they disappeared. I apologize for not being more careful about investigating the root cause before taking any acti

6. 6. 2025 13:40:07 | Fast company - tech
Waymo is winning in San Francisco

The self-driving car service Waymo has been active in San Francisco for 20 months and has already captured 27% of the city’s rideshare market, according to

6. 6. 2025 13:40:05 | Fast company - tech