Why experts are using the word ‘bullshit’ to describe AI’s flaws

Since the launch of ChatGPT in late 2022 we’ve known that artificial intelligence language models are prone to spewing falsehoods, otherwise known as hallucinations

The AI companies have been telling us that it’s a problem that can be cured. And with the use of technologies like retrieval-augmented generation (in which the AI calls on a database of reliable information), hallucinations have indeed decreased in many contexts. But they persist, as a recent Wired investigation of the Perplexity AI search tool illustrated. The Wired story features an especially bold headline: “Perplexity Is a Bullshit Machine.”

The use of “bullshit” made more than a grabby headline; it was a reference to recently published research from a trio of philosophy professors at Glasgow University. The report, titled “ChatGPT is bullshit,” argues that calling the false output of large language models hallucination is misleading; what LLMs are really spouting, they argue, is more like bullshit. And not just any bullshit: Bullshit as defined by the late moral philosopher Harry Frankfurt in his 2005 bestseller, On Bullshit. 

Frankfurt’s book is principally concerned with defining the difference between a “liar” and a “bullshitter.” A bullshitter (aka “bullshit artist”) doesn’t use facts to come off as credible and therefore persuasive, he explained, but is content to say things that sound true to get the same result. For example, a used-car salesman who is practicing bullshit uses a set of talking points he thinks will lead someone to buy a car. Some of the talking points may be true, some may be false; to him it doesn’t matter: He would use the same set of points whether they happened to be true or not. “[He] does not reject the authority of the truth, as the liar does, and oppose himself to it,” Frankfurt wrote. “He pays no attention to it at all. By virtue of this, bullshit is a greater enemy of the truth than lies are.” 

In their recent report, the Glasgow researchers—Michael Townsen Hicks, James Humphries, and Joe Slater—argue that Frankfurt’s bullshit definition fits the behavior of LLMs better than the term hallucinate. In order to hallucinate, the researchers argue, one must have some awareness or regard for the truth; LLMs, by contrast, work with probabilities, not binary correct/incorrect judgments. Based on a huge many-dimensional map of words created by processing huge amounts of text, LLMs decide which words (based on meaning and current context) would most likely follow from the words used in a prompt. They’re inherently more concerned with sounding truthy than delivering a factually correct response, the researchers conclude.

Importantly, the LLM doesn’t always choose the word that is statistically most likely to follow. Letting the model choose between a set of more or less likely candidates for the next word gives the output an unexpected, creative, or even human quality. This quality can even be modulated using a control that AI developers call “temperature.” But dialing up the model’s temperature increases the chances it will generate falsehoods.

When LLMs “hallucinate,” it’s normally because they’ve encountered a dearth of information about the particular subject or context within their training data. They rely on their existing probability settings and, in a sense, wing it. Frankfurt’s “bullshit” definition seems to capture this behavior remarkably well. “Bullshit is unavoidable whenever circumstances require someone to talk without knowing what he is talking about,” he wrote.

And these metaphors matter, the researchers say. “Calling their mistakes ‘hallucinations’ isn’t harmless: It lends itself to the confusion that the machines are in some way misperceiving but are nonetheless trying to convey something that they believe or have perceived,” they write, concluding that the term also “feeds into overblown hype about their abilities among technology cheerleaders.”

Of course, Frankfurt was talking about human bullshitters, not AIs. But human metaphors are the best tools we have for understanding AI. Because of the opaque and/or alien quality of the inner workings of LLMs, the public has relied on AI researchers and Big Tech companies to provide those metaphors. That might have been fine before generative AI became a very big business—and before there was a profit motive to spin and control narratives. With profit-hungry companies like Google and OpenAI choosing their words carefully, AI regulators and consumers alike should insist on hearing a diversity of opinions, especially when it involves crucial issues such as bias and safety. 

<hr class=“wp-block-separator is-style-wide”/> https://www.fastcompany.com/91145865/why-experts-are-using-the-word-bullshit-to-describe-ais-flaws?partner=rss&amp;utm_source=rss&amp;utm_medium=feed&amp;utm_campaign=rss+fastcompany&amp;utm_content=rss

Creată 11mo | 25 iun. 2024, 13:20:03


Autentifică-te pentru a adăuga comentarii

Alte posturi din acest grup

Shimmers, floating toolbars, and radical transparency: Here’s what iOS 26 could look like

In less than two weeks, on June 9, Apple will kick off its annual Worldwide Developers Conference (WWDC), in which it will showcase the next versions of the operating systems that power its myriad

31 mai 2025, 09:20:03 | Fast company - tech
‘I still think I’m dreaming’: Kai Cenat’s ‘Streamer University’ now has its first graduating class

The class of 2025 have now graduated from Kai Cenat’s “Streamer University.”

Last week, 120 students—handpicked from

30 mai 2025, 21:40:04 | Fast company - tech
This startup is bringing photos—and even video—to 911 calls

It’s become commonplace to message someone a photo, text them an address, and switch to a video chat all in the middle of a phone call.

But 911 systems, largely designed for the er

30 mai 2025, 17:10:05 | Fast company - tech
Bluesky is most definitely alive and kicking

Last weekend, an ugly rumor of a tragic death spread began rocketing around Bluesky. What made it odd was the identity of the dearly departed: Bluesky itself.

It’s not entirely clear wha

30 mai 2025, 14:40:05 | Fast company - tech
Tesla’s best-selling Model Y could be dethroned by this newly launched Chinese EV model

Xiaomi rolled out its new sports utility vehicle in Beijing on Thursday, as the firm best kno

30 mai 2025, 14:40:05 | Fast company - tech
For CEOs, AI tech literacy is no longer optional

Artificial intelligence has been the subject of unprecedented levels of investment and enthusiasm over the past three years, driven by a tide of hype that promises revolutionary transformation acr

30 mai 2025, 10:10:04 | Fast company - tech