What ChatGPT’s ‘sycophancy’ failure teaches publishers about AI and trust

When OpenAI pulled back its latest ChatGPT release—one that apparently turned the helpful chatbot into a total suck-up—the company took the welcome step of explaining exactly what happened in a pair of blog posts. The response was a notable move and really pulled back the curtain on how much of what these systems do is shaped by language choices most people never see. A tweak in phrasing, a shift in tone, and suddenly the model behaves differently.

For journalists, this shouldn’t be surprising. Many editorial meetings are spent agonizing over framing, tone, and headline language. But what is surprising—and maybe even a little disorienting—is that the same editorial sensitivity now needs to be applied not just to headlines and pull quotes, but to algorithms, prompts, and workflows that live in the guts of newsroom technology.

Before we connect the dots to newsroom AI, a quick recap: OpenAI’s latest update to GPT-4o involved an extensive process for testing the outputs, and it scored well on the factors the testers could measure: accuracy, safety, and helpfulness, among others. However, some evaluators doing more qualitative testing said the model felt “off,” but without more to go on, OpenAI released it anyway.

Within a day, it was clear the evaluators’ vibe-checks were onto something. Apparently the release had substantially increased “sycophancy,” or the model’s tendency to flatter and support the user, regardless of whether it was ultimately helpful. In its post announcing the rollback, OpenAI said it would refine ChatGPT’s system prompt—the invisible language that serves as kind of an “umbrella” instruction for every query and conversation with the public chatbot.

Lost in translation

The first thing that strikes you about this: We’re talking about changes to language, not code. In reaction to the recall, a former OpenAI employee posted on X about a conversation he had with a senior colleague at the company about how the change of a single word in the system prompt induced ChatGPT to behave in different ways. And the only way to know this was to make the change and try it out.

If you’re familiar with AI and prompting, this isn’t a shock. But on a fundamental level, it kind of is. I’m not saying the new release of GPT-4o was entirely about changing language in the system prompt, but the system prompt is a crucial element—altering it was the only temporary fix OpenAI could implement before engaging in the careful process of rolling back the release.

For anyone in communications or journalism, this should be somewhat reassuring. We’re in the business of words, after all. And words are no longer just the way we communicate about technology—they’re a crucial part of how these systems work.

An editorial and product hybrid

OpenAI’s ordeal has two important takeaways for how the media deals with AI: First, that editorial staff have a vital role to play in building the AI systems that govern their operations. (Outside frontier labs, tool building often amounts to prompt engineering paired with automations.) And second, transparency is the path to preserving user trust.

On the first point, the way AI directly affects content, and the need for good prompting to do that well, has a consequence for how media companies are organized: Editorial and product teams are becoming more like each other. The more journalists incorporate AI into their process, the more they end up creating their own tools. Think custom GPTs for writing assistance, NotebookLM knowledge bases for analyzing documents, or even browser extensions for fact-checking on the fly.

On the product side, the idea that media technology today isn’t just presenting content, but remixing and sometimes creating it is a massive change. To ensure those outputs adhere to journalistic principles, it doesn’t just make sense to have writers and editors be a part of that process—it’s necessary.

What results, then, is a journalist-product manager hybrid. These kinds of roles aren’t entirely new, but they’re generally senior leadership roles with words like “newsroom innovation” in the title. What AI does is encourage each side to adopt the skills of the other all the way down. Every reporter adopts a product mindset. Every product manager prioritizes brevity and accuracy.

Audience trust starts with transparency

The audience is the silent partner in this relationship, and OpenAI’s incident also serves as an example of how to best include them—through radical transparency. It’s hard to think of a way OpenAI could have better restored trust with its users other than its decision to fully explain how the problems got by its review process, and what it’s doing to improve.

While it’s unusual among the major AI labs (can you imagine xAI or DeepSeek writing a similar note?), this isn’t out of character for OpenAI. Sam Altman often shares on his X account announcements and behind-the-scenes observations from his vantage point as CEO, and while those are probably more calculated than they seem, they’ve earned the company a certain amount of respect.

This approach provides a road map for how to publicly communicate about AI strategy, especially for the media. Typically, when a publication creates an AI media policy, the focus is on disclosures and guidelines. Those are great first steps, but without a clearer window into the specific process, indicators such as “This article is AI assisted” aren’t that helpful, and audiences will be inclined to assume the worst when something goes wrong.

Better to be transparent from the start. When CNET used AI writers in the early days of generative AI to disastrous results, it published a long explanation of what went wrong, but it didn’t come until well after it had been called out. If the publication had been out front with what it was doing—not just saying it was using AI, but explaining how it was building, using, and evaluating it—things might have turned out differently.

Journalists can shape AI—and should

In its second post about the sycophancy fiasco, OpenAI revealed that a big part of its concern was the surprising number of people who now use ChatGPT for personal advice, an activity that wasn’t that significant a year ago. That growth is a testament to how fast the technology is improving and taking hold in various aspects of our lives. While it’s only just beginning to alter the media ecosystem, it could quickly become more deeply embedded than we had predicted.

Building AI systems that people trust starts with the people building them. By leveraging the natural talents of journalists on product teams, those systems will have the best chance of success. But when they screw up—and they will—preserving that trust will depend on how clear the window is on how they were built. Best to start polishing it now.

https://www.fastcompany.com/91331042/openai-chatgpt-sycophancy-failure-teaches-publishers-about-ai-trust?partner=rss&utm_source=rss&utm_medium=feed&utm_campaign=rss+fastcompany&utm_content=rss

Vytvořeno 2mo | 9. 5. 2025 8:50:02


Chcete-li přidat komentář, přihlaste se

Ostatní příspěvky v této skupině

A newly discovered exoplanet rekindles humanity’s oldest question: Are we alone?

Child psychologists tell us that around the age of five or six, children begin to seriously contemplate the world around them. It’s a glorious moment every parent recognizes—when young minds start

13. 7. 2025 11:10:06 | Fast company - tech
How Watch Duty became a go-to app during natural disasters

During January’s unprecedented wildfires in Los Angeles, Watch Duty—a digital platform providing real-time fire data—became the go-to app for tracking the unfolding disaster and is credit

13. 7. 2025 6:30:05 | Fast company - tech
Why the AI pin won’t be the next iPhone

One of the most frequent questions I’ve been getting from business execs lately is whether the

12. 7. 2025 12:10:02 | Fast company - tech
Microsoft will soon delete your Authenticator passwords. Here are 3 password manager alternatives

Users of Microsoft apps are having a rough year. First, in May, the Windows maker

12. 7. 2025 9:40:03 | Fast company - tech
Yahoo Creators platform hits record revenue as publisher bets big on influencer-led content

Yahoo’s bet on creator-led content appears to be paying off. Yahoo Creators, the media company’s publishing platform for creators, had its most lucrative month yet in June.

Launched in M

11. 7. 2025 17:30:04 | Fast company - tech
GameStop’s Nintendo Switch 2 stapler sells for more than $100,000 on eBay after viral mishap

From being the face of memestock mania to going viral for inadvertently stapling the screens of brand-new video game consoles, GameStop is no stranger to infamy.

Last month, during the m

11. 7. 2025 12:50:04 | Fast company - tech
Don’t take the race for ‘superintelligence’ too seriously

The technology industry has always adored its improbably audacious goals and their associated buzzwords. Meta CEO Mark Zuckerberg is among the most enamored. After all, the name “Meta” is the resi

11. 7. 2025 12:50:02 | Fast company - tech