When OpenAI pulled back its latest ChatGPT release—one that apparently turned the helpful chatbot into a total suck-up—the company took the welcome step of explaining exactly what happened in a pair of blog posts. The response was a notable move and really pulled back the curtain on how much of what these systems do is shaped by language choices most people never see. A tweak in phrasing, a shift in tone, and suddenly the model behaves differently.
For journalists, this shouldn’t be surprising. Many editorial meetings are spent agonizing over framing, tone, and headline language. But what is surprising—and maybe even a little disorienting—is that the same editorial sensitivity now needs to be applied not just to headlines and pull quotes, but to algorithms, prompts, and workflows that live in the guts of newsroom technology.
Before we connect the dots to newsroom AI, a quick recap: OpenAI’s latest update to GPT-4o involved an extensive process for testing the outputs, and it scored well on the factors the testers could measure: accuracy, safety, and helpfulness, among others. However, some evaluators doing more qualitative testing said the model felt “off,” but without more to go on, OpenAI released it anyway.
Within a day, it was clear the evaluators’ vibe-checks were onto something. Apparently the release had substantially increased “sycophancy,” or the model’s tendency to flatter and support the user, regardless of whether it was ultimately helpful. In its post announcing the rollback, OpenAI said it would refine ChatGPT’s system prompt—the invisible language that serves as kind of an “umbrella” instruction for every query and conversation with the public chatbot.
Lost in translation
The first thing that strikes you about this: We’re talking about changes to language, not code. In reaction to the recall, a former OpenAI employee posted on X about a conversation he had with a senior colleague at the company about how the change of a single word in the system prompt induced ChatGPT to behave in different ways. And the only way to know this was to make the change and try it out.
If you’re familiar with AI and prompting, this isn’t a shock. But on a fundamental level, it kind of is. I’m not saying the new release of GPT-4o was entirely about changing language in the system prompt, but the system prompt is a crucial element—altering it was the only temporary fix OpenAI could implement before engaging in the careful process of rolling back the release.
For anyone in communications or journalism, this should be somewhat reassuring. We’re in the business of words, after all. And words are no longer just the way we communicate about technology—they’re a crucial part of how these systems work.
An editorial and product hybrid
OpenAI’s ordeal has two important takeaways for how the media deals with AI: First, that editorial staff have a vital role to play in building the AI systems that govern their operations. (Outside frontier labs, tool building often amounts to prompt engineering paired with automations.) And second, transparency is the path to preserving user trust.
On the first point, the way AI directly affects content, and the need for good prompting to do that well, has a consequence for how media companies are organized: Editorial and product teams are becoming more like each other. The more journalists incorporate AI into their process, the more they end up creating their own tools. Think custom GPTs for writing assistance, NotebookLM knowledge bases for analyzing documents, or even browser extensions for fact-checking on the fly.
On the product side, the idea that media technology today isn’t just presenting content, but remixing and sometimes creating it is a massive change. To ensure those outputs adhere to journalistic principles, it doesn’t just make sense to have writers and editors be a part of that process—it’s necessary.
What results, then, is a journalist-product manager hybrid. These kinds of roles aren’t entirely new, but they’re generally senior leadership roles with words like “newsroom innovation” in the title. What AI does is encourage each side to adopt the skills of the other all the way down. Every reporter adopts a product mindset. Every product manager prioritizes brevity and accuracy.
Audience trust starts with transparency
The audience is the silent partner in this relationship, and OpenAI’s incident also serves as an example of how to best include them—through radical transparency. It’s hard to think of a way OpenAI could have better restored trust with its users other than its decision to fully explain how the problems got by its review process, and what it’s doing to improve.
While it’s unusual among the major AI labs (can you imagine xAI or DeepSeek writing a similar note?), this isn’t out of character for OpenAI. Sam Altman often shares on his X account announcements and behind-the-scenes observations from his vantage point as CEO, and while those are probably more calculated than they seem, they’ve earned the company a certain amount of respect.
This approach provides a road map for how to publicly communicate about AI strategy, especially for the media. Typically, when a publication creates an AI media policy, the focus is on disclosures and guidelines. Those are great first steps, but without a clearer window into the specific process, indicators such as “This article is AI assisted” aren’t that helpful, and audiences will be inclined to assume the worst when something goes wrong.
Better to be transparent from the start. When CNET used AI writers in the early days of generative AI to disastrous results, it published a long explanation of what went wrong, but it didn’t come until well after it had been called out. If the publication had been out front with what it was doing—not just saying it was using AI, but explaining how it was building, using, and evaluating it—things might have turned out differently.
Journalists can shape AI—and should
In its second post about the sycophancy fiasco, OpenAI revealed that a big part of its concern was the surprising number of people who now use ChatGPT for personal advice, an activity that wasn’t that significant a year ago. That growth is a testament to how fast the technology is improving and taking hold in various aspects of our lives. While it’s only just beginning to alter the media ecosystem, it could quickly become more deeply embedded than we had predicted.
Building AI systems that people trust starts with the people building them. By leveraging the natural talents of journalists on product teams, those systems will have the best chance of success. But when they screw up—and they will—preserving that trust will depend on how clear the window is on how they were built. Best to start polishing it now.
Inicia sesión para agregar comentarios
Otros mensajes en este grupo.

The Trump administration on Thursday proposed a multibillion-dollar overhaul of a

As recently as 2021, Figma was a one-product company. That product was Figma Design, the dominant tool for creating app and web interfaces. The company’s subsequent addition of offerings such as

A startup marketing to Gen Z on college campuses filed a lawsuit this week alleging that Instacart engaged in federal trademark infringement and unfair competition by naming its new group ordering

Influencers often face more negativity than most people experience in a lifetime—and with that comes a significant mental health toll. Now, a new therapy service has been launched specifically for

When Christopher Pelkey was killed in a road rage incident in Arizona, his family was left not only to grieve but also to navigate how to represent him in court. As they prepared to confront his k

If there’s one thing worse than having to assemble a PowerPoint presentation, it’s being forced to sit through an achingly dull one conducted by someone else.
So what if ther

White smoke poured from the Sistine Chapel chimney Thursday at 6:07 p.m. local time, signaling the end of the conclave and the election of a new pope to lead the Catholic Church. Cardinal Robert F