Anthropic's Claude AI now has the ability to end 'distressing' conversations

Anthropic's latest feature for two of its Claude AI models could be the beginning of the end for the AI jailbreaking community. The company announced in a post on its website that the Claude Opus 4 and 4.1 models now have the power to end a conversation with users. According to Anthropic, this feature will only be used in "rare, extreme cases of persistently harmful or abusive user interactions."

To clarify, Anthropic said those two Claude models could exit harmful conversations, like "requests from users for sexual content involving minors and attempts to solicit information that would enable large-scale violence or acts of terror." With Claude Opus 4 and 4.1, these models will only end a conversation "as a last resort when multiple attempts at redirection have failed and hope of a productive interaction has been exhausted," according to Anthropic. However, Anthropic claims most users won't experience Claude cutting a conversation short, even when talking about highly controversial topics, since this feature will be reserved for "extreme edge cases."

Anthropic's example of Claude ending a conversation
Anthropic

In the scenarios where Claude ends a chat, users can no longer send any new messages in that conversation, but can start a new one immediately. Anthropic added that if a conversation is ended, it won't affect other chats and users can even go back and edit or retry previous messages to steer towards a different conversational route.

For Anthropic, this move is part of its research program that studies the idea of AI welfare. While the idea of anthropomorphizing AI models remains an ongoing debate, the company said the ability to exit a "potentially distressing interaction" was a low-cost way to manage risks for AI welfare. Anthropic is still experimenting with this feature and encourages its users to provide feedback when they encounter such a scenario.

This article originally appeared on Engadget at https://www.engadget.com/ai/anthropics-claude-ai-now-has-the-ability-to-end-distressing-conversations-201427401.html?src=rss https://www.engadget.com/ai/anthropics-claude-ai-now-has-the-ability-to-end-distressing-conversations-201427401.html?src=rss
Creado 4h | 17 ago 2025, 22:20:12


Inicia sesión para agregar comentarios

Otros mensajes en este grupo.

MasterClass deal: Subscriptions are 40 percent off right now

If you want to brush up on some skills or learn new ones,

17 ago 2025, 13:10:07 | Engadget
Roblox cracks down on its user-created content following multiple child safety lawsuits

Following a wave of lawsuits alleging that Roblox doesn't provide a safe environment for its underage users, the gaming platform made a series of sweeping updates to its policies. To address recent

16 ago 2025, 20:50:05 | Engadget
Apple's iPad Air M3 is cheaper than ever right now

For a lot of people, the iPad Air is Apple’s goldilocks tablet. It’s more powerful and available in a larger size than the entry-level iPad, but a lot less expensive than the iPad Pro, which can be

16 ago 2025, 16:20:10 | Engadget
One of our favorite Bluetooth speakers drops to $60 on Amazon

Yes, summer might be coming to a close sooner than any of us would like, but that doesn't mean the outdoor fun has to end.

Currently,

16 ago 2025, 16:20:07 | Engadget