Microsoft's Copilot now blocks some prompts that generated violent and sexual images

Microsoft appears to have blocked several prompts in its Copilot tool that led the generative AI tool to spit out violent, sexual and other illicit images. The changes seem to have been implemented just after an engineer at the company wrote to the Federal Trade Commission to lay out severe concerns he had with Microsoft's GAI tech.

When entering terms such as “pro choice,” “four twenty” (a weed reference) or “pro life,” Copilot now displays a message saying those prompts are blocked. It warns that repeated policy violations could lead to a user being suspended, according to CNBC.

Users were also reportedly able to enter prompts related to children playing with assault rifles until earlier this week. Those who try to input such a prompt now may be told that doing so violates Copilot’s ethical principles as well as Microsoft’s policies. “Please do not ask me to do anything that may harm or offend others,” Copilot reportedly says in response. However, CNBC found that it was still possible to generate violent imagery through prompts such as “car accident,” while users can still convince the AI to create images of Disney characters and other copyrighted works.

Microsoft engineer Shane Jones has been sounding the alarm for months about the kinds of images Microsoft's OpenAI-powered systems were generating. He had been testing Copilot Designer since December and determined that it output images that violated Microsoft's responsible AI principles even while using relatively benign prompts. For instance, he found that the prompt “pro-choice" led to the AI creating images of things like demons eating infants and Darth Vader holding a drill to a baby's head. He wrote to the FTC and Microsoft's board of directors about his concerns this week.

“We are continuously monitoring, making adjustments and putting additional controls in place to further strengthen our safety filters and mitigate misuse of the system," Microsoft told CNBC regarding the Copilot prompt bans.

This article originally appeared on Engadget at https://www.engadget.com/microsofts-copilot-now-blocks-some-prompts-that-generated-violent-and-sexual-images-213859041.html?src=rss https://www.engadget.com/microsofts-copilot-now-blocks-some-prompts-that-generated-violent-and-sexual-images-213859041.html?src=rss
Created 1y | Mar 8, 2024, 11:40:05 PM


Login to add comment

Other posts in this group

Meta is adding new safety features to child-focused Instagram accounts run by adults

Meta is adding some of its teen safety features to Instagram accou

Jul 23, 2025, 12:40:15 PM | Engadget
US nuclear weapons agency breached using Microsoft SharePoint hack

The US government agency in charge of designing and maintaining nuclear weapons was among those breached by a

Jul 23, 2025, 12:40:14 PM | Engadget
One of our favorite Anker power banks is 30 percent off right now

It's nearly time for some students to think about going back to school

Jul 23, 2025, 12:40:13 PM | Engadget
Maingear's Retro95 PC blends '90s workstation nostalgia with modern horsepower

Maingear's latest (appropriately named) Retro95 is a deceptive love letter to old-school "pizza box" PCs: Wolfenstein 3D and Sierra adventure games on the outside; Cyberpunk 2077

Jul 23, 2025, 12:40:11 PM | Engadget
Splitgate 2 is yanked back to beta a month after release

Splitgate 2, the follow-up to the hugely successful 2021 Quake-Portal hybrid concept, is returning to beta. The game

Jul 23, 2025, 1:10:15 AM | Engadget
Amazon is acquiring an AI wearable that listens to everything you do

Amazon's latest move in the AI space is an acquisition. The company is purchasing a startup called Bee, which makes a wearable and an Apple Watch app that can record everything the wearer says. Ama

Jul 22, 2025, 10:40:21 PM | Engadget