Meta’s oversight board spotlights the social network’s problem with explicit AI deepfakes

Meta’s Oversight Board says the social media company fell short in its response to a pair of high-profile explicit, AI-generated images of female public figures on its sites and is calling for the company to update its policies and make the language of those policies even more clear to users.

The decision, announced Thursday, comes after a three-month investigation following user posts of a deepfake nude image of a public figure from India, as well as a more graphic image of a public figure from the U.S. Neither Meta nor the Oversight Board named the victims of the deepfakes.

The Board found that both images violated Meta’s rule prohibiting “derogatory sexualized photoshop” images, part of its Bullying and Harassment policy, even though Photoshop was not used in their creation.

“Removing both posts was in line with Meta’s human rights responsibilities,” the finding reads.

Unfortunately, that removal didn’t happen as fast as it should have. In the case of the Indian woman, a user reported the content to Meta, citing pornography. That report was closed as it was not reviewed within 48 hours. An appeal was made, but that was also automatically closed, meaning the image remained viewable. The user then appealed to the Oversight Board, at which point Meta reversed course and said the decision to leave the image up was a mistake and removed it.

However, the image of the American celebrity, which showed her nude while a man groped her breasts, was removed promptly and added to the company’s automated enforcement system, which instantly removes reposts of offensive images.

When the Oversight Committee asked about the discrepancy, Meta replied that the company relied on media reports to add the image resembling the American public figure to the bank, but there was no such news coverage in the Indian case.

“This is worrying because many victims of deepfake intimate images are not in the public eye and are forced to either accept the spread of their non-consensual depictions or search for and report every instance,” the Board’s ruling reads.

Clearer language

One of the chief takeaways from the Board’s recommendations is that the language Meta uses to prohibit these sort of images needs to be updated. It suggested the company move the prohibition on “derogatory sexualized photoshop” into the Adult Sexual Exploitation Community Standard.”

It also asked for some changes to the “derogatory sexualized photoshop” term, urging Meta to modify “derogatory” to “non-consensual” and to replace “Photoshop” with a more generalized term for manipulated media.

Additionally, it recommended that policies on non-consensual content include AI-generated or manipulated images and more. “For content with this specific context, the policy should also specify that it need not be ‘non-commercial or produced in a private setting’ to be violating,” it wrote.

The Board also noted concerns about the auto-closing of appeals for image-based sexual abuse, saying “even waiting 48 hours for a review can be harmful given the damage caused. The Board does not yet have sufficient information on Meta’s use of auto­ closing generally but considers this an issue that could have a significant human rights impact, requiring risk assessment and mitigation.”

Deepfake sexually-explicit or nude images and videos are overwhelmingly aimed at women, who made up 99% of the targeted individuals, according to a 2023 report from Home Security Heroes.

“Experts consulted by the Board noted that this content can be particularly damaging in socially conservative communities,” the Oversight Board wrote. “For instance, an 18-year-old woman was reportedly shot dead by her father and uncle in Pakistan’s remote Kohistan region after a digitally altered photograph of her with a man went viral.”

As offensive as the posting of deepfakes often is, those images are not always necessarily meant to harass the victim. External research, the Board wrote, shows that “users post such content for many reasons besides harassment and trolling, including a desire to build an audience, monetize pages or direct users to other sites, including pornographic ones.”

By clarifying the rules, it said, and putting the emphasis on the lack of consent, that could potentially reduce their proliferation of AI deepfakes on the company’s social media sites.

https://www.fastcompany.com/91162426/meta-oversight-board-ai-created-deepfakes?partner=rss&utm_source=rss&utm_medium=feed&utm_campaign=rss+fastcompany&utm_content=rss

Létrehozva 9mo | 2024. júl. 25. 12:20:09


Jelentkezéshez jelentkezzen be

EGYÉB POSTS Ebben a csoportban

Goodbye human drivers? Waymo’s robotaxis are now fully operational

Summoning a robotaxi from your phone is not a futuristic fantasy since Waymo achieved full commercial deployment.

https://www.fastcompany.com/91325288/goodbye-human-drivers-waymos-robotaxis-a

2025. máj. 6. 8:50:02 | Fast company - tech
‘You got to be really careful what you tie your name to’: The Hawk Tuah girl is planning a rebrand

Haliey Welch, better known as the Hawk Tuah girl, is ready for a rebrand.

After being thrust into the spotlight in 2024, thanks to her now-iconic “Hawk Tuah” catchphrase—featured in a vi

2025. máj. 5. 23:30:07 | Fast company - tech
Anthropic hires a top Biden official to lead its new AI-for-social-good team (exclusive)

Anthropic is turning to a Biden administration alum to run its new Beneficial Deployments team, which is tasked with helping extend the benefits of its AI to organizations focused on social good—p

2025. máj. 5. 21:20:03 | Fast company - tech
Speed-limiting devices could be coming for reckless U.S. drivers in these states

A teenager who admitted being “addicted to speed” behind the wheel had totaled two other cars in the year before he slammed into a minivan at 112 mph (180 kph) in a Seattle suburb,

2025. máj. 5. 16:40:03 | Fast company - tech
Nvidia chips could face new tracking rules under a bipartisan bill to stop chip smuggling to China

A U.S. lawmaker plans to introduce legislation in coming weeks to verify the location of

2025. máj. 5. 16:40:02 | Fast company - tech
Meta’s AI social feed is a privacy disaster waiting to happen

Since ChatGPT sparked the generative AI revolution in November 2022, interacting with AI has felt like using a digital confession booth—private, intimate, and shielded from public view (unless you

2025. máj. 5. 14:20:05 | Fast company - tech