Meta oversight board reviews handling of AI-created celebrity porn

industry executives have called for legislation to criminalize the creation of harmful "deep fakes"


Reuters April 16, 2024
A combination photograph showing an image purporting to be of British student and freelance writer Oliver Taylor (L) and a heat map of the same photograph produced by Tel Aviv-based deepfake detection company Cyabra is seen in this undated handout photo obtained by Reuters. The heat map, which was produced using one of Cyabra's algorithms, highlights areas of suspected computer manipulation. The digital inconsistencies were one of several indicators used by experts to determine that Taylor was an online mirage. PHOTO: REUTERS

NEWYORK:

Meta Platforms' (META.O), opens new tab Oversight Board is reviewing the company's handling of two sexually explicit AI-generated images of female celebrities that circulated on its Facebook and Instagram services, the board said on Tuesday.

The board, which is funded by the social media giant but operates independently from it, will use the two examples to assess the overall effectiveness of Meta's policies and enforcement practices around pornographic fakes created using artificial intelligence, it said in a blog post.

It provided descriptions of the images in question but did not name the famous women depicted in them in order to "prevent further harm," a board spokesperson said.

Advances in AI technology have made fabricated images, audio clips and videos virtually indistinguishable from real human-created content, resulting in a spate of sexual fakes proliferating online, mostly depicting women and girls.

Read: Research firm warns of threat from deepfakes

In an especially high-profile case earlier this year, Elon Musk-owned social media platform X briefly blocked users from searching for all images of U.S. pop star Taylor Swift after struggling to control the spread of fake explicit images of her.

Some industry executives have called for legislation to criminalize the creation of harmful "deep fakes" and require that tech companies prevent such uses of their products.

According to the Oversight Board's descriptions of its cases, one involves an AI-generated image of a nude woman resembling a public figure from India, posted by an account on Instagram that only shares AI-generated images of Indian women.

The other image, the board said, appeared in a Facebook group for sharing AI creations and featured an AI-generated depiction of a nude woman resembling "an American public figure" with a man groping her breast.

Read: Getty Images launches an AI image generator to iStock

Meta removed the image depicting the American woman for violating its bullying and harassment policy, which bars "derogatory sexualized photoshops or drawings," but initially left up the one featuring the Indian woman and only reversed course after the board selected it for review.

In a separate post, Meta acknowledged the cases and pledged to implement the board's decisions.

COMMENTS (1)

Falisha Manpower | 7 months ago | Reply The Meta oversight board conducts thorough reviews of Meta s content moderation policies and practices. They assess Meta s handling of controversial content including decisions on user bans content removals and misinformation. Through independent scrutiny and recommendations the oversight board aims to promote transparency fairness and accountability in Meta s content moderation efforts. Apply Now For Best Jobs.best recruitment agencies in pakistan
Replying to X

Comments are moderated and generally will be posted if they are on-topic and not abusive.

For more information, please see our Comments FAQ