Here’s What Meta’s AI Warning Labels Will Look Like on Images, Video and Audio Posts – CNET

author
3 minutes, 1 second Read

Get ready for more and more content on Facebook and Instagram to be labeled for being manipulated or generated by artificial intelligence. Meta, which also owns WhatsApp and Threads, said Friday it plans to expand AI warnings online. It follows a growing number of services, including YouTube and TikTok, that are responding to this issue.

Meta said that, starting in May, it will label video, audio and images as “Made with AI” either when its systems detect AI involvement, or when creators disclose it during an upload. The company also said it may add a more prominent label if the content has “a particularly high risk of materially deceiving the public on a matter of importance.”

“We plan to start labeling AI-generated content in May 2024, and we’ll stop removing content solely on the basis of our manipulated video policy in July,” Monika Bickert, Meta’s vice presidnet of content policy, wrote in a blog post. “This overall approach gives people more information about the content so they can better assess it and so they will have context if they see the same content elsewhere.”

Meta said it came to its decision to balance transparency with the need to avoid unnecessarily restricting freedom of expression online.

Read moreAI Atlas: Your Guide to Today’s Artificial Intelligence

The move marks another way the tech industry is responding to growing concerns about the pervasiveness of AI-generated content and its risk to the public. Videos generated by AI technology like OpenAI’s Sora look increasingly lifelike. And though that tool hasn’t been made widely available to the public, other AI technologies have already begun to cause public confusion. 

Earlier this year, a political consultant made mass-scale robocalls using President Joe Biden’s voice, re-created by AI, encouraging people in New Hampshire not to vote in the primary election. Experts say more AI disinformation is likely on the way, particularly with the upcoming 2024 presidential election.

Meta isn’t the only social media company working to identify AI-powered content. TikTok said last year that it will launch a tool to help creators label manipulated content, noting that it also prohibits deepfakes — videos, images or audio created to mislead viewers about real events or people. Meanwhile, Google’s YouTube subsidiary began requiring disclosure of AI-manipulated videos from creators last month, saying that some examples included “realistic” likenesses of people or scenes, as well as altered footage of real events or places. 

What Facebook and Instagram AI warning labels will look like

Meta showed off what AI-made Instagram labels will look like.

Meta

Meta showed a mockup of what it’ll look like in May when more content begins being labeled — Instagram’s AI warning labels, for instance, will say “Made with AI” right under the creator’s username, with a little square icon with a plus in the corner.

Meta cited a survey it conducted with more than 23,000 respondents in 13 countries, in which 82% favored labels on AI-generated content “that depicts people saying things they did not say.” 

For more serious content, Meta’s blog post said, the company “will remove content, regardless of whether it is created by AI or a person, if it violates our policies against voter interference, bullying and harassment, violence and incitement, or any other policy in our Community Standards.”

Read moreHow Close Is That Photo to the Truth? What to Know in the Age of AI

Editors’ note: CNET used an AI engine to help create several dozen stories, which are labeled accordingly. The note you’re reading is attached to articles that deal substantively with the topic of AI but are created entirely by our expert editors and writers. For more, see our AI policy.

This post was originally published on this site

Similar Posts