SnogBlogy

Instagram Developing Nudity Protection Feature to Block Unsolicited Nude Photos in DMs

Instagram is working on a way to let users block unsolicited nude photos that they’re sent via direct messages — using a “nudity protection” filter to screen out and hide such pics.

The social media giant confirmed the development of the feature but did not provide a timeline for when it would become available. “We’re working closely with experts to ensure these new features preserve people’s privacy, while giving them control over the messages they receive,” a spokesperson for Meta, Instagram’s parent company, said in a statement.

According to Instagram, the “nudity protection” feature will be optional. It’s similar to Hidden Words, an opt-in feature Instagram launched last year that automatically filters out abusive DM requests based on keywords and diverts them into a hidden folder so users never have to see them.

Related Stories

Illustration of a TV set with a desert scene with a tumbleweed blowing across the screen VIP+

Fall Season’s Scripted Reduction Bodes Badly for Broadcast TV

The Late Show with Stephen Colbert and guest Vice President Kamala Harris during Wednesday’s March 15, 2023 show. Photo: Scott Kowalchyk/CBS ©2023 CBS Broadcasting Inc. All Rights Reserved.

Kamala Harris To Make First Late Night Appearance as Presidential Nominee This Tuesday on 'The Late Show With Stephen Colbert'

The company’s confirmation of the feature, as first reported by the Verge, came after app researcher Alessandro Paluzzi posted a screenshot of the forthcoming message-protection features earlier this week on Twitter. The Instagram notice says that “Technology on your device covers photos that may contain nudity in chats. Instagram can’t access the photos.” If the app detects nudity in a DM, Instagram will hide the photos unless a user chooses to view them, according to the screenshot.

Popular on Variety

A study released in April by nonprofit group Center for Countering Digital Hate analyzed direct messages sent to five high-profile women — including actor Amber Heard, who has been the target of widespread hateful attacks online amid her legal battle with Johnny Depp — and recorded 125 separate examples of image-based sexual abuse. Instagram “failed to act on every single example of image-based sexual abuse 48 hours after the abuse was reported to the platform,” according to the org.

According to a 2020 study from the Kinsey Institute at Indiana University, nearly 50% of women surveyed said they had ever received an image of male genitalia (aka a “dick pic”).

https://twitter.com/alex193a/status/1571924946183397377

VIP+ Survey: The Demographic Divide Across Media Platforms

Read More About:

Jump to Comments

More from Variety

Most Popular

Must Read

Sign Up for Variety Newsletters

By providing your information, you agree to our Terms of Use and our Privacy Policy.We use vendors that may also process your information to help provide our services. // This site is protected by reCAPTCHA Enterprise and the Google Privacy Policy and Terms of Service apply.

Variety Confidential

ncG1vNJzZmiukae2psDYZ5qopV9nfXN%2Bjp2goKGklrlwusSwqmihnqjBorPRmqRmpqWZtrXFjKmpqKyVmMGqu81mmaWnk6B6sbTOraasZZSiwG59kWxsbHBgaIR6ew%3D%3D

Aldo Pusey

Update: 2024-02-28