Instagram says it's deploying new tools to protect young people and combat sexual extortion, including a feature that will automatically blur nudity in direct messages. The social media platform said in a blog post Thursday that it's testing the features as part of its campaign to fight sexual scams and other forms of "image abuse," and to make it tougher for criminals to contact teens, per the AP. Sexual extortion, or sextortion, involves persuading a person to send explicit photos online and then threatening to make the images public unless the victim pays money or engages in sexual favors.
Instagram said scammers often use direct messages to ask for "intimate images." To counter this, it will soon start testing out a nudity-protection feature for direct messages that blurs any images with nudity "and encourages people to think twice before sending nude images." The feature "is designed not only to protect people from seeing unwanted nudity in their DMs, but also to protect them from scammers who may send nude images to trick people into sending their own images in return," Instagram said. The feature will be turned on by default globally for teens under 18. Adult users will get a notification encouraging them to activate it.
Images with nudity will be blurred with a warning, giving users the option to view it. They'll also get an option to block the sender and report the chat. For people sending direct messages with nudity, they will get a message reminding them to be cautious when sending "sensitive photos." They'll also be informed that they can unsend the photos if they change their mind, but that there's a chance others may have already seen them. Instagram owner Meta Platforms will not use the blur feature on Facebook or WhatsApp. (Meta's Mark Zuckerberg previously apologized to parents at a congressional hearing for not better protecting young people.)