Instagram* is working on a mechanism that will protect users of the platform from unwanted nude photos in private messages. A spokesperson for the service’s owner, Meta*, confirmed that the feature is indeed in development, after researcher Alessandro Paluzzi published relevant screenshots of the application, informs The Verge.
According to a Meta* spokesperson, the new security features will help users protect themselves from “nudity” and other unwanted messages. Similar features have appeared before – one of them filters requests through the private messaging system, blocking offensive content. The company clarified that Meta* specialists will not have access to user correspondence and will not be able to share this data with third parties. In more detail, the company will talk about new protection tools in the coming weeks, closer to the start of testing.
Meanwhile, the UK non-profit organization Center for Countering Digital Hate in April establishedthat the Instagram* protection system does not work effectively enough, skipping 9 out of 10 offensive messages against famous women: the protection did not respond to either male intimate pictures or foul language. And early last year, the Pew Research Center published reportaccording to which 33% of women under the age of 35 have experienced online harassment.
* Included in the list of public associations and religious organizations in respect of which the court made a final decision to liquidate or ban activities on the grounds provided for by Federal Law No. 114-FZ of July 25, 2002 “On countering extremist activity.”
If you notice an error, select it with the mouse and press CTRL + ENTER.