There's been a creepy rise in people using AI apps to 'undress' women in photos, data shows

  • AI apps and websites that create fake nude images are seeing a rise in the number of visitors.

  • Analysis firm Graphika found a group of 34 such platforms garnered 24 million visitors in September.

  • The platforms use AI to make photos of fully clothed women appear naked.

AI apps that can "undress" women in photos are seeing higher numbers of users, data showed.

Such "undressing" sites and apps let users upload an image of someone fully clothed. AI is then used to generate a fake nude image of them.

A group of 34 websites attracted more than 24 million unique visitors in September, social-media analysis firm Graphika said in its December report, "A Revealing Picture." The report cited data from traffic-analysis site Similarweb.

The report also said that the number of referral link spam for these sites and apps has risen by more than 2,000% since the beginning of this year on platforms including X and Reddit.

A further 53 Telegram groups that are used to access those services have at least 1 million users, the report said.

The report's researchers said, "We assess the increasing prominence and accessibility of these services will very likely lead to further instances of online harm, such as the creation and dissemination of non-consensual nude images, targeted harassment campaigns, sextortion, and the generation of child sexual abuse material."

Apps that create deepfake nudes, digitally altered images to make a person appear nude in a photograph, have been around for several years. They are typically created without the subject's consent and have previously targeted celebrities and online personalities.

In February, Twitch streamer QTCinderella fell victim to deepfake porn videos of her circulating online.

The day after a screenshot from the video was shared online, the streamer said in an X post: "The amount of body dysmorphia I've experienced since seeing those photos has ruined me."

She added: "It's not as simple as 'just' being violated. It's so much more than that."

In September, fake naked images of more than 20 girls were circulated in several Spanish schools. The photos were processed by an AI application, per El País.

Read the original article on Business Insider