Once something has been shared online, it never truly goes away. This adage is particularly relevant for DeepNude, software that uses AI to create fake nude images of women.
The app came to public attention last week after a report from Motherboard highlighted its existence. Shortly afterward, the app’s creator pulled it from the web, saying that the probability the software would be misused to harass and shame women was “too high.”
Of course, the app is still available, with numerous copies floating around forums and message boards. The Verge was able to find links that ostensibly offer downloads of DeepNude in a variety of places, including Telegram channels, message boards like 4chan, YouTube video descriptions, and even on the…
from The Verge – All Posts https://ift.tt/2YqeEUh