Have you been targeted by non-consensual deepfake pornography?
While “revenge porn” – a form of harassment that involves posting explicit images of someone online without their consent – has been an issue for years, new AI tools make it possible to artificially generate nude images that look strikingly real.
Women around the world, from Taylor Swift and Rep. Alexandria Ocasio-Cortez to high school girls in New Jersey and Texas, have been targeted by non-consensual, AI-generated sexual images in recent months. Social media companies are facing new pressure to address the spread of non-consensual images. And lawmakers are now scrambling to find a way to crack down on the people who create such images and the online tools that enable them.
Have you or your child been the target of this form of harassment? We want to hear about what it meant for your life and how (and if) law enforcement or school officials responded.
Tell us about your experiences in the form below. Providing commentary anonymously can be discussed.
Inresponse to the rising issue, various tech companies are actively exploring ways to improve their algorithms and detect AI-generated non-consensual images in their platforms. Furthermore, businesses and tech giants are under scrutiny to incorporate stronger privacy measures and ethical guidelines to prevent such misuse of technology.