"Fake Nudes: Youths Confront Harms"
Teens increasingly traumatized by deepfake nudes clearly understand that the AI-generated images are harmful. A surprising recent Thorn survey suggests there's growing consensus among young people under 20 that making and sharing fake nudes is obviously abusive. The stigma surrounding creating and distributing non-consensual nudes appears to be shifting, with many teens now recognizing it as a serious form of abuse.
- As the normalization of deepfakes in entertainment becomes more widespread, it will be crucial for tech companies and lawmakers to adapt their content moderation policies and regulations to protect young people from AI-generated sexual material.
- What role can educators and mental health professionals play in supporting young victims of non-consensual sharing of fake nudes, particularly in schools that lack the resources or expertise to address this issue?