Child Helpline International, Safe Online, Internet Watch Foundation, National Center for Missing & Exploited Children, Offlimits, WeProtect Global Alliance, INHOPE
Date
Feb. 10. 2026
Unifying Voices Worldwide: No to Nudify
Nudifying tools have no place in any society that seeks to protect its children.
By
Child Helpline International, Safe Online, Internet Watch Foundation, National Center for Missing & Exploited Children, Offlimits, WeProtect Global Alliance, INHOPE
Date
Feb. 10. 2026
Our message is clear: this functionality serves no good purpose and should be explicitly and universally
illegal. AI must serve as an ally, not a threat, to children. To achieve this mission, we need an immediate, global ban on nudifying tools.
Nudifying tools use AI to generate nude images from clothed photos. Though mainly marketed for adults, these tools are often misused to create non-consensual nudified images of people and are increasingly being used to produce illegal images of children. The companies, developers, and individuals who create or distribute them must be held accountable and face legal and criminal consequences.
Technology that enables the creation of nude and abusive imagery is having an insidious impact on children’s lives around the world. Images of children, including survivors of child sexual abuse, are being manipulated into increasingly extreme and abusive forms.
Deepfakes are now closely linked to sexual coercion, extortion, and blackmail. The alarming ease of these AI tools means abusers no longer need to obtain photographic intimate images; they can create them artificially, at scale, with terrifying efficiency. Offenders, including young people themselves, are already monetising these images, creating new abusive economies.
This technology is not just a tool, it is a new vector for exploitation, putting women and children - particularly girls - at unprecedented risk. They have no place in any society that values the safety and sanctity of childhood. In pursuit of the pact on the dignity and protection of the youngest, we are calling upon governments and legislators to urgently enact and enforce regulation, at the latest within the next two years, to prohibit nudifying tools and ensure they are universally inaccessible.
We also call on institutions, companies, and citizens to take immediate action towards eliminating the use of nudifying tools by committing to the following:
Recognise that nudifying tools inflict irreparable harm on individuals, enabling the indefensible abuse and exploitation of women and children in particular.
Acknowledge the profound societal damage this functionality causes, including the normalisation of explicit imagery and gender-based violence, as well as the erosion of trust and safety.
Demand accountability and innovation, requiring technology companies to implement safety by design, fast-track the development and delivery of effective protections, and provide clear transparency.
We are failing children and ourselves if we continue to ignore these risks. We must act now. We must stand together. We must stop AI-facilitated sexual abuse before it spreads further.
Our Call to Governments and Legislators is to set out an urgent agenda over the next two years
Within these two years, every jurisdiction must:
Ban nudifying tools outright. Outlaw the development, distribution, possession, and commercial use of nudifying apps and nudifying functionalities of AI systems with clear exemptions for legitimate use and safeguards to avoid criminalising children and survivors.
Establish criminal and civil liability. Hold companies, their owners, developers, platforms, and individuals accountable for enabling or profiting from the creation or circulation of nudified imagery.
Mandate universal accessibility blocks. Require platforms, app stores, hosting services, and ISPs to prevent access to nudification technologies.
Ensure (online) media literacy programmes in schools or any other form for citizens of all ages that include guidance on risks of nudifying apps and tools.
Our Call to Technology Companies, Platforms, and Developers
Effective immediately, companies must:
Implement safety-by-design. Ensure no product, model, or feature capable of generating nudified deepfakes can be released without proven protections.
Detect and block nudified content of children. Deploy transparent, robust, privacy preserving detection systems, including cross-platform hashing, model-level safeguards, and proactive scanning.
Prohibit nudification tools across their ecosystems. Ban them from app stores, hosting services, plugins, APIs, and third-party integrations.
Disrupt the abuse economy. Remove monetisation avenues, report offenders, and work with industry partners to cut off payment flows.
Provide transparency. Publish clear reporting on detection, enforcement, model safeguards, and any misuse linked to their tools.
Our Call to Institutions and Citizens
Recognise the harm. Nudifying tools cause severe and lasting damage to victims, enabling abuse, humiliation, and lifelong digital violations.
Understand the societal cost. These tools normalize explicit imagery, drive desensitisation, fuel gender-based violence, and erode trust and safety across communities.
Refuse to tolerate their use. Report them, condemn them, (re)educate others, and help close social and commercial spaces where they thrive.
And especially, keep children safe in every professional or personal capacity from these tools. It takes a village to safely raise a child.
If you’ve been affected by a nudifying app: you are not alone.
Your Rights Matter: You have the right to be heard, to privacy, to protection, and to receive support without blame or shame. Here are some steps you can take right now:
1. Reach Out for Help
If you’re in immediate danger, contact your local emergency numbers (e.g., 112 in the EU, 999 in the UK, 911 in the US) or the police.
If you are in Europe, you can also get support through the Safer Internet Centres (SICs) coordinated by Insafe. SICs provide helplines, reporting guidance, and youth-friendly advice on online harm, including image-based abuse.
2. Report
If someone has generated a nudified image of yourself and you are under 18, you can anonymously report it at your local hotline to have it removed - click here to find your hotline. All you need to provide is the URL where you saw the image. If you can’t find the URL, you can share screenshots, usernames, timestamps, or any other links or group details available.
3. Tell Someone You Trust
Speak with an adult you feel comfortable with - this could be a parent, family member, teacher, school counselor, or social worker. They can help you stay safe and access the support you need.
Want to Join the Call?
This Safer Internet Day, 10 February, we are uniting voices worldwide to demand safety-by-design in AI tools and a global ban on nudifying functionalities. Together, we can make it clear: nudifying tools have no place in a society that values children’s safety and dignity.
Add your name to the call and help amplify this message by filling out this form. To support your advocacy, request a communications package with ready-to-use promotional materials by contacting communications@inhope.org.
Cookies
This website requires the use of cookies to function properly. To learn more, please refer to our cookie policies.