This Is More Common Than You Think
AI "nudify" tools—sometimes called deepnude, undress AI, clothesoff, or similar names—have exploded in availability. They run on websites, Telegram bots, and smartphone apps. Millions of images have been generated.
You're not alone. The people using these tools against others are criminals, and the legal system is catching up.
What Are AI Nudify Apps?
Understanding the technology helps you understand your options
How They Work
AI nudify tools use machine learning to take any clothed photo and generate a fake "nude" version. They analyze the body shape and generate synthetic nudity—none of it is real.
Where They Exist
Telegram bots (most common), dedicated websites, mobile apps, and Discord servers. Many operate from overseas but are still subject to US law if they serve US users.
Who Uses Them
Unfortunately, anyone with a photo can use these tools. Victims include high school students targeted by classmates, women targeted by acquaintances, and public figures targeted by strangers.
Your Legal Rights in 2025
The law has evolved to specifically address AI-generated content
Federal Law: TAKE IT DOWN Act
- Creating AI deepfake porn without consent is a federal crime
- Platforms must remove content within 48 hours
- Penalties include up to 2-3 years in prison
- Explicitly covers AI-generated intimate imagery
State Laws
- California: Civil action with $1,500+ statutory damages
- Texas: Class A misdemeanor, up to 1 year jail
- Virginia: Criminal penalties + civil remedies
- New York: Civil action with up to $30,000 damages
Take Action Now
Here's exactly what to do in the next 24-48 hours
Document the Evidence
Screenshot everything: the fake images, URLs, usernames, the tool/app name if visible, any messages from the creator. Save the date and time. This evidence is essential for reports and potential legal action.
Report to Platforms (48-Hour Response Required)
Report the content using each platform's NCII (non-consensual intimate imagery) reporting process. Clearly state the images are AI-generated fakes. Under the TAKE IT DOWN Act, platforms must respond within 48 hours.
File with Google
Google has a specific form for removing fake pornography from search results. Even if the source site is uncooperative, getting delisted from Google means no one can find it by searching your name.
Create Hashes with StopNCII.org
If you have access to copies of the fake images, create digital 'fingerprints' that will automatically block them on Meta, TikTok, Reddit, and other platforms—preventing re-upload.
Report to Law Enforcement (If You Know Who Did It)
If you know or suspect who created the images, file a police report citing your state's deepfake law or the federal TAKE IT DOWN Act. Also report to FBI IC3 at ic3.gov.
Consider a Civil Lawsuit
If you can identify the creator, you can sue for damages—including emotional distress, statutory damages (in some states), and attorney's fees. Many attorneys handle these cases on contingency.
Where to Report
Direct links to reporting for common platforms where this content appears
Telegram Bots
Most nudify tools run on Telegram. Report the bot and any channels sharing the content.
→ Report to @notoscam or abuse@telegram.org
Google Search
Get fake porn removed from Google search results.
→ Use fake pornography removal form
Discord Servers
Many tools are shared in Discord communities.
→ Trust & Safety report
Report any posts or subreddits sharing the content.
→ Select 'Non-consensual intimate media'
X (Twitter)
Report accounts sharing nudify content.
→ Intimate media report
StopNCII.org
Create hashes to auto-block content across platforms.
→ Free hash creation tool
Common Questions
These images are fake—does that make them less harmful?
No. The emotional impact on victims is real regardless of whether the images are authentic. The law recognizes this: AI-generated intimate imagery is treated the same as real NCII under both federal and most state laws.
The person used a Telegram bot—can anything be done?
Yes. While Telegram is notoriously difficult for takedowns, there are strategies that work: reporting to @notoscam, filing DMCA claims, targeting the bot's payment processors, and getting associated links/channels banned. The content can also be blocked from spreading to other platforms.
What if I don't know who created the images?
You can still get the content removed from platforms, delisted from Google, and blocked from spreading via StopNCII.org. Legal action against the creator requires identifying them, but content removal doesn't.
I'm a student and someone at my school did this—what do I do?
Report to school administration immediately—schools can take disciplinary action including expulsion. Also file with police. If you're under 18, report to NCMEC's CyberTipline (1-800-843-5678). This is taken extremely seriously when minors are involved.
How do I prevent this from happening again?
While you can't completely prevent someone from using your public photos, you can: limit high-quality full-body photos on public social media, use privacy settings, set up Google Alerts for your name, and use StopNCII.org proactively if you have concerns.
Will people be able to tell the images are fake?
Often, yes—AI-generated images frequently have telltale artifacts, strange lighting, or body proportions that don't match. But even convincing fakes can be contextualized as fake. If needed, you can proactively explain the situation to people who might see them.
This Is Not Your Fault
Having photos on social media—even bikini photos, dating app photos, or any other photos—doesn't make you responsible for this abuse. The person who fed your photos into an AI nudify tool is 100% responsible.
Don't let shame prevent you from taking action. You are the victim of a crime, and you deserve support.