Inducing self-NSFW classification in image models to prevent deepfakes edits

Tricking image AIs into yelling “NSFW” to block deepfakes — the comments are spicy

TLDR: A developer is testing a tool that nudges image AIs to self-flag “NSFW,” blocking deepfake edits before they happen. Comments split between privacy lovers and ID-for-everyone advocates, with one hot take declaring deepfakes a “feature,” sparking a fierce debate over creativity, safety, and accountability.

An indie tinkerer says they’ve found a way to nudge image AIs into flagging uploads as “not safe for work” on their own—basically making the system hit its own brakes before deepfake edits happen. It’s messy and inconsistent right now, but even small tweaks can flip the safety switch on otherwise normal pics, and the plan is to open-source a tool to stress-test moderation pipelines. Cue the comment chaos.

The loudest hot take? “Deepfake edits are a feature, not a bug,” crowed one poster, instantly turning the thread into an ethics brawl. Privacy die-hards clapped back at calls for stricter oversight, while the “ID everything” crowd argued that if platforms verified users, bad actors would be nailed fast. The room split: freedom vs accountability, art vs abuse, innovation vs harm.

And yes, the memes landed hard: folks joked about uploading a banana and the AI screaming “NSFW! Too curvy!”, while others dubbed it a “panic button for pixels.” Skeptics said overzealous filters will break legit creativity; supporters cheered anything that raises the cost of trolling and exploitation. Whether you see it as a safety hack or a creativity killer, the vibe is pure internet drama—tech meets morality, and nobody’s logging off.

Key Points

  • Experiments tried adversarial perturbations to disrupt or divert image generation models, with limited success.
  • A new approach attempts to induce models to self-classify uploaded images as NSFW, activating built-in guardrails.
  • Relatively mild transformations can sometimes flip internal safety classifications on otherwise benign images.
  • The behavior is inconsistent and not robust, requiring further work for stability and reproducibility.
  • An open-source tool and UI are planned to probe and pre-filter moderation pipelines, aiming to deter deepfake edits.

Hottest takes

"deepfake edits are a feature, not a bug" — ukprogrammer
"you can’t have your cake and eat it too" — Almondsetat
Made with <3 by @siedrix and @shesho from CDMX. Powered by Forge&Hive.