What Is the TAKE IT DOWN Act?
The TAKE IT DOWN Act is the 2026 federal law establishing a 48-hour removal mandate for non-consensual intimate imagery, including AI-generated deepfakes, on covered online platforms. It is codified at 47 U.S.C. § 223a.
The Targeting and Eradicating Non-Consensual Images to Obstruct Wrongful Nudes — TAKE IT DOWN — Act was signed into law in 2026 after years of advocacy by NCII survivors and organizations. It creates three key obligations: covered platforms must establish a mechanism to receive removal notices; they must remove verified NCII within 48 hours of receipt; and they must implement reasonable measures to prevent re-upload of removed content. The law covers both authentic and AI-generated intimate imagery and specifically addresses the re-upload problem that plagued earlier state-level NCII statutes.
Key facts about this term
-
Platforms must have a notice mechanism All covered platforms must establish and publicize a process for submitting NCII removal notices. Platforms that lack such a mechanism are in violation before any content is even reported.
-
48-hour removal is mandatory, not discretionary Unlike DMCA takedown, which creates a safe harbor framework, the TAKE IT DOWN Act creates an affirmative obligation. Platforms that act within 48 hours are protected. Those that do not face civil and criminal liability.
-
Re-upload prevention is required Platforms must implement 'reasonable technical measures' to prevent re-upload of content that has been removed under a valid notice. This was a key advocacy priority to prevent the 'whack-a-mole' problem.
Frequently asked questions
Which platforms are covered by the TAKE IT DOWN Act?
Covered platforms include any internet-based service that hosts or shares user-generated content and has more than a threshold number of U.S. users. This includes social media, adult content platforms, image hosting sites, and messaging platforms.
What happens if a platform ignores my removal notice?
Platform non-compliance creates civil liability. Victims can sue non-compliant platforms and may be entitled to statutory damages. The Department of Justice may also initiate criminal proceedings in egregious cases.
Find and remove your photos now
Upload a photo. We scan 2.4 billion face embeddings and send legal removal notices in 48 hours.
Remove photos now
ScanErase