The "TAKE IT DOWN Act" criminalizes the intentional sharing of nonconsensual intimate visual depictions, including digital forgeries, and requires online platforms to establish a process for the removal of such content upon request.
Ted Cruz
Senator
TX
The TAKE IT DOWN Act aims to combat the nonconsensual sharing of intimate visual depictions, including deepfakes, by establishing criminal penalties for those who knowingly publish such content and requiring online platforms to create a process for individuals to request the removal of these depictions. Platforms must remove the content within 48 hours of a valid request and make reasonable efforts to remove any known identical copies. Failure to comply with these takedown obligations will be treated as a violation under the Federal Trade Commission Act. The Act also defines key terms related to consent, digital forgeries, and covered platforms.
Party | Total Votes | Yes | No | Did Not Vote |
---|---|---|---|---|
Republican | 220 | 207 | 2 | 11 |
Democrat | 213 | 202 | 0 | 11 |
The TAKE IT DOWN Act is a new bill that tackles the growing problem of nonconsensual intimate images, including deepfakes, being shared online. It does this by creating criminal penalties for those who share such material and by forcing websites to take down this content quickly when notified. Here's a breakdown of what it all means:
This bill sets up some serious consequences for anyone who intentionally shares intimate images or videos of someone without their permission, especially if it's meant to cause harm. This includes "digital forgeries" – or what most of us call deepfakes. If you're caught sharing this kind of content of an adult, you could face fines and up to 2 years in prison (Sec. 2). If the content involves a minor, the penalties jump up to 3 years. Even threatening to share these images can land you in jail. Importantly, the bill clarifies that just because someone agreed to create an image doesn't mean they agreed to have it shared online (Sec. 2).
This is where things get interesting for websites and apps. The bill puts the heat on "covered platforms" – basically, any site or app where users can post their own content (think social media, forums, etc.) – to create a system for people to report nonconsensual intimate images and get them removed (Sec. 3). These platforms have one year to set up a clear, easy-to-find reporting process. Once someone makes a valid request, the platform has just 48 hours to remove the image and try to find any identical copies (Sec. 3). The upside for these platforms? If they act in "good faith" to remove content, they're protected from liability, even if it turns out the content wasn't actually illegal (Sec. 3).
While the bill aims to protect people from having their intimate images used against them, there are a few things to consider. First, what exactly counts as an "intimate visual depiction" is pretty broad (Sec. 2), and some worry this could lead to legitimate content being taken down. Second, that 48-hour takedown window could be tough for smaller platforms to meet consistently. It is also important to know that the definition of "covered platform" has some exceptions. It doesn't include things like internet service providers or email services, or sites that mainly offer pre-selected content where user interaction is limited (Sec. 4).
The TAKE IT DOWN Act tries to strike a balance between protecting individuals and not crushing free speech online. It gives the FTC significant power to enforce these new rules, and it will be interesting to see how platforms adapt. Whether it truly solves the problem of nonconsensual intimate images, or creates new headaches, remains to be seen.