This bill establishes a voluntary framework, overseen by NIST, for vetting AI training data to detect, remove, and report child pornography, while granting limited liability protection to developers and data collectors who comply.
John Cornyn
Senator
TX
This bill, the PROACTIV Artificial Intelligence Data Act of 2025, directs the National Institute of Standards and Technology (NIST) to create voluntary guidelines for detecting, removing, and reporting child pornography found in datasets used to train AI systems. It establishes definitions for key industry roles and provides limited legal liability protection to AI developers and data collectors who follow these new guidelines when handling such illegal material. The legislation also mandates federal support for research into better detection methods for this harmful content.
The aptly named PROACTIV Artificial Intelligence Data Act of 2025 tackles a serious, gross problem: the presence of child pornography in the massive datasets used to train AI models. This bill aims to set up a system for finding, removing, and reporting this illegal content, but it does so by creating a specific trade-off: industry best practices in exchange for legal protection.
At its core, this bill requires the Director of the National Institute of Standards and Technology (NIST) to develop a voluntary framework within one year. This framework is essentially a set of instructions for AI Developers and Data Collectors on how to spot, remove, and report child pornography found in their Covered Datasets—the data scraped from the web specifically for AI training (Sec. 3). NIST has to work with groups like the National Center for Missing and Exploited Children (NCMEC) and the Department of Justice to build this rulebook. Think of it as a standardized safety protocol for cleaning up the digital swamp before AI drinks from it.
Here’s where the bill gets interesting for anyone in the tech industry or anyone who might ever want to sue a tech company. If an AI Developer or Data Collector follows the voluntary framework when handling this illegal material—finding it, taking it down, and reporting it—they are granted immunity from civil lawsuits related to those specific actions (Sec. 4). This means if someone tries to sue them for the way they handled the data cleanup, the case must be dismissed. This is a huge incentive for companies to adopt the voluntary framework, as it acts like a legal shield against costly litigation.
That immunity isn't absolute, which is important for accountability. The liability shield disappears if the developer or collector was acting with actual malice, gross negligence, or if they intentionally caused harm (Sec. 4). If a company is seriously sloppy or intentionally malicious in how they handle or report the material, they can still be sued. However, defining terms like “gross negligence” is always a battleground in court, which could mean years of legal wrangling before regular folks see justice.
One key detail is who gets this legal protection. The immunity applies only to the AI Developer (the company building the system) and the Data Collector (the specialist gathering the training data). It specifically excludes the Artificial Intelligence Deployer—the company that takes a finished AI system and builds it into its own product. If you’re a small business that uses an AI model built by a huge tech firm, and your product somehow runs into data issues, you don't get the same legal protection as the original builder. This creates an uneven playing field where the biggest players get the liability shield, while downstream users and smaller companies are left exposed.
While the bill is strong on providing liability protection, the framework itself is voluntary (Sec. 3). This is the biggest practical challenge. If a large AI developer decides the framework is too costly or cumbersome to implement, they can simply skip it. They won't get the immunity, but they also aren't legally required to clean their data using the NIST guidelines. For everyday people, this means the effectiveness of the bill hinges entirely on whether industry players see the liability protection as valuable enough to justify the effort of scrubbing their data.