PolicyBrief
S. 2381
119th CongressJul 22nd 2025
Preventing Recurring Online Abuse of Children Through Intentional Vetting of Artificial Intelligence Data Act of 2025
IN COMMITTEE

This bill establishes a voluntary framework, overseen by NIST, for vetting AI training data to detect, remove, and report child pornography, while granting limited liability protection to developers and data collectors who comply.

John Cornyn
R

John Cornyn

Senator

TX

LEGISLATION

AI Data Bill Offers Immunity to Developers Who Voluntarily Scan for Child Exploitation Material

The aptly named PROACTIV Artificial Intelligence Data Act of 2025 tackles a serious, gross problem: the presence of child pornography in the massive datasets used to train AI models. This bill aims to set up a system for finding, removing, and reporting this illegal content, but it does so by creating a specific trade-off: industry best practices in exchange for legal protection.

The Cleanup Crew Gets a Rulebook

At its core, this bill requires the Director of the National Institute of Standards and Technology (NIST) to develop a voluntary framework within one year. This framework is essentially a set of instructions for AI Developers and Data Collectors on how to spot, remove, and report child pornography found in their Covered Datasets—the data scraped from the web specifically for AI training (Sec. 3). NIST has to work with groups like the National Center for Missing and Exploited Children (NCMEC) and the Department of Justice to build this rulebook. Think of it as a standardized safety protocol for cleaning up the digital swamp before AI drinks from it.

The Liability Shield: What Developers Get

Here’s where the bill gets interesting for anyone in the tech industry or anyone who might ever want to sue a tech company. If an AI Developer or Data Collector follows the voluntary framework when handling this illegal material—finding it, taking it down, and reporting it—they are granted immunity from civil lawsuits related to those specific actions (Sec. 4). This means if someone tries to sue them for the way they handled the data cleanup, the case must be dismissed. This is a huge incentive for companies to adopt the voluntary framework, as it acts like a legal shield against costly litigation.

Where the Legal Shield Breaks

That immunity isn't absolute, which is important for accountability. The liability shield disappears if the developer or collector was acting with actual malice, gross negligence, or if they intentionally caused harm (Sec. 4). If a company is seriously sloppy or intentionally malicious in how they handle or report the material, they can still be sued. However, defining terms like “gross negligence” is always a battleground in court, which could mean years of legal wrangling before regular folks see justice.

Who’s Covered and Who Isn’t

One key detail is who gets this legal protection. The immunity applies only to the AI Developer (the company building the system) and the Data Collector (the specialist gathering the training data). It specifically excludes the Artificial Intelligence Deployer—the company that takes a finished AI system and builds it into its own product. If you’re a small business that uses an AI model built by a huge tech firm, and your product somehow runs into data issues, you don't get the same legal protection as the original builder. This creates an uneven playing field where the biggest players get the liability shield, while downstream users and smaller companies are left exposed.

The Catch: It’s All Voluntary

While the bill is strong on providing liability protection, the framework itself is voluntary (Sec. 3). This is the biggest practical challenge. If a large AI developer decides the framework is too costly or cumbersome to implement, they can simply skip it. They won't get the immunity, but they also aren't legally required to clean their data using the NIST guidelines. For everyday people, this means the effectiveness of the bill hinges entirely on whether industry players see the liability protection as valuable enough to justify the effort of scrubbing their data.