- The Senate rejected a bipartisan proposal targeting the removal of legal protections for artificial intelligence (AI) technology on Wednesday.
- Republican Missouri Sen. Josh Hawley and Democratic Connecticut Sen. Richard Blumenthal initially introduced their No Section 230 Immunity for AI Act in June, which would have empowered Americans to sue tech platforms over text and visual content their AI produces.
- “The bill is troubling because it would actively undermine one of the legal foundations of the Digital Revolution, thus harming the ability of American innovators to be as innovative and globally competitive as they otherwise would be,” R Street Institute Senior Fellow Adam Thierer told the DCNF. “Section 230 is the secret sauce behind America’s success in digital markets globally. We shouldn’t abandon it lightly.”
The Senate shot down a bipartisan bill Wednesday aimed at stripping legal liability protections for artificial intelligence (AI) technology.
Republican Missouri Sen. Josh Hawley and Democratic Connecticut Sen. Richard Blumenthal first introduced their No Section 230 Immunity for AI Act in June and Hawley put it up for an unanimous consent motion on Wednesday. The bill would have eliminated Section 230 protections that currently grant tech platforms immunity from liability for the text and visual content their AI produces, enabling Americans to file lawsuits against them. (RELATED: White House Announces Artificial Intelligence Vows From Big Tech … But Experts Are Unimpressed)
Today, Sen. Hawley’s expected to ask for unanimous consent to S.1993, the “No Section 230 Immunity for AI Act,” which takes an untargeted/overbroad approach to important/emerging questions.
It would also end important policy debate before it even begins: https://t.co/yXt4bNPrx5
— Center for Democracy & Technology (@CenDemTech) December 12, 2023
Republican Texas Sen. Ted Cruz objected to the unanimous consent motion, expressing concerns about a lack of debate on the bill and its potential “unintended consequences.”
The bill would have had major implications on speech and innovation, experts previously told the Daily Caller News Foundation. Section 230 of the Communications Decency Act of 1996 asserts that internet companies cannot be held liable for third-party speech posted on their platforms.
The No Section 230 Immunity for AI Act defines generative AI as “an artificial intelligence system that is capable of generating novel text, video, images, audio, and other media based on prompts or other forms of data provided by a person.’’ Since it was an unanimous consent motion, it required no objections in order to pass, according to the Senate glossary.
The bill “would threaten freedom of expression, content moderation, and innovation,” a coalition, including technology and liberty advocacy groups, wrote to Senate Majority Leader Chuck Schumer of New York and Senate Minority Leader Mitch McConnell in a Monday letter. “Far from targeting any clear problem, the bill takes a sweeping, overly broad approach, preempting an important public policy debate without sufficient consideration of the complexities at hand.”
NetChoice, a group whose members include companies like Google and TikTok, says the definition is so vague and broad that it could apply to any actions computers take and even calculators.
“Lawmakers are misdirecting their anger at tech platforms when they should be focusing on how the government is trying to use its power to force companies to do the government’s bidding,” NetChoice’s Vice President & General Counsel Carl Szabo told the DCNF.
It could also crush American innovation, R Street Institute Senior Fellow Adam Thierer told the DCNF.
“The bill is troubling because it would actively undermine one of the legal foundations of the Digital Revolution, thus harming the ability of American innovators to be as innovative and globally competitive as they otherwise would be,” Thierer told the DCNF. “Section 230 is the secret sauce behind America’s success in digital markets globally. We shouldn’t abandon it lightly.”
Democratic Oregon Sen. Ron Wyden, one of the authors of Section 230, said it should not apply to AI in comments to The Washington Post in March.
“AI tools like ChatGPT, Stable Diffusion and others being rapidly integrated into popular digital services should not be protected by Section 230,” he told the Post. “And it isn’t a particularly close call … Section 230 is about protecting users and sites for hosting and organizing users’ speech” and it “has nothing to do with protecting companies from the consequences of their own actions and products.”
Hawley and Blumenthal have argued that providing Section 230 liability safeguards to AI would protect tech firms from accountability for the perceived harms of their products.
“We can’t make the same mistakes with generative AI as we did with Big Tech on Section 230,” Hawley stated. “When these new technologies harm innocent people, the companies must be held accountable. Victims deserve their day in court.”
“AI companies should be forced to take responsibility for business decisions as they’re developing products — without any Section 230 legal shield,” Blumenthal stated.
All content created by the Daily Caller News Foundation, an independent and nonpartisan newswire service, is available without charge to any legitimate news publisher that can provide a large audience. All republished articles must include our logo, our reporter’s byline and their DCNF affiliation. For any questions about our guidelines or partnering with us, please contact licensing@dailycallernewsfoundation.org.