OpenAI accused of trying to profit off AI model inspection in court

Date:

Share:



The AISI is supposed to protect the US from risky AI models by conducting safety testing to detect harms before models are deployed. Testing should “address risks to human rights, civil rights, and civil liberties, such as those related to privacy, discrimination and bias, freedom of expression, and the safety of individuals and groups,” President Joe Biden said in a national security memo last month, urging that safety testing was critical to support unrivaled AI innovation.

“For the United States to benefit maximally from AI, Americans must know when they can trust systems to perform safely and reliably,” Biden said.

But the AISI’s safety testing is voluntary, and while companies like OpenAI and Anthropic have agreed to the voluntary testing, not every company has. Hansen is worried that AISI is under-resourced and under-budgeted to achieve its broad goals of safeguarding America from untold AI harms.

“The AI Safety Institute predicted that they’ll need about $50 million in funding, and that was before the National Security memo, and it does not seem like they’re going to be getting that at all,” Hansen told Ars.

Biden had $50 million budgeted for AISI in 2025, but Donald Trump has threatened to dismantle Biden’s AI safety plan upon taking office.

The AISI was probably never going to be funded well enough to detect and deter all AI harms, but with its future unclear, even the limited safety testing the US had planned could be stalled at a time when the AI industry continues moving full speed ahead.

That could largely leave the public at the mercy of AI companies’ internal safety testing. As frontier models from big companies will likely remain under society’s microscope, OpenAI has promised to increase investments in safety testing and help establish industry-leading safety standards.

According to OpenAI, that effort includes making models safer over time, less prone to producing harmful outputs, even with jailbreaks. But OpenAI has a lot of work to do in that area, as Hansen told Ars that he has a “standard jailbreak” for OpenAI’s most popular release, ChatGPT, “that almost always works” to produce harmful outputs.



Source link

━ more like this

How Constructive Shareholder Activism Is Reshaping Public Company Governance – Insights Success

The Changing Image of Shareholder ActivismFor many years, shareholder activism had a reputation for confrontation. Activist investors were often associated with public disputes,...

Is the Dell XPS 16 good for working on the go? Does the Dell XPS have all-day battery life?

The Dell XPS 16 is positioned as a powerful, modern productivity laptop, built for professionals, creatives, and users who need excellent performance without...

Beyond the Boundary Wire: How Yardcare and the New N1600PRO are Leading the Robotic Mower Revolution

The weekend morning dread is a real situation—the realization that the next few hours will be spent toiling with the hum of a...

Watch this moonwalking humanoid robot impress with lifelike agility

A new video (above) out of South Korea features the field tests and interaction capabilities of KAIST Humanoid v0.7, developed at the Korea...

AI frame generation is coming to the PS5 Pro

Sony is preparing to take its PlayStation graphics technology a step further, with AI-powered frame generation now confirmed as a future feature for...
spot_img