The US Government Wants You—Yes, You—to Hunt Down Generative AI Flaws

Date:

Share:


At the 2023 Defcon hacker conference in Las Vegas, prominent AI tech companies partnered with algorithmic integrity and transparency groups to sic thousands of attendees on generative AI platforms and find weaknesses in these critical systems. This “red-teaming” exercise, which also had support from the US government, took a step in opening these increasingly influential yet opaque systems to scrutiny. Now, the ethical AI and algorithmic assessment nonprofit Humane Intelligence is taking this model one step further. On Wednesday, the group announced a call for participation with the US National Institute of Standards and Technology, inviting any US resident to participate in the qualifying round of a nationwide red-teaming effort to evaluate AI office productivity software.

The qualifier will take place online and is open to both developers and anyone in the general public as part of NIST’s AI challenges, known as Assessing Risks and Impacts of AI, or ARIA. Participants who pass through the qualifying round will take part in an in-person red-teaming event at the end of October at the Conference on Applied Machine Learning in Information Security (CAMLIS) in Virginia. The goal is to expand capabilities for conducting rigorous testing of the security, resilience, and ethics of generative AI technologies.

“The average person utilizing one of these models doesn’t really have the ability to determine whether or not the model is fit for purpose,” says Theo Skeadas, CEO of the AI governance and online safety group Tech Policy Consulting, which works with Humane Intelligence. “So we want to democratize the ability to conduct evaluations and make sure everyone using these models can assess for themselves whether or not the model is meeting their needs.”

The final event at CAMLIS will split the participants into a red team trying to attack the AI systems and a blue team working on defense. Participants will use NIST’s AI risk management framework, known as AI 600-1, as a rubric for measuring whether the red team is able to produce outcomes that violate the systems’ expected behavior.

“NIST’s ARIA is drawing on structured user feedback to understand real-world applications of AI models,” says Humane Intelligence founder Rumman Chowdhury, who is also a contractor in NIST’s Office of Emerging Technologies and a member of the US Department of Homeland Security AI safety and security board. “The ARIA team is mostly experts on sociotechnical test and evaluation, and [is] using that background as a way of evolving the field toward rigorous scientific evaluation of generative AI.”

Chowdhury and Skeadas say the NIST partnership is just one of a series of AI red team collaborations that Humane Intelligence will announce in the coming weeks with US government agencies, international governments, and NGOs. The effort aims to make it much more common for the companies and organizations that develop what are now black-box algorithms to offer transparency and accountability through mechanisms like “bias bounty challenges,” where individuals can be rewarded for finding problems and inequities in AI models.

“The community should be broader than programmers,” Skeadas says. “Policymakers, journalists, civil society, and nontechnical people should all be involved in the process of testing and evaluating of these systems. And we need to make sure that less represented groups like individuals who speak minority languages or are from nonmajority cultures and perspectives are able to participate in this process.”



Source link

━ more like this

From Microsoft to “microslop”: The AI backlash that forced a reset

At some point in 2025, Windows stopped feeling like an operating system and started feeling like a demo for AI. Open Notepad to...

Apple smart glasses might avoid the creepy reputation of Meta Ray-Bans with a light trick

Apple’s upcoming smart glasses could sidestep one of the biggest issues facing the category – privacy concerns – by rethinking something as simple...

The MacBook Neo is moonlighting as a Windows gaming machine, and it’s doing it well

Apple didn’t position its most affordable MacBook as a gaming machine. The MacBook Neo, a budget-leaning laptop that runs on Apple’s A18 Pro...

Apple glasses won’t go brand shopping like Meta did with Ray-Ban and Oakley

When it comes to smart glasses, Apple seems to be taking the road less traveled. While others have leaned on big-name eyewear brands...

I tried this Pokémon-inspired weather app, and checking the weather now feels like a Pokédex hunt

Weather apps are usually one of the most boring things on your phone. You open one, glance at the temperature, maybe check if...
spot_img