"Are You a CEO, Director, or Founder interested in a Feature Interview?"
All Interviews are 100% FREE of Charge
At the 2023 Defcon hacker conference in Las Vegas, prominent AI technology companies partnered with a group that seeks algorithmic integrity and transparency to drive thousands of participants into generative AI platforms to uncover weaknesses in these critical systems. The “red team” exercise, which was also backed by the U.S. government, was a step toward scrutinizing these increasingly influential but opaque systems. Now, Humane Intelligence, a nonprofit that works on ethical AI and algorithmic evaluation, is taking this model a step further. On Wednesday, The group invited participation In collaboration with the National Institute of Standards and Technology, we opened up the qualifying rounds of a nationwide red team competition to evaluate AI office productivity software to any U.S. resident.
The qualifying round will be held online, with developers Anyone in the general public This is part of NIST’s AI challenge, “Assessing the Risks and Impacts of AI (ARIA).” Participants who make it through the preliminary round will participate in an in-person red team event at the Conference on Applied Machine Learning in Information Security (CAMLIS) in Virginia at the end of October. The goal is to expand the capacity to conduct rigorous testing of generative AI technologies for security, resilience, and ethics.
“The public who use these models don’t really have the ability to judge whether they’re fit for purpose,” says Theo Skjærdas, CEO of Tech Policy Consulting, an AI governance and online safety group that partners with Humane Intelligence. “So we want to democratize the ability to do the evaluation, so that anyone who uses these models can evaluate for themselves whether they’re meeting their needs.”
In the final CAMLIS event, participants will be split into two teams: the red team, which attacks the AI system, and the blue team, which defends it. NIST AI Risk Management FrameworkThe test, known as AI 600-1, is used as a benchmark to measure whether red teams can produce results that violate a system’s expected behavior.
“NIST’s ARIA uses structured user feedback to understand real-world applications of AI models,” said Ruman Chowdhury, founder of Humane Intelligence, a NIST Emerging Technologies Office contractor and member of the Department of Homeland Security’s AI Safety and Security Committee. “The ARIA team is made up of mostly socio-technical testing and evaluation experts. [is] We will use that background to advance the field toward rigorous scientific evaluation of generative AI.”
Chowdhury and Skeerdas say the NIST partnership is just one in a series of AI red team collaborations Human Intelligence will announce with U.S. government agencies, international governments and NGOs in the coming weeks. The effort aims to make it more common for companies and organizations that develop algorithms that are currently black boxes to provide transparency and accountability through mechanisms such as the “Bias Bounty Challenge” to reward individuals who find problems or unfairness in AI models.
“The community should be broader than programmers,” Skærdas says. “Policymakers, journalists, civil society, and non-technical people should all be involved in the process of testing and evaluating these systems, and we need to ensure that under-represented groups, like speakers of minority languages and people with non-majority cultures and perspectives, can participate in this process.”
"Elevate Your Brand with an Exclusive Feature Interview!"