This year's DEF CON AI Village has invited hackers to show up, dive in, and find bugs and biases in large language models (LLMs) built by OpenAI, Google, Anthropic, and others.
The collaborative event, which AI Village organizers describe as "the largest red teaming exercise ever for any group of AI models," will host "thousands" of people, including "hundreds of students from overlooked institutions and communities," all of whom will be tasked with finding flaws in LLMs that power today's chat bots and generative AI.
Think: traditional bugs in code, but also problems more specific to machine learning, such as bias, hallucinations, and jailbreaks — all of which ethical and security professionals are now having to grapple with as these technologies scale.
DEF CON is set to run from August 10 to 13 this year in Las Vegas, USA.
My guess is this will be a target rich environment.
Some questions. Are there hard walls (you may not pass this point, you may not change this description) built into the LLM's?
ReplyDeleteIf these systems are so sophisticated, as is the general public's fear, can they not find their own flaws?
Who will eventually determine if a system is unbiased
(it certainly can't be politicians or the media)?