72 subscribers
Treceți offline cu aplicația Player FM !
Red Teaming o1 Part 1/2– Automated Jailbreaking with Haize Labs' Leonard Tang, Aidan Ewart, and Brian Huang
Manage episode 439899337 series 3452589
In this Emergency Pod of The Cognitive Revolution, Nathan provides crucial insights into OpenAI's new o1 and o1-mini reasoning models. Featuring exclusive interviews with members of the o1 Red Team from Apollo Research and Haize Labs, we explore the models' capabilities, safety profile, and OpenAI's pre-release testing approach. Dive into the implications of these advanced AI systems, including their potential to match or exceed expert performance in many areas. Join us for an urgent and informative discussion on the latest developments in AI technology and their impact on the future.
- o1 Safety Card
- Haize Labs
- Endless Jailbreaks with Bijection Learning: a Powerful, Scale-Agnostic Attack Method
- Haize Labs Job board
Papers mentioned:
Apply to join over 400 Founders and Execs in the Turpentine Network: https://www.turpentinenetwork.co/
SPONSORS:
Oracle: Oracle Cloud Infrastructure (OCI) is a single platform for your infrastructure, database, application development, and AI needs. OCI has four to eight times the bandwidth of other clouds; offers one consistent price, and nobody does data better than Oracle. If you want to do more and spend less, take a free test drive of OCI at https://oracle.com/cognitive
Brave: The Brave search API can be used to assemble a data set to train your AI models and help with retrieval augmentation at the time of inference. All while remaining affordable with developer first pricing, integrating the Brave search API into your workflow translates to more ethical data sourcing and more human representative data sets. Try the Brave search API for free for up to 2000 queries per month at https://bit.ly/BraveTCR
Omneky: Omneky is an omnichannel creative generation platform that lets you launch hundreds of thousands of ad iterations that actually work customized across all platforms, with a click of a button. Omneky combines generative AI and real-time advertising data. Mention "Cog Rev" for 10% off https://www.omneky.com/
Squad: Head to Squad to access global engineering without the headache and at a fraction of the cost: head to https://choosesquad.com/ and mention “Turpentine” to skip the waitlist.
RECOMMENDED PODCAST:
This Won't Last.
Eavesdrop on Keith Rabois, Kevin Ryan, Logan Bartlett, and Zach Weinberg's monthly backchannel. They unpack their hottest takes on the future of tech, business, venture, investing, and politics.
Apple Podcasts: https://podcasts.apple.com/us/podcast/id1765665937
Spotify: https://open.spotify.com/show/2HwSNeVLL1MXy0RjFPyOSz
YouTube: https://www.youtube.com/@ThisWontLastpodcast
CHAPTERS:
(00:00:00) About the Show
(00:00:22) About the Episode
(00:05:03) Introduction and Haize Labs Overview
(00:07:36) Universal Jailbreak Technique and Attacks
(00:13:47) Automated vs Manual Red Teaming
(00:17:15) Qualitative Assessment of Model Jailbreaking (Part 1)
(00:19:38) Sponsors: Oracle | Brave
(00:21:42) Qualitative Assessment of Model Jailbreaking (Part 2)
(00:26:21) Context-Specific Safety Considerations
(00:32:26) Model Capabilities and Safety Correlation (Part 1)
(00:36:22) Sponsors: Omneky | Squad
(00:37:48) Model Capabilities and Safety Correlation (Part 2)
(00:44:42) Model Behavior and Defense Mechanisms
(00:52:47) Challenges in Preventing Jailbreaks
(00:56:24) Safety, Capabilities, and Model Scale
(01:00:56) Model Classification and Preparedness
(01:04:40) Concluding Thoughts on o1 and Future Work
(01:05:54) Outro
211 episoade
Red Teaming o1 Part 1/2– Automated Jailbreaking with Haize Labs' Leonard Tang, Aidan Ewart, and Brian Huang
"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis
Manage episode 439899337 series 3452589
In this Emergency Pod of The Cognitive Revolution, Nathan provides crucial insights into OpenAI's new o1 and o1-mini reasoning models. Featuring exclusive interviews with members of the o1 Red Team from Apollo Research and Haize Labs, we explore the models' capabilities, safety profile, and OpenAI's pre-release testing approach. Dive into the implications of these advanced AI systems, including their potential to match or exceed expert performance in many areas. Join us for an urgent and informative discussion on the latest developments in AI technology and their impact on the future.
- o1 Safety Card
- Haize Labs
- Endless Jailbreaks with Bijection Learning: a Powerful, Scale-Agnostic Attack Method
- Haize Labs Job board
Papers mentioned:
Apply to join over 400 Founders and Execs in the Turpentine Network: https://www.turpentinenetwork.co/
SPONSORS:
Oracle: Oracle Cloud Infrastructure (OCI) is a single platform for your infrastructure, database, application development, and AI needs. OCI has four to eight times the bandwidth of other clouds; offers one consistent price, and nobody does data better than Oracle. If you want to do more and spend less, take a free test drive of OCI at https://oracle.com/cognitive
Brave: The Brave search API can be used to assemble a data set to train your AI models and help with retrieval augmentation at the time of inference. All while remaining affordable with developer first pricing, integrating the Brave search API into your workflow translates to more ethical data sourcing and more human representative data sets. Try the Brave search API for free for up to 2000 queries per month at https://bit.ly/BraveTCR
Omneky: Omneky is an omnichannel creative generation platform that lets you launch hundreds of thousands of ad iterations that actually work customized across all platforms, with a click of a button. Omneky combines generative AI and real-time advertising data. Mention "Cog Rev" for 10% off https://www.omneky.com/
Squad: Head to Squad to access global engineering without the headache and at a fraction of the cost: head to https://choosesquad.com/ and mention “Turpentine” to skip the waitlist.
RECOMMENDED PODCAST:
This Won't Last.
Eavesdrop on Keith Rabois, Kevin Ryan, Logan Bartlett, and Zach Weinberg's monthly backchannel. They unpack their hottest takes on the future of tech, business, venture, investing, and politics.
Apple Podcasts: https://podcasts.apple.com/us/podcast/id1765665937
Spotify: https://open.spotify.com/show/2HwSNeVLL1MXy0RjFPyOSz
YouTube: https://www.youtube.com/@ThisWontLastpodcast
CHAPTERS:
(00:00:00) About the Show
(00:00:22) About the Episode
(00:05:03) Introduction and Haize Labs Overview
(00:07:36) Universal Jailbreak Technique and Attacks
(00:13:47) Automated vs Manual Red Teaming
(00:17:15) Qualitative Assessment of Model Jailbreaking (Part 1)
(00:19:38) Sponsors: Oracle | Brave
(00:21:42) Qualitative Assessment of Model Jailbreaking (Part 2)
(00:26:21) Context-Specific Safety Considerations
(00:32:26) Model Capabilities and Safety Correlation (Part 1)
(00:36:22) Sponsors: Omneky | Squad
(00:37:48) Model Capabilities and Safety Correlation (Part 2)
(00:44:42) Model Behavior and Defense Mechanisms
(00:52:47) Challenges in Preventing Jailbreaks
(00:56:24) Safety, Capabilities, and Model Scale
(01:00:56) Model Classification and Preparedness
(01:04:40) Concluding Thoughts on o1 and Future Work
(01:05:54) Outro
211 episoade
Toate episoadele
×1 Material Progress: Developing AI's Scientific Intuition, with Orbital Materials' Jonathan Godwin & Tim Duignan 1:40:50
1 Dodging Latent Space Detectors: Obfuscated Activation Attacks with Luke, Erik, and Scott. 2:10:23
1 Gene Hunting with o1-pro: Reasoning about Rare Diseases with ChatGPT Pro Grantee Dr. Catherine Brownstein 1:33:29
1 AI AMA – Part 2: AI Utopia, Consciousness, and the Future of Work 2:01:36
1 AI AMA – Part 1: OpenAI’s o3, Deliberative Alignment, and AI Surprises of 2024 2:06:57
1 Teaching AI to See: A Technical Deep-Dive on Vision Language Models with Will Hardman of Veratai 3:56:09
1 roon's Heroic Duty: Will "the Good Guys" Build AGI First? (from Doom Debates) 1:57:58
1 Emad Mostaque on the Intelligent Internet and Universal Basic AI 2:11:52
1 Can AIs do AI R&D? Reviewing REBench Results with Neev Parikh of METR 1:47:58
1 Breakthroughs in AI for Biology: AI Lab Groups & Protein Model Interpretability with Prof James Zou 1:02:49
1 Scouting Frontiers in AI for Biology: Dynamics, Diffusion, and Design, with Amelie Schreiber 1:47:28
1 Building Government's Largest Civilian AI Team with DHS AI Corps' Director, Michael Boyce 1:30:11
1 Emergency Pod: o1 Schemes Against Users, with Alexander Meinke from Apollo Research 2:06:52
1 Automating Scientific Discovery, with Andrew White, Head of Science at Future House 1:58:32
1 The Evolution of AI Agents: Lessons from 2024, with MultiOn CEO Div Garg 1:30:21
Bun venit la Player FM!
Player FM scanează web-ul pentru podcast-uri de înaltă calitate pentru a vă putea bucura acum. Este cea mai bună aplicație pentru podcast și funcționează pe Android, iPhone și pe web. Înscrieți-vă pentru a sincroniza abonamentele pe toate dispozitivele.