filmov
tv
Red Teaming o1 Part 1/2–Automated Jailbreaking w/ Haize Labs' Leonard Tang, Aidan Ewart& Brian Huang
Показать описание
In this Emergency Pod of The Cognitive Revolution, Nathan provides crucial insights into OpenAI's new o1 and o1-mini reasoning models. Featuring exclusive interviews with members of the o1 Red Team from Apollo Research and Haize Labs, we explore the models' capabilities, safety profile, and OpenAI's pre-release testing approach. Dive into the implications of these advanced AI systems, including their potential to match or exceed expert performance in many areas. Join us for an urgent and informative discussion on the latest developments in AI technology and their impact on the future.
Papers mentioned:
SPONSORS:
RECOMMENDED PODCAST:
This Won't Last.
Eavesdrop on Keith Rabois, Kevin Ryan, Logan Bartlett, and Zach Weinberg's monthly backchannel. They unpack their hottest takes on the future of tech, business, venture, investing, and politics.
CHAPTERS:
(00:00:00) About the Show
(00:00:22) About the Episode
(00:05:03) Introduction and Haize Labs Overview
(00:07:36) Universal Jailbreak Technique and Attacks
(00:09:59) Red Teaming Setup for o1
(00:13:47) Automated vs Manual Red Teaming
(00:17:15) Qualitative Assessment of Model Jailbreaking (Part 1)
(00:19:38) Sponsors: Oracle | Brave
(00:21:42) Qualitative Assessment of Model Jailbreaking (Part 2)
(00:21:47) Challenges with Dual Use Cases
(00:26:21) Context-Specific Safety Considerations
(00:32:26) Model Capabilities and Safety Correlation (Part 1)
(00:36:22) Sponsors: Omneky | Squad
(00:37:48) Model Capabilities and Safety Correlation (Part 2)
(00:39:14) New Attack Techniques and Insights
(00:44:42) Model Behavior and Defense Mechanisms
(00:48:23) Current State of Model Jailbreaking
(00:50:33) Automated Jailbreaking Efforts
(00:52:47) Challenges in Preventing Jailbreaks
(00:56:24) Safety, Capabilities, and Model Scale
(01:00:56) Model Classification and Preparedness
(01:02:46) Transparency and Whistleblowing Mechanisms
(01:04:40) Concluding Thoughts on o1 and Future Work
(01:05:54) Outro
SOCIAL LINKS:
Papers mentioned:
SPONSORS:
RECOMMENDED PODCAST:
This Won't Last.
Eavesdrop on Keith Rabois, Kevin Ryan, Logan Bartlett, and Zach Weinberg's monthly backchannel. They unpack their hottest takes on the future of tech, business, venture, investing, and politics.
CHAPTERS:
(00:00:00) About the Show
(00:00:22) About the Episode
(00:05:03) Introduction and Haize Labs Overview
(00:07:36) Universal Jailbreak Technique and Attacks
(00:09:59) Red Teaming Setup for o1
(00:13:47) Automated vs Manual Red Teaming
(00:17:15) Qualitative Assessment of Model Jailbreaking (Part 1)
(00:19:38) Sponsors: Oracle | Brave
(00:21:42) Qualitative Assessment of Model Jailbreaking (Part 2)
(00:21:47) Challenges with Dual Use Cases
(00:26:21) Context-Specific Safety Considerations
(00:32:26) Model Capabilities and Safety Correlation (Part 1)
(00:36:22) Sponsors: Omneky | Squad
(00:37:48) Model Capabilities and Safety Correlation (Part 2)
(00:39:14) New Attack Techniques and Insights
(00:44:42) Model Behavior and Defense Mechanisms
(00:48:23) Current State of Model Jailbreaking
(00:50:33) Automated Jailbreaking Efforts
(00:52:47) Challenges in Preventing Jailbreaks
(00:56:24) Safety, Capabilities, and Model Scale
(01:00:56) Model Classification and Preparedness
(01:02:46) Transparency and Whistleblowing Mechanisms
(01:04:40) Concluding Thoughts on o1 and Future Work
(01:05:54) Outro
SOCIAL LINKS:
Комментарии