Jailbreaking LLMs and Agentic Systems

Attacks, Defenses, and Evaluations

ICML 2025 Tutorial

Hamed Hassani

Hamed Hassani

Penn & Google

Amin Karbasi

Amin Karbasi

Yale & Robust Intelligence

Alex Robey

Alex Robey

CMU & Gray Swan

Abstract. This website contains resources from the ICML 2025 tutorial Jailbreaking LLMs and Agentic Systems. The tutorial provides an overview of recent advances in jailbreaking attacks, defenses, and evaluation methods, with particular emphasis on applications in robotics and agentic systems.