Cover Image for Escaping Alignment: LLM Jailbreaks and their defenses
Cover Image for Escaping Alignment: LLM Jailbreaks and their defenses
31 Went

Escaping Alignment: LLM Jailbreaks and their defenses

Registration
Past Event
Welcome! To join the event, please register below.
About Event

Samuel Šimko will share his recent work on jailbreaking LLMs and why it matters for AI-safety research.

The speaker will walk you through his contest-winning manual jailbreaking techniques, explore the current landscape
of automated attacks and defenses, and conclude with a brief presentation of his master's thesis, which proposes a novel
defense against adversarial attacks on LLMs.

Expect an accessible session aimed at students and practitioners curious about Large Language Model security.

Location
ETH Zürich Hauptgebäude
Rämistrasse 101, 8092 Zürich, Switzerland
HG E22
31 Went