Cover Image for Paper Reading Session - Knowledge Distillation of Large Language Models
Cover Image for Paper Reading Session - Knowledge Distillation of Large Language Models
Avatar for Unify
Presented by
Unify
Build AI Your Way ✨
49 Went

Paper Reading Session - Knowledge Distillation of Large Language Models

Google Meet
Registration
Past Event
Welcome! To join the event, please register below.
About Event

In this session, we're excited to welcome Yuxian Gu, who will be presenting his work on: Knowledge Distillation of Large Language Models. Yuxian is a third year PhD student in the conversational AI group at Tsinghua University, advised by Prof. Minlie Huang. In this work, they replace the forward Kullback-Leibler divergence (KLD) objective in the standard KD approaches with reverse KLD. This prevents the student model from overestimating low-probability regions of the teacher distribution, and the resultant models (MiniLLMs) generate more precise responses with higher overall quality compared to KD baselines. Exciting stuff, see you there!

Avatar for Unify
Presented by
Unify
Build AI Your Way ✨
49 Went