

Multimodal Weekly 77: Video Frame Interpolation, Video Restoration, and Multi-Shot Video Understanding
In the 77th session of Multimodal Weekly, we have three exciting presentations on video frame interpolation, video restoration, and multi-shot video understanding.
✅ Zujin Guo will present Generalizable Implicit Motion Modeling (GIMM), a novel and effective approach to motion modeling for Video Frame Interpolation.
✅ Kamran Janjua and Amirhosein Gashemabadi will present Turtle, a method to learn the truncated causal history model for efficient and high-performing video restoration.
✅ Mingfei Han will present Shot2Story, a new multi-shot video understanding benchmark dataset with detailed shot-level captions, comprehensive video summaries and question-answering pairs.
Join the Multimodal Minds community to connect with the speakers!
Multimodal Weekly is organized by Twelve Labs, a startup building multimodal foundation models for video understanding. Learn more about Twelve Labs here: https://twelvelabs.io/