vLLM: Easy, Fast, and Cheap LLM Serving for Everyone
โFor developers, builders, AI enthusiasts, and anyone looking to optimize LLM serving and opportunity to contribute to open source project.
โ
๐
When: July 18th, 2024
โโฐ Time: 10:30 AM PST
โ๐ Where: Zoom Webinar
โ
๐ Agenda:
โThis presentation introduces vLLM, a high-performance open-source LLM inference engine. We'll cover:
โOverview of vLLM and its key benefits
โDeep dive into features like pipeline parallelism and speculative decoding
โLive demonstration of setting up and optimizing LLM inference
โContribution opportunities and current development priorities
โ
๐ค Speakers:
โWoosuk Kwon - Ph.D. student at UC Berkeley, creator of vLLM
โKaichao You - Ph.D. student at Tsinghua University, vLLM contributor
โ๐ซ What's the Deal?
โWe promise 1 hour of cutting-edge AI technology, insightful demonstrations, and the chance to connect with the creators of a amazing open-source project. Learn how to significantly improve your AI infrastructure's performance and cost-efficiency!
โ๐ Resources:
โvLLM GitHub: https://github.com/vllm-project/vllm