Cover Image for vLLM Inference Meetup: Pune, India
Cover Image for vLLM Inference Meetup: Pune, India
Avatar for vLLM Meetups and Events
Join the vLLM community to discuss optimizing LLM inference!
Registration
Welcome! To join the event, please register below.
About Event

We are excited to invite you to the inaugural vLLM meetup in India hosted by Red Hat at Pune.

​This is your chance to connect with a growing community of vLLM users, developers, maintainers, and engineers from Red Hat. We'll dive deep into technical talks, share insights, and discuss our journey in optimizing LLM inference for performance and efficiency.

​​What to expect:
Technical insights
Networking with industry experts
Hands-on learning & demos

Agenda

09:30-10:00: Registration and Opening Remarks
10:00-10:30: Turning GenAI Investments into Results: Why Inference Matters
10:30-11:00: Introduction to vLLM Inference Engine
11:00-11:30: Advanced Inference Techniques: Quantization, KV Cache, Paged-Attention, and Continuous Batching
11:30-12:00: vLLM Inference Demo – NVIDIA GPU Accelerator
12:00-12:30: Break
12:30-02:00: Hands-on Lab: vLLM Inference

Bring your laptop with SSH installed. GPU instances provided by organizers.

Hosts:
eprasad96@gmail.com
ompragash.viswanathan@gmail.com

Location
Red Hat India Private Limited
Level 3, Tower 6, Magarpatta Inner Circle, Cybercity, Magarpatta, Hadapsar, Pune, Maharashtra 411013, India
Avatar for vLLM Meetups and Events
Join the vLLM community to discuss optimizing LLM inference!