

Productizing DeepSeek
DeepSeek in shapes and sizes – whether you’re running a nimble 1.5b or a massive 671b DeepSeek variant, you’re already in the game!
With the rise of diverse DeepSeek models, it’s time to get hands-on with self-hosting, regardless of the model’s footprint. In this webinar, we’re going to explore:
Deploying Every Variant: Learn how to deploy models sized 1.5b, 7b, 8b, 14b, 32b, 70b, and 671b.
Savoring the Flavors: Discover the differences between distilled, quantized, and KV cached models.
Crunching the Costs: Find out what each option will take in terms of resources and expense.
Managed vs. Self-Hosted: Compare traditional managed solutions with self-hosting strategies.
Distributed vLLM Deployment: Dive into the techniques for scaling your deployment across multiple GPUs (get ready for some GPUs to go Brrr!).
Join me next Wednesday for a deep dive into how to productize DeepSeek – no matter the size!