StreamYard On-Air

Lessons from Scaling llm-d on DigitalOcean

Want to scale beyond a single-GPU demo—without getting buried in infrastructure? Join DigitalOcean’s Solution Architects for a live walkthrough of how they deployed llm-d with vLLM on DigitalOcean Kubernetes (DOKS). You’ll see how to spin up a cluster, load a 7B model, and serve real-time completions—all with a starter repo you can use right away. Skip the guesswork, grab the template, and help guide future deep dives. See you there!

Already registered?