Meetup··Jaipur, India
Stop the GPU Madness! Making LLM Inference Actually Efficient on K8s
AWS User Group Jaipur
LLMKubernetesGPUInferenceAWS
Abstract
AWS User Group Jaipur — main auditorium, RIC Jaipur. A meetup talk on running LLM inference workloads on Kubernetes without burning through GPU budgets.
Resources
More Talks
- Conference
Help! My LLM is a Resource Hog: How We Tamed Inference with Kubernetes and Open Source Muscle
KubeCon + CloudNativeCon North America 2025 · Atlanta, USA
- Meetup
Accelerating CI Pipelines: Rapid Kubernetes Testing with vCluster
Cloud Native & AI Day — Beyond ChatBots · Bengaluru, India
- Meetup
Multitenancy in the Kubernetes Era
Cloud Native Taiwan User Group Meetup · Taipei, Taiwan
- Meetup
Multitenancy in the Kubernetes Era
Incident Management & Cloud Native Meetup · Bengaluru, India