Aflevering 116: Running AI on Kubernetes: From GPUs to CRO

Jan Stomphorst
Ronald Kers
Luister naar deze aflevering op jouw favoriete platform!
Apple Podcast Icon - Radio Webflow TemplateSpotify Icon- Radio Webflow TemplateGoogle Podcast Icon - Radio Webflow TemplateAnchor Icon - Radio Webflow TemplateSoundCloud Icon - Radio Webflow Template
Aflevering 116: Running AI on Kubernetes: From GPUs to CRO
October 28, 2025
42
 MIN

Aflevering 116: Running AI on Kubernetes: From GPUs to CRO

Carlos joins us to explore what it really takes to run AI workloads on Kubernetes, from GPU scheduling to scaling inference and training efficiently across clusters.

Samenvatting

In this episode of De Nederlandse Kubernetes Podcast, we talk with Carlos Santana, Principal Partner Solution Architect at AWS and long-time contributor to the Kubernetes and AI communities.

Carlos joins us to explore what it really takes to run AI workloads on Kubernetes, from GPU scheduling to scaling inference and training efficiently across clusters. We discuss how AI and machine learning are transforming the cloud-native ecosystem — and why orchestration is becoming just as important as the models themselves.

He shares insights into:

  • 💡 The challenges of scheduling and sharing GPUs in multi-tenant Kubernetes clusters
  • ⚙️ Why Kubernetes Resource Orchestrator (CRO) could be the next big abstraction layer
  • 🚀 The balance between performance, cost efficiency, and developer experience
  • 🧠 His hands-on experiments with Jetson devices, edge computing, and model optimization
  • 🌐 How open source projects and cloud providers are shaping the future of AI infrastructure

A forward-looking conversation about where AI, Kubernetes, and cloud-native engineering are heading — from someone building that future at scale.

Stuur ons een bericht.

ACC ICT Specialist in IT-CONTINUÏTEIT
Bedrijfskritische applicaties én data veilig beschikbaar, onafhankelijk van derden, altijd en overal

Support the show

Like and subscribe! It helps out a lot.