Kubernetes 1.35: In-Place Pod Resource Updates and AI Model Image Volumes

Kubernetes 1.35, released in January 2026 and now supported on Amazon EKS and EKS Distro, marks a significant milestone in container orchestration—particularly for AI/ML workloads. This release introduces In-Place Pod Resource Updates, allowing you to resize CPU and memory without restarting pods, and Image Volumes, a game-changer for delivering large AI models using OCI container […]

Read more →

Mastering GKE: A Deep Dive into Google Kubernetes Engine for Production Workloads

Introduction: Google Kubernetes Engine represents the gold standard for managed Kubernetes, built on the same infrastructure that runs Google’s own containerized workloads at massive scale. This deep dive explores GKE’s enterprise capabilities—from Autopilot mode that eliminates node management to advanced features like workload identity, binary authorization, and multi-cluster service mesh. After deploying production Kubernetes clusters […]

Read more →

Running LLMs on Kubernetes: Production Deployment Guide

Deploying LLMs on Kubernetes requires careful planning. After deploying 25+ LLM models on Kubernetes, I’ve learned what works. Here’s the complete guide to running LLMs on Kubernetes in production. Figure 1: Kubernetes LLM Architecture Why Kubernetes for LLMs Kubernetes offers significant advantages for LLM deployment: Scalability: Auto-scale based on demand Resource management: Efficient GPU and […]

Read more →