
Sign up to save your podcasts
Or


Running multiple Kubernetes clusters on AWS with the cluster autoscaler? Every four months, you face the same grind: upgrading Kubernetes versions, recreating auto scaling groups, and hoping instance type changes stick.
Adhi Sutandi, DevOps Engineer at Beekeeper by LumApps, shares how his team migrated from the cluster autoscaler to Karpenter across eight EKS clusters — and the hard lessons they learned along the way.
In this episode:
Why AWS auto scaling groups are immutable and how that creates upgrade bottlenecks at scale
How the latest AMI tag accidentally turned less critical clusters into chaos engineering environments, dropping SLOs before anyone realized Karpenter was the cause
Why pre-stop sleep hooks solved pod restartability problems that Quarkus's built-in graceful shutdown couldn't
The case for pod disruption budgets over Karpenter annotations when protecting critical workloads during node rotations
How Karpenter's implicit 10% disruption budget caught the team off guard — and the explicit configuration that fixed it
Sponsor
This episode is sponsored by LearnKube — get started on your Kubernetes journey through comprehensive online, in-person or remote training.
More info
Find all the links and info for this episode here: https://ku.bz/XyVfsSQPr
Interested in sponsoring an episode? Learn more.
By KubeFM5
22 ratings
Running multiple Kubernetes clusters on AWS with the cluster autoscaler? Every four months, you face the same grind: upgrading Kubernetes versions, recreating auto scaling groups, and hoping instance type changes stick.
Adhi Sutandi, DevOps Engineer at Beekeeper by LumApps, shares how his team migrated from the cluster autoscaler to Karpenter across eight EKS clusters — and the hard lessons they learned along the way.
In this episode:
Why AWS auto scaling groups are immutable and how that creates upgrade bottlenecks at scale
How the latest AMI tag accidentally turned less critical clusters into chaos engineering environments, dropping SLOs before anyone realized Karpenter was the cause
Why pre-stop sleep hooks solved pod restartability problems that Quarkus's built-in graceful shutdown couldn't
The case for pod disruption budgets over Karpenter annotations when protecting critical workloads during node rotations
How Karpenter's implicit 10% disruption budget caught the team off guard — and the explicit configuration that fixed it
Sponsor
This episode is sponsored by LearnKube — get started on your Kubernetes journey through comprehensive online, in-person or remote training.
More info
Find all the links and info for this episode here: https://ku.bz/XyVfsSQPr
Interested in sponsoring an episode? Learn more.

274 Listeners

288 Listeners

2,009 Listeners

631 Listeners

276 Listeners

153 Listeners

583 Listeners

287 Listeners

44 Listeners

167 Listeners

179 Listeners

206 Listeners

62 Listeners

98 Listeners

68 Listeners