Kubernetes Autoscaling

Introduction

Course Introduction

Kubernetes has emerged as the de facto platform for container orchestration—often called “the Linux of the cloud.” With the AI revolution accelerating, many services you use (e.g., PoE, Anthropic Cloud, ChatGPT) run on Kubernetes combined with specialized orchestrators. As AI adoption grows, so does the demand for Kubernetes expertise.

According to a 2022 survey by Indeed, Kubernetes led job-search growth with a 173% year-over-year increase:

The image is an Indeed survey report showing the year-over-year growth in job search interest for various tech skills, with Kubernetes leading at 173% growth. Other skills listed include Magento, Verilog, and Golang.

This upward trend continued through 2023 and into 2024, making it essential to develop your Kubernetes skills to boost your market value.

Welcome to Kubernetes on a Scaling. I’m Michael Forrester, and in this course you’ll master efficient scaling of Kubernetes workloads through hands-on labs. You’ll learn by experimenting, troubleshooting, and building confidence to tackle scaling challenges in production environments.

Table of Contents

Scaling MethodFocusKey Benefits
Manual ScalingBasic kubectl commandsUnderstand core autoscaling primitives
Horizontal Pod Autoscaler (HPA)CPU, memory, custom & external metricsAutomated pod count adjustments
Vertical Pod Autoscaler (VPA)CPU & memory resource recommendationsOptimized resource allocation per pod
Cluster Proportional AutoscalerComponent scaling based on cluster sizeBalanced control-plane scaling
Event-Driven Autoscaling (KEDA)Event triggers (Cron, Redis, Prometheus)Dynamic response to external events

Manual Scaling Fundamentals

We begin with manual scaling to grasp how Kubernetes handles resource changes under the hood. You’ll use basic kubectl commands to adjust and observe workload sizes, gaining a solid foundation before moving on to automated methods.

Automating with HPA

The Horizontal Pod Autoscaler (HPA) automatically adjusts the number of pod replicas based on observed metrics like CPU, memory, and custom metrics. You’ll explore the HPA architecture, configure necessary adapters, and create HPA objects to automate scaling in various scenarios.

Exploring VPA

The Vertical Pod Autoscaler (VPA) recommends or applies resource request and limit adjustments for containers. Through labs, you’ll learn how VPA helps optimize CPU and memory allocation per pod, improving application stability and efficiency.

Cluster Proportional Autoscaler (CPA)

The Cluster Proportional Autoscaler scales control-plane components—such as the API server, scheduler, and controller manager—based on cluster size. You’ll configure CPA policies to ensure your control-plane scales in line with workload demands.

Event-Driven Autoscaling with KEDA

KEDA (Kubernetes Event-Driven Autoscaling) enables scaling based on external events—cron schedules, message queues, Redis lists, Prometheus alerts, and more. In this module, you’ll connect KEDA to various event sources and implement flexible autoscaling strategies.

Join the Community

Engage with fellow learners in our vibrant forum—ask questions, share best practices, and collaborate on scaling strategies. Active participation accelerates your learning and helps others.

Let’s get started on mastering Kubernetes autoscaling techniques. See you in class!

Watch Video

Watch video content