Kubernetes memory scalability

Description

Verify that your Kubernetes service scales as expected when memory is limited. Gremlin increases memory utilization in three stages—50%, 75%, and 90%—to validate Kubernetes memory management.

What this Scenario does

This Scenario progressively consumes memory within your Kubernetes pods in three stages: 50%, 75%, and 90%. This simulates memory pressure from traffic spikes, memory leaks, or under-provisioned pod resource limits.

Why run this Scenario?

  • Identify memory leaks or poor garbage collection within pod workloads before they trigger OOMKiller.
  • Test how Kubernetes memory resource limits and requests interact with the OOMKiller under pressure.
  • Verify that Kubernetes pod autoscaling (HPA or VPA) responds to memory pressure as configured.
  • Optimize pod memory resource limits to balance cost, performance, and cluster stability.

Expected outcome

If memory usage spikes to 90% within a Kubernetes pod, the application continues to function without performance degradation for at least 5 minutes.

Target
Kubernetes
Experiments
Memory
Memory
Preview
Runtime:  
20 minutes