Pod IO Stress Details
Experiment Metadata
Type | Description | Tested K8s Platform |
---|---|---|
Generic | Inject IO stress on the application container | GKE, Packet(Kubeadm), Minikube |
Prerequisites
- Ensure that the Litmus Chaos Operator is running by executing
kubectl get pods
in operator namespace (typically,litmus
). If not, install from here - Ensure that the
pod-io-stress
experiment resource is available in the cluster by executingkubectl get chaosexperiments
in the desired namespace. If not, install from here - Cluster must run docker container runtime
Entry Criteria
- Application pods are healthy on the respective nodes before chaos injection
Exit Criteria
- Application pods are healthy on the respective nodes post chaos injection
Details
- This experiment causes disk stress on the application pod. The experiment aims to verify the resiliency of applications that share this disk resource for ephemeral or persistent storage purposes
Integrations
- Pod IO Stress can be effected using the chaos library:
pumba
Steps to Execute the Chaos Experiment
This Chaos Experiment can be triggered by creating a ChaosEngine resource on the cluster. To understand the values to provide in a ChaosEngine specification, refer Getting Started
Follow the steps in the sections below to create the
chaosServiceAccount
, prepare the ChaosEngine & execute the experiment.
Prepare chaosServiceAccount
Use this sample RBAC manifest to create a chaosServiceAccount
in the desired (app) namespace. This example consists of the minimum necessary role permissions to execute the experiment.
Sample Rbac Manifest
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: pod-io-stress-sa
namespace: default
labels:
name: pod-io-stress-sa
---
apiVersion: rbac.authorization.k8s.io/v1beta1
kind: Role
metadata:
name: pod-io-stress-sa
namespace: default
labels:
name: pod-io-stress-sa
rules:
- apiGroups: ["","litmuschaos.io","batch"]
resources: ["pods","jobs","events","pods/log","pods/exec","chaosengines","chaosexperiments","chaosresults"]
verbs: ["create","list","get","patch","update","delete","deletecollection"]
---
apiVersion: rbac.authorization.k8s.io/v1beta1
kind: RoleBinding
metadata:
name: pod-io-stress-sa
namespace: default
labels:
name: pod-io-stress-sa
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: Role
name: pod-io-stress-sa
subjects:
- kind: ServiceAccount
name: pod-io-stress-sa
namespace: default
Prepare ChaosEngine
- Provide the application info in
spec.appinfo
- Provide the auxiliary applications info (ns & labels) in
spec.auxiliaryAppInfo
- Override the experiment tunables if desired in
experiments.spec.components.env
- To understand the values to provided in a ChaosEngine specification, refer ChaosEngine Concepts
Supported Experiment Tunables
Variables | Description | Type | Notes | |
---|---|---|---|---|
FILESYSTEM_UTILIZATION_PERCENTAGE | Specify the size as percentage of free space on the file system | Optional | Default to 10% | |
FILESYSTEM_UTILIZATION_BYTES | Specify the size in GigaBytes(GB). FILESYSTEM_UTILIZATION_PERCENTAGE & FILESYSTEM_UTILIZATION_BYTES are mutually exclusive. If both are provided, FILESYSTEM_UTILIZATION_PERCENTAGE is prioritized. |
Optional | ||
NUMBER_OF_WORKERS | It is the number of IO workers involved in IO disk stress | Optional | Default to 4 | |
TARGET_POD | Name of the application pod subjected to IO stress chaos | Optional | If not provided it will select from the app label provided | |
TOTAL_CHAOS_DURATION | The time duration for chaos (seconds) | Optional | Default to 120s | |
LIB | The chaos lib used to inject the chaos | Optional | Default to pumba |
|
LIB_IMAGE | Image used to run the stress command | Optional | Default to gaiaadm/pumba |
|
TARGET_POD | Name of the application pod subjected to pod io stress chaos | Optional | If not provided it will select from the app label provided | |
PODS_AFFECTED_PERC | The Percentage of total pods to target | Optional | Default to 0% (corresponds to 1 replica) | |
RAMP_TIME | Period to wait before and after injection of chaos in sec | Optional | ||
INSTANCE_ID | A user-defined string that holds metadata/info about current run/instance of chaos. Ex: 04-05-2020-9-00. This string is appended as suffix in the chaosresult CR name. | Optional | Ensure that the overall length of the chaosresult CR is still < 64 characters |
Sample ChaosEngine Manifest
apiVersion: litmuschaos.io/v1alpha1
kind: ChaosEngine
metadata:
name: nginx-chaos
namespace: default
spec:
# It can be true/false
annotationCheck: 'true'
# It can be active/stop
engineState: 'active'
#ex. values: ns1:name=percona,ns2:run=nginx
auxiliaryAppInfo: ''
appinfo:
appns: 'default'
applabel: 'app=nginx'
appkind: 'deployment'
chaosServiceAccount: pod-io-stress-sa
monitoring: false
# It can be delete/retain
jobCleanUpPolicy: 'delete'
experiments:
- name: pod-io-stress
spec:
components:
env:
# set chaos duration (in sec) as desired
- name: TOTAL_CHAOS_DURATION
value: '120'
## specify the size as percentage of free space on the file system
- name: FILESYSTEM_UTILIZATION_PERCENTAGE
value: '10'
- name: TARGET_POD
value: ''
Create the ChaosEngine Resource
Create the ChaosEngine manifest prepared in the previous step to trigger the Chaos.
kubectl apply -f chaosengine.yml
If the chaos experiment is not executed, refer to the troubleshooting section to identify the root cause and fix the issues.
Watch Chaos progress
View the status of the pods as they are subjected to IO disk stress.
watch -n 1 kubectl get pods -n <application-namespace>
Monitor the capacity filled up on the host filesystem
watch du -h
Check Chaos Experiment Result
Check whether the application stack is resilient to IO stress on the app replica, once the experiment (job) is completed. The ChaosResult resource name is derived like this:
<ChaosEngine-Name>-<ChaosExperiment-Name>
.kubectl describe chaosresult nginx-chaos-pod-io-stress -n <application-namespace>
Pod IO Stress Experiment Demo
- The Demo Video will be Added soon.