push (chore)cleanup stale chaos experiments from chaoshub (#585)

Signed-off-by: ksatchit <karthik.s@harness.io>
This commit is contained in:
Karthik Satchitanand
2023-02-15 17:22:53 +05:30
committed by GitHub
parent 6ae06914d9
commit 89f495c5ba
171 changed files with 0 additions and 7201 deletions

View File

@@ -1,43 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
createdAt: 2019-12-30T10:28:08Z
name: coredns-pod-delete
version: 0.1.7
annotations:
categories: CoreDNS
vendor: CNCF
support: https://slack.cncf.io/
spec:
displayName: coredns-pod-delete
categoryDescription: |
Pod delete contains chaos to disrupt state of coredns resources. Experiments can inject random pod delete failures against specified application.
- Causes (forced/graceful) pod failure of coredns replicas of an application deployment.
- Tests of successful resolution of app services during chaos injection
keywords:
- Kubernetes
- K8S
- CoreDNS
- DNS
- Pod
platforms:
- GKE
- Minikube
maturity: alpha
maintainers:
- name: Raj Babu Das
email: raj.das@mayadata.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
labels:
app.kubernetes.io/component: chartserviceversion
app.kubernetes.io/version: latest
links:
- name: Source Code
url: https://github.com/litmuschaos/litmus-ansible/tree/master/experiments/coredns/pod_delete
- name: Documentation
url: https://v1-docs.litmuschaos.io/docs/coredns-pod-delete/
- name: Video
url: https://www.youtube.com/watch?v=pwo5idKW7q8
chaosexpcrdlink: https://raw.githubusercontent.com/litmuchaos/chaos-charts/master/charts/coredns/coredns-pod-delete/experiment.yaml

View File

@@ -1,30 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
kind: ChaosEngine
metadata:
name: engine-coredns
namespace: kube-system
spec:
appinfo:
appns: 'kube-system'
applabel: 'k8s-app=kube-dns'
appkind: 'deployment'
# It can be active/stop
engineState: 'active'
#ex. values: ns1:name=percona,ns2:run=nginx
auxiliaryAppInfo: ''
chaosServiceAccount: coredns-pod-delete-sa
experiments:
- name: coredns-pod-delete
spec:
components:
env:
# set chaos duration (in sec) as desired
- name: TOTAL_CHAOS_DURATION
value: '30'
# set chaos interval (in sec) as desired
- name: CHAOS_INTERVAL
value: '10'
- name: CHAOS_NAMESPACE
value: 'kube-system'

View File

@@ -1,72 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
description:
message: |
Deletes coredns pod in kube-system namespace
kind: ChaosExperiment
metadata:
name: coredns-pod-delete
labels:
name: coredns-pod-delete
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: chaosexperiment
app.kubernetes.io/version: latest
spec:
definition:
permissions:
- apiGroups:
- ""
- "batch"
- "litmuschaos.io"
resources:
- "services"
- "jobs"
- "pods"
- "pods/log"
- "events"
- "chaosengines"
- "chaosexperiments"
- "chaosresults"
verbs:
- "get"
- "list"
- "patch"
- "create"
- "update"
- "delete"
image: "litmuschaos/ansible-runner:latest"
imagePullPolicy: Always
args:
- -c
- ansible-playbook ./experiments/coredns/pod_delete/pod_delete_ansible_logic.yml -i /etc/ansible/hosts -vv; exit 0
command:
- /bin/bash
env:
- name: ANSIBLE_STDOUT_CALLBACK
value: 'default'
- name: TOTAL_CHAOS_DURATION
value: '15'
# provide the kill count
- name: KILL_COUNT
value: ''
- name: FORCE
value: 'true'
- name: CHAOS_INTERVAL
value: '5'
#supported libs are litmus and powerfulseal
- name: LIB
value: 'litmus'
- name: LIB_IMAGE
value: 'litmuschaos/pod-delete-helper:latest'
labels:
name: coredns-pod-delete
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: experiment-job
app.kubernetes.io/version: latest

View File

@@ -1,36 +0,0 @@
apiVersion: v1
kind: ServiceAccount
metadata:
name: coredns-pod-delete-sa
namespace: kube-system
labels:
name: coredns-pod-delete-sa
app.kubernetes.io/part-of: litmus
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: coredns-pod-delete-sa
labels:
name: coredns-pod-delete-sa
app.kubernetes.io/part-of: litmus
rules:
- apiGroups: ["","litmuschaos.io","batch"]
resources: ["services", "pods","jobs","events","pods/log","chaosengines","chaosexperiments","chaosresults"]
verbs: ["create","list","get","patch","update","delete"]
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: coredns-pod-delete-sa
labels:
name: coredns-pod-delete-sa
app.kubernetes.io/part-of: litmus
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: coredns-pod-delete-sa
subjects:
- kind: ServiceAccount
name: coredns-pod-delete-sa
namespace: kube-system

View File

@@ -1,41 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
createdAt: 2019-12-30T10:28:08Z
name: coredns
version: 0.1.4
annotations:
categories: Kubernetes
chartDescription: Injects coredns chaos
spec:
displayName: CoreDNS chaos
categoryDescription: >
CoreDNS is an open-source system for DNS management of containerized applications.
It helps to resolve services to IP address. It will install all the experiments which can be used to inject chaos into containerized applications.
experiments:
- coredns-pod-delete
keywords:
- Kubernetes
- K8S
- Pod
- CoreDNS
- DNS
maintainers:
- name: Raj Babu Das
email: raj.das@mayadata.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
links:
- name: CoreDNS Website
url: https://coredns.io
- name: Source Code
url: https://github.com/coredns/coredns
- name: CoreDNS Slack
url: https://slack.cncf.io/
- name: Documentation
url: https://coredns.io/manual/toc/
icon:
- url: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/coredns/icons/coredns-pod-delete.png
mediatype: image/png
chaosexpcrdlink: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/coredns/experiments.yaml

View File

@@ -1,5 +0,0 @@
packageName: coredns
experiments:
- name: coredns-pod-delete
CSV: coredns-pod-delete.chartserviceversion.yaml
desc: "coredns pod delete experiment"

View File

@@ -1,74 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
description:
message: |
Deletes coredns pod in kube-system namespace
kind: ChaosExperiment
metadata:
name: coredns-pod-delete
labels:
name: coredns-pod-delete
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: chaosexperiment
app.kubernetes.io/version: latest
spec:
definition:
permissions:
- apiGroups:
- ""
- "batch"
- "litmuschaos.io"
resources:
- "services"
- "jobs"
- "pods"
- "pods/log"
- "events"
- "chaosengines"
- "chaosexperiments"
- "chaosresults"
verbs:
- "get"
- "list"
- "patch"
- "create"
- "update"
- "delete"
image: "litmuschaos/ansible-runner:latest"
imagePullPolicy: Always
args:
- -c
- ansible-playbook ./experiments/coredns/pod_delete/pod_delete_ansible_logic.yml -i /etc/ansible/hosts -vv; exit 0
command:
- /bin/bash
env:
- name: ANSIBLE_STDOUT_CALLBACK
value: 'default'
- name: TOTAL_CHAOS_DURATION
value: '15'
# provide the kill count
- name: KILL_COUNT
value: ''
- name: FORCE
value: 'true'
- name: CHAOS_INTERVAL
value: '5'
#supported libs are litmus and powerfulseal
- name: LIB
value: 'litmus'
- name: LIB_IMAGE
value: 'litmuschaos/pod-delete-helper:latest'
labels:
name: coredns-pod-delete
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: experiment-job
app.kubernetes.io/version: latest
---

Binary file not shown.

Before

Width:  |  Height:  |  Size: 20 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 20 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 2.2 KiB

View File

@@ -1,32 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
createdAt: 2020-03-17T10:28:08Z
name: longhorn
version: 0.0.2
annotations:
categories: longhorn
chartDescription: Injects chaos on longhorn components
spec:
displayName: Longhorn Chaos
categoryDescription: >
Longhorn is a distributed block storage system for Kubernetes
experiments:
keywords:
- Kubernetes
- Longhorn
maintainers:
- name: ksatchit
email: karthik.s@mayadata.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
links:
- name: Longhorn
url: https://rancher.com/project-longhorn-now-available-kubernetes/
- name: Documentation
url: https://docs.rancher.cn/longhorn/#current-status
icon:
- url:
mediatype:
chaosexpcrdlink: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/longhorn/experiments.yaml

View File

@@ -1,2 +0,0 @@
packageName: longhorn
experiments:

Binary file not shown.

Before

Width:  |  Height:  |  Size: 40 KiB

View File

@@ -1,30 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
createdAt: 2020-03-17T10:28:08Z
name: mysql
version: 0.0.2
annotations:
categories: mysql
chartDescription: Injects chaos on mysql components
spec:
displayName: MySQL Chaos
categoryDescription: >
MySQL is an open-source relational database management system.
experiments:
keywords:
- Kubernetes
- MySQL
maintainers:
- name: atulabhi
email: atul.abhishek@openebs.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
links:
- name: mysql Website
url: https://www.mysql.com/
icon:
- url:
mediatype:
chaosexpcrdlink: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/mysql/experiments.yaml

View File

@@ -1,2 +0,0 @@
packageName: mysql
experiments:

File diff suppressed because it is too large Load Diff

Binary file not shown.

Before

Width:  |  Height:  |  Size: 12 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 12 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 12 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 12 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 12 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 12 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 1.4 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 3.1 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 1.2 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 2.7 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 2.6 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 12 KiB

View File

@@ -1,20 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
kind: ChaosEngine
metadata:
name: control-plane-chaos
namespace: openebs
spec:
# It can be active/stop
engineState: 'active'
appinfo:
appns: 'openebs'
applabel: 'name=maya-apiserver'
appkind: 'deployment'
chaosServiceAccount: control-plane-sa
experiments:
- name: openebs-control-plane-chaos
spec:
components:
env:
- name: FORCE
value: ''

View File

@@ -1,79 +0,0 @@
---
apiVersion: litmuschaos.io/v1alpha1
description:
message: |
Kill all openebs control plane pod and check if gets scheduled again
kind: ChaosExperiment
metadata:
labels:
litmuschaos.io/name: openebs
name: openebs-control-plane-chaos
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: chaosexperiment
app.kubernetes.io/version: latest
name: openebs-control-plane-chaos
spec:
definition:
scope: Namespaced
permissions:
- apiGroups:
- ""
- "litmuschaos.io"
- "batch"
- "apps"
resources:
- "pods"
- "pods/log"
- "deployments"
- "events"
- "jobs"
- "configmaps"
- "chaosengines"
- "chaosexperiments"
- "chaosresults"
verbs:
- "create"
- "list"
- "get"
- "patch"
- "update"
- "delete"
- apiGroups:
- ""
resources:
- "nodes"
verbs:
- "get"
- "list"
image: "litmuschaos/ansible-runner:latest"
imagePullPolicy: Always
args:
- -c
- ansible-playbook ./experiments/openebs/openebs-control-plane-chaos/openebs_control_plane_chaos_ansible_logic.yml -i /etc/ansible/hosts -vv; exit 0
command:
- /bin/bash
env:
- name: ANSIBLE_STDOUT_CALLBACK
value: 'default'
- name: OPENEBS_NAMESPACE
value: 'openebs'
## Period to wait before injection of chaos
- name: RAMP_TIME
value: ''
- name: FORCE
value: ''
## env var that describes the library used to execute the chaos
## default: litmus. Supported values: litmus, powerfulseal
- name: LIB
value: 'litmus'
labels:
name: openebs-control-plane-chaos
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: experiment-job
app.kubernetes.io/version: latest

View File

@@ -1,42 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
name: openebs-control-plane-chaos
version: 0.1.4
annotations:
categories: OpenEBS
vendor: CNCF
support: https://slack.openebs.io/
spec:
displayName: openebs-control-plane-chaos
categoryDescription: >
Kill all openebs control plane pods and check if gets scheduled again.
keywords:
- Kubernetes
- K8S
- OpenEBS
platforms:
- GKE
- EKS
maturity: alpha
chaosType: infra
maintainers:
- name: Raj Babu Das
email: raj.das@mayadata.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
labels:
app.kubernetes.io/component: chartserviceversion
app.kubernetes.io/version: latest
links:
- name: Source Code
url: https://github.com/litmuschaos/litmus-ansible/tree/master/experiments/openebs/openebs-control-plane-chaos
- name: Documentation
url: https://v1-docs.litmuschaos.io/docs/openebs-control-plane-chaos/
- name: Video
url:
icon:
- url: ""
mediatype: ""
chaosexpcrdlink: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/openebs/openebs-control-plane-chaos/experiment.yaml

View File

@@ -1,42 +0,0 @@
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: control-plane-sa
namespace: openebs
labels:
name: control-plane-sa
app.kubernetes.io/part-of: litmus
---
apiVersion: rbac.authorization.k8s.io/v1
kind: Role
metadata:
name: control-plane-sa
namespace: openebs
labels:
name: control-plane-sa
app.kubernetes.io/part-of: litmus
rules:
- apiGroups: ["","litmuschaos.io","batch","apps"]
resources: ["pods","deployments","pods/log","events","jobs","configmaps","chaosengines","chaosexperiments","chaosresults"]
verbs: ["create","list","get","patch","update","delete"]
- apiGroups: [""]
resources: ["nodes"]
verbs: ["get","list"]
---
apiVersion: rbac.authorization.k8s.io/v1
kind: RoleBinding
metadata:
name: control-plane-sa
namespace: openebs
labels:
name: control-plane-sa
app.kubernetes.io/part-of: litmus
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: Role
name: control-plane-sa
subjects:
- kind: ServiceAccount
name: control-plane-sa
namespace: openebs

View File

@@ -1,40 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
kind: ChaosEngine
metadata:
name: nfs-chaos
namespace: default
spec:
# It can be active/stop
engineState: 'active'
appinfo:
appns: 'minio'
applabel: 'app=minio'
appkind: 'deployment'
chaosServiceAccount: nfs-chaos-sa
experiments:
- name: openebs-nfs-provisioner-kill
spec:
components:
env:
# provide the total chaos duration
- name: TOTAL_CHAOS_DURATION
value: '20'
- name: NFS_PROVISIONER_NAMESPACE
value: 'app-nfs-ns'
- name: NFS_PROVISIONER_LABEL
value: 'app=nfs'
- name: NFS_PVC
value: 'nfs-pvc-claim'
- name: NFS_SVC
value: 'nfs-provisioner'
- name: TARGET_CONTAINER
value: 'nfs-provisioner'
# EXTERNAL_APP_CHECK can be true/false
- name: EXTERNAL_APP_CHECK
value: 'true'

View File

@@ -1,100 +0,0 @@
---
apiVersion: litmuschaos.io/v1alpha1
description:
message: |
Kill the OpenEBS NFS provisioner container and check if pods consuming the NFS PVs continue to be available and volumes are writable (RWM mode)
kind: ChaosExperiment
metadata:
labels:
litmuschaos.io/name: openebs
name: openebs-nfs-provisioner-kill
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: chaosexperiment
app.kubernetes.io/version: latest
name: openebs-nfs-provisioner-kill
spec:
definition:
scope: Cluster
permissions:
- apiGroups:
- ""
- "apps"
- "litmuschaos.io"
- "batch"
- "extensions"
- "storage.k8s.io"
resources:
- "pods"
- "pods/exec"
- "pods/log"
- "deployments"
- "events"
- "jobs"
- "configmaps"
- "services"
- "persistentvolumeclaims"
- "storageclasses"
- "persistentvolumes"
- "chaosexperiments"
- "chaosresults"
- "chaosengines"
verbs:
- "create"
- "list"
- "get"
- "patch"
- "update"
- "delete"
- apiGroups:
- ""
resources:
- "nodes"
verbs:
- "get"
- "list"
image: "litmuschaos/ansible-runner:latest"
imagePullPolicy: Always
args:
- -c
- ansible-playbook ./experiments/openebs/openebs-nfs-provisioner-kill/openebs_nfs_provisioner_kill_ansible_logic.yml -i /etc/ansible/hosts -vv; exit 0
command:
- /bin/bash
env:
- name: ANSIBLE_STDOUT_CALLBACK
value: 'default'
# NFS default container
- name: TARGET_CONTAINER
value: 'nfs-provisioner'
# Period to wait before injection of chaos in sec
- name: RAMP_TIME
value: ''
# It supports pumba and containerd
- name: LIB
value: 'pumba'
# LIB_IMAGE can be - gaiaadm/pumba:0.6.5, gprasath/crictl:ci
# For pumba image use: gaiaadm/pumba:0.6.5
# For containerd image use: gprasath/crictl:ci
- name: LIB_IMAGE
value: 'gaiaadm/pumba:0.6.5'
# provide the chaos interval
- name: CHAOS_INTERVAL
value: '10'
# provide the total chaos duration
- name: TOTAL_CHAOS_DURATION
value: '20'
labels:
name: openebs-nfs-provisioner-kill
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: experiment-job
app.kubernetes.io/version: latest
configmaps:
- name: openebs-nfs-provisioner-kill
mountPath: /mnt/

View File

@@ -1,42 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
name: openebs-nfs-provisioner-kill
version: 0.1.3
annotations:
categories: OpenEBS
vendor: CNCF
support: https://slack.openebs.io/
spec:
displayName: openebs-nfs-kill-chaos
categoryDescription: >
Kill the OpenEBS NFS provisioner container and check if pods consuming the NFS PVs continue to be available and volumes are writable (RWM mode)
keywords:
- Kubernetes
- K8S
- OpenEBS
- NFS
platforms:
- GKE
maturity: alpha
chaosType: infra
maintainers:
- name: Raj Babu Das
email: raj.das@mayadata.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
labels:
app.kubernetes.io/component: chartserviceversion
app.kubernetes.io/version: latest
links:
- name: Source Code
url: https://github.com/litmuschaos/litmus-ansible/tree/master/experiments/openebs/openebs-nfs-provisioner-kill
- name: Documentation
url: https://v1-docs.litmuschaos.io/docs/openebs-nfs-provisioner-kill/
- name: Video
url:
icon:
- url: ""
mediatype: ""
chaosexpcrdlink: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/openebs/openebs-nfs-provisioner-kill/experiment.yaml

View File

@@ -1,37 +0,0 @@
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: nfs-chaos-sa
namespace: default
labels:
name: nfs-chaos-sa
app.kubernetes.io/part-of: litmus
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: nfs-chaos-sa
labels:
name: nfs-chaos-sa
app.kubernetes.io/part-of: litmus
rules:
- apiGroups: ["","apps","litmuschaos.io","batch","extensions","storage.k8s.io"]
resources: ["pods","pods/exec","pods/log", "deployments","events","jobs","configmaps","services","persistentvolumeclaims","storageclasses","persistentvolumes","chaosexperiments","chaosresults","chaosengines"]
verbs: ["create","list","get","patch","update","delete"]
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: nfs-chaos-sa
labels:
name: nfs-chaos-sa
app.kubernetes.io/part-of: litmus
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: nfs-chaos-sa
subjects:
- kind: ServiceAccount
name: nfs-chaos-sa
namespace: default

View File

@@ -1,29 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
kind: ChaosEngine
metadata:
name: pool-chaos
namespace: default
spec:
# It can be active/stop
engineState: 'active'
#ex. values: ns1:name=percona,ns2:run=nginx
auxiliaryAppInfo: ''
appinfo:
appns: 'default'
applabel: 'app=nginx'
appkind: 'deployment'
chaosServiceAccount: pool-container-failure-sa
experiments:
- name: openebs-pool-container-failure
spec:
components:
env:
# provide the total chaos duration
- name: TOTAL_CHAOS_DURATION
value: '20'
- name: APP_PVC
value: 'demo-nginx-claim'
- name: DEPLOY_TYPE
value: 'deployment'

View File

@@ -1,97 +0,0 @@
---
apiVersion: litmuschaos.io/v1alpha1
description:
message: |
Kill the pool container and check if gets scheduled again
kind: ChaosExperiment
metadata:
labels:
litmuschaos.io/name: openebs
name: openebs-pool-container-failure
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: chaosexperiment
app.kubernetes.io/version: latest
name: openebs-pool-container-failure
spec:
definition:
scope: Cluster
permissions:
- apiGroups:
- ""
- "extensions"
- "apps"
- "batch"
- "litmuschaos.io"
- "openebs.io"
- "storage.k8s.io"
resources:
- "replicasets"
- "events"
- "jobs"
- "pods"
- "pods/log"
- "pods/exec"
- "configmaps"
- "secrets"
- "persistentvolumeclaims"
- "cstorvolumereplicas"
- "chaosengines"
- "chaosexperiments"
- "chaosresults"
verbs:
- "create"
- "get"
- "delete"
- "list"
- "patch"
- "update"
image: "litmuschaos/ansible-runner:latest"
imagePullPolicy: Always
args:
- -c
- ansible-playbook ./experiments/openebs/openebs-pool-container-failure/openebs_pool_container_failure_ansible_logic.yml -i /etc/ansible/hosts -vv; exit 0
command:
- /bin/bash
env:
- name: ANSIBLE_STDOUT_CALLBACK
value: 'default'
- name: OPENEBS_NS
value: 'openebs'
- name: APP_PVC
value: ''
- name: LIVENESS_APP_LABEL
value: ''
# only pumba supported
# For pumba image use : gaiaadm/pumba:0.6.5
- name: LIB_IMAGE
value: 'gaiaadm/pumba:0.6.5'
- name: LIVENESS_APP_NAMESPACE
value: ''
# provide the chaos interval
- name: CHAOS_INTERVAL
value: '10'
# provide the total chaos duration
- name: TOTAL_CHAOS_DURATION
value: '20'
- name: DATA_PERSISTENCE
value: ''
- name: CHAOS_ITERATIONS
value: '2'
labels:
name: openebs-pool-container-failure
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: experiment-job
app.kubernetes.io/version: latest
#configmaps:
#- name: openebs-pool-container-failure
# mountPath: /mnt

View File

@@ -1,45 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
createdAt: 2019-11-13T10:28:08Z
name: openebs-pool-container-failure
version: 0.1.13
annotations:
categories: OpenEBS
vendor: CNCF
support: https://slack.openebs.io/
spec:
displayName: openebs-pool-container-failure
categoryDescription: >
Kill the pool container and check if gets scheduled again.
keywords:
- Kubernetes
- K8S
- Storage Pool
- OpenEBS
platforms:
- GKE
- EKS
- OpenShift
maturity: alpha
chaosType: infra
maintainers:
- name: shubham chaudhary
email: shubham.chaudhary@mayadata.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
labels:
app.kubernetes.io/component: chartserviceversion
app.kubernetes.io/version: latest
links:
- name: Source Code
url: https://github.com/litmuschaos/litmus-ansible/tree/master/experiments/openebs/openebs-pool-container-failure
- name: Documentation
url: https://v1-docs.litmuschaos.io/docs/openebs-pool-container-failure/
- name: Video
url:
icon:
- url: ""
mediatype: ""
chaosexpcrdlink: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/openebs/openebs-pool-container-failure/experiment.yaml

View File

@@ -1,38 +0,0 @@
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: pool-container-failure-sa
namespace: default
labels:
name: pool-container-failure-sa
app.kubernetes.io/part-of: litmus
---
# Source: openebs/templates/clusterrole.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: pool-container-failure-sa
labels:
name: pool-container-failure-sa
app.kubernetes.io/part-of: litmus
rules:
- apiGroups: ["","apps","litmuschaos.io","batch","extensions","storage.k8s.io","openebs.io"]
resources: ["pods","jobs","events","pods/log","replicasets","pods/exec","configmaps","secrets","persistentvolumeclaims","cstorvolumereplicas","chaosexperiments","chaosresults","chaosengines"]
verbs: ["create","list","get","patch","update","delete"]
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: pool-container-failure-sa
labels:
name: pool-container-failure-sa
app.kubernetes.io/part-of: litmus
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: pool-container-failure-sa
subjects:
- kind: ServiceAccount
name: pool-container-failure-sa
namespace: default

View File

@@ -1,51 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
kind: ChaosEngine
metadata:
name: pool-chaos
namespace: default
spec:
# It can be active/stop
engineState: 'active'
#ex. values: ns1:name=percona,ns2:run=busybox
auxiliaryAppInfo: ''
appinfo:
appns: 'default'
applabel: 'app=nginx'
appkind: 'deployment'
chaosServiceAccount: pool-disk-loss-sa
experiments:
- name: openebs-pool-disk-loss
spec:
components:
env:
# provide the total chaos duration
- name: TOTAL_CHAOS_DURATION
value: '60'
- name: APP_PVC
value: 'demo-nginx-claim'
# GKE and AWS supported
- name: CLOUD_PLATFORM
value: 'GKE'
# Enter the project id for gcp only
- name: PROJECT_ID
value: 'litmus-demo-123'
# Enter the node name
- name: NODE_NAME
value: 'demo-node-123'
# Enter the disk name
- name: DISK_NAME
value: 'demo-disk-123 '
# Enter the device name
- name: DEVICE_NAME
value: '/dev/sdb'
# Enter the zone name
- name: ZONE_NAME
value: 'us-central1-a'

View File

@@ -1,111 +0,0 @@
---
apiVersion: litmuschaos.io/v1alpha1
description:
message: |
OpenEBS Pool Disk Loss contains chaos to disrupt state of infra resources. Experiments can inject disk loss against openEBS pool.
kind: ChaosExperiment
metadata:
labels:
litmuschaos.io/name: openebs
name: openebs-pool-disk-loss
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: chaosexperiment
app.kubernetes.io/version: latest
name: openebs-pool-disk-loss
spec:
definition:
scope: Cluster
permissions:
- apiGroups:
- ""
- "apps"
- "litmuschaos.io"
- "batch"
- "extensions"
- "storage.k8s.io"
- "openebs.io"
resources:
- "pods"
- "jobs"
- "pods/log"
- "events"
- "pods/exec"
- "cstorpools"
- "configmaps"
- "secrets"
- "storageclasses"
- "persistentvolumes"
- "persistentvolumeclaims"
- "cstorvolumereplicas"
- "chaosexperiments"
- "chaosresults"
- "chaosengines"
verbs:
- "create"
- "list"
- "get"
- "patch"
- "update"
- "delete"
image: "litmuschaos/ansible-runner:latest"
imagePullPolicy: Always
args:
- -c
- ansible-playbook ./experiments/openebs/openebs-pool-disk-loss/openebs_pool_disk_loss_ansible_logic.yml -i /etc/ansible/hosts -vv; exit 0
command:
- /bin/bash
env:
- name: ANSIBLE_STDOUT_CALLBACK
value: 'default'
- name: TOTAL_CHAOS_DURATION
value: '60'
# Period to wait before injection of chaos in sec
- name: RAMP_TIME
value: ''
# GKE and AWS supported
- name: CLOUD_PLATFORM
value: 'GKE'
- name: PROJECT_ID
value: ''
- name: NODE_NAME
value: ''
- name: DISK_NAME
value: ''
- name: ZONE_NAME
value: ''
- name: DEVICE_NAME
value: ''
- name: LIVENESS_APP_LABEL
value: ''
- name: LIVENESS_APP_NAMESPACE
value: ''
- name: DATA_PERSISTENCE
value: ''
- name: OPENEBS_NAMESPACE
value: 'openebs'
labels:
name: openebs-pool-disk-loss
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: experiment-job
app.kubernetes.io/version: latest
configmaps:
- name: openebs-pool-disk-loss
mountPath: /mnt/
secrets:
- name: cloud-secret
mountPath: /tmp/

View File

@@ -1,45 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
createdAt: 2020-03-13T10:28:08Z
name: openebs-pool-disk-loss
version: 0.1.3
annotations:
categories: OpenEBS
vendor: CNCF
support: https://slack.openebs.io/
spec:
displayName: openebs-pool-disk-loss
categoryDescription: >
OpenEBS Pool Disk Loss contains chaos to disrupt state of infra resources. Experiments can inject disk loss against openEBS pool.
- Causes (forced/graceful) disk loss from the node disk pool.
- Tests deployment sanity (replica availability & uninterrupted service) and recovery workflows of the application pod.
keywords:
- Kubernetes
- K8S
- Storage Pool
- OpenEBS
platforms:
- GKE
maturity: alpha
chaosType: infra
maintainers:
- name: Raj Babu Das
email: raj.das@mayadata.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
labels:
app.kubernetes.io/component: chartserviceversion
app.kubernetes.io/version: latest
links:
- name: Source Code
url: https://github.com/litmuschaos/litmus-ansible/tree/master/experiments/openebs/openebs-pool-disk-loss
- name: Documentation
url: https://v1-docs.litmuschaos.io/docs/openebs-pool-disk-loss/
- name: Video
url:
icon:
- url: ""
mediatype: ""
chaosexpcrdlink: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/openebs/openebs-pool-disk-loss/experiment.yaml

View File

@@ -1,38 +0,0 @@
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: pool-disk-loss-sa
namespace: default
labels:
name: pool-disk-loss-sa
app.kubernetes.io/part-of: litmus
---
# Source: openebs/templates/clusterrole.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: pool-disk-loss-sa
labels:
name: pool-disk-loss-sa
app.kubernetes.io/part-of: litmus
rules:
- apiGroups: ["","apps","litmuschaos.io","batch","extensions","storage.k8s.io","openebs.io"]
resources: ["pods", "pods/log", "jobs", "events", "pods/exec", "cstorpools", "configmaps", "secrets", "storageclasses", "persistentvolumes", "persistentvolumeclaims", "cstorvolumereplicas", "chaosexperiments", "chaosresults", "chaosengines"]
verbs: ["create","list","get","patch","update","delete"]
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: pool-disk-loss-sa
labels:
name: pool-disk-loss-sa
app.kubernetes.io/part-of: litmus
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: pool-disk-loss-sa
subjects:
- kind: ServiceAccount
name: pool-disk-loss-sa
namespace: default

View File

@@ -1,25 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
kind: ChaosEngine
metadata:
name: pool-chaos
namespace: default
spec:
appinfo:
appns: 'default'
applabel: 'app=nginx'
appkind: 'deployment'
chaosServiceAccount: pool-network-delay-sa
experiments:
- name: openebs-pool-network-delay
spec:
components:
env:
- name: TOTAL_CHAOS_DURATION
value: '60' # in seconds
- name: APP_PVC
value: 'demo-nginx-claim'
# in milliseconds
- name: NETWORK_DELAY
value: '60000'

View File

@@ -1,93 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
description:
message: |
Network delay to pool pod belonging to a OpenEBS PVC
This experiment is using pumba lib for network chaos
kind: ChaosExperiment
metadata:
labels:
litmuschaos.io/name: openebs
name: openebs-pool-network-delay
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: chaosexperiment
app.kubernetes.io/version: latest
name: openebs-pool-network-delay
spec:
definition:
scope: Cluster
permissions:
- apiGroups:
- ""
- "apps"
- "litmuschaos.io"
- "batch"
- "extensions"
- "storage.k8s.io"
- "openebs.io"
resources:
- "pods"
- "pods/exec"
- "jobs"
- "pods/log"
- "events"
- "configmaps"
- "services"
- "persistentvolumeclaims"
- "storageclasses"
- "persistentvolumes"
- "chaosengines"
- "chaosexperiments"
- "chaosresults"
- "cstorpools"
- "cstorvolumereplicas"
- "replicasets"
verbs:
- "create"
- "get"
- "list"
- "patch"
- "update"
- "delete"
image: "litmuschaos/ansible-runner:latest"
imagePullPolicy: Always
args:
- -c
- ansible-playbook ./experiments/openebs/openebs-pool-network-delay/openebs_pool_network_delay_ansible_logic.yml -i /etc/ansible/hosts -vv; exit 0
command:
- /bin/bash
env:
- name: ANSIBLE_STDOUT_CALLBACK
value: 'default'
- name: OPENEBS_NAMESPACE
value: 'openebs'
# only pumba supported
# For pumba image use : gaiaadm/pumba:0.6.5
- name: LIB_IMAGE
value: 'gaiaadm/pumba:0.6.5'
# in milliseconds
- name: NETWORK_DELAY
value: '60000'
- name: TC_IMAGE
value: 'gaiadocker/iproute2'
- name: TOTAL_CHAOS_DURATION
value: '60' # in seconds
- name: LIVENESS_APP_LABEL
value: ''
- name: LIVENESS_APP_NAMESPACE
value: ''
- name: DATA_PERSISTENCE
value: ''
labels:
name: openebs-pool-network-delay
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: experiment-job
app.kubernetes.io/version: latest

View File

@@ -1,44 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
name: openebs-pool-network-delay
version: 0.1.4
annotations:
categories: OpenEBS
vendor: CNCF
support: https://slack.openebs.io/
spec:
displayName: openebs-pool-network-delay
categoryDescription: >
Network delay to pool pod belonging to a OpenEBS PVC
keywords:
- Kubernetes
- K8S
- Storage Pool
- OpenEBS
- Network
platforms:
- GKE
- EKS
maturity: alpha
chaosType: infra
maintainers:
- name: Raj Babu Das
email: raj.das@mayadata.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
labels:
app.kubernetes.io/component: chartserviceversion
app.kubernetes.io/version: latest
links:
- name: Source Code
url: https://github.com/litmuschaos/litmus-ansible/tree/master/experiments/openebs/openebs-pool-network-delay
- name: Documentation
url: https://v1-docs.litmuschaos.io/docs/openebs-pool-network-delay/
- name: Video
url:
icon:
- url: ""
mediatype: ""
chaosexpcrdlink: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/openebs/openebs-pool-network-delay/experiment.yaml

View File

@@ -1,34 +0,0 @@
apiVersion: v1
kind: ServiceAccount
metadata:
name: pool-network-delay-sa
namespace: default
labels:
name: pool-network-delay-sa
---
# Source: openebs/templates/clusterrole.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: pool-network-delay-sa
labels:
name: pool-network-delay-sa
rules:
- apiGroups: ["","apps","litmuschaos.io","batch","extensions","storage.k8s.io","openebs.io"]
resources: ["pods","pods/exec","pods/log","events","jobs","configmaps","services","persistentvolumeclaims","storageclasses","persistentvolumes","chaosengines","chaosexperiments","chaosresults","cstorpools","cstorvolumereplicas","replicasets"]
verbs: ["create","list","get","patch","update","delete"]
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: pool-network-delay-sa
labels:
name: pool-network-delay-sa
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: pool-network-delay-sa
subjects:
- kind: ServiceAccount
name: pool-network-delay-sa
namespace: default

View File

@@ -1,27 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
kind: ChaosEngine
metadata:
name: pool-chaos
namespace: default
spec:
auxiliaryAppInfo: ''
# It can be active/stop
engineState: 'active'
appinfo:
appns: 'default'
applabel: 'app=nginx'
appkind: 'deployment'
chaosServiceAccount: pool-network-loss-sa
experiments:
- name: openebs-pool-network-loss
spec:
components:
env:
- name: TOTAL_CHAOS_DURATION
value: '120' # in seconds
- name: APP_PVC
value: 'demo-nginx-claim'
- name: NETWORK_PACKET_LOSS_PERCENTAGE
value: '100'

View File

@@ -1,92 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
description:
message: |
Network loss to pool pod belonging to a OpenEBS PVC
kind: ChaosExperiment
metadata:
labels:
litmuschaos.io/name: openebs
name: openebs-pool-network-loss
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: chaosexperiment
app.kubernetes.io/version: latest
name: openebs-pool-network-loss
spec:
definition:
scope: Cluster
permissions:
- apiGroups:
- ""
- "apps"
- "litmuschaos.io"
- "batch"
- "extensions"
- "storage.k8s.io"
- "openebs.io"
resources:
- "pods"
- "pods/log"
- "pods/exec"
- "events"
- "jobs"
- "configmaps"
- "services"
- "persistentvolumeclaims"
- "storageclasses"
- "persistentvolumeclaims"
- "persistentvolumes"
- "chaosengines"
- "chaosexperiments"
- "chaosresults"
- "cstorpools"
- "cstorvolumereplicas"
- "replicasets"
verbs:
- "create"
- "get"
- "delete"
- "list"
- "patch"
- "update"
image: "litmuschaos/ansible-runner:latest"
imagePullPolicy: Always
args:
- -c
- ansible-playbook ./experiments/openebs/openebs-pool-network-loss/openebs_pool_network_loss_ansible_logic.yml -i /etc/ansible/hosts -vv; exit 0
command:
- /bin/bash
env:
- name: ANSIBLE_STDOUT_CALLBACK
value: 'default'
- name: OPENEBS_NAMESPACE
value: 'openebs'
# only pumba supported
# For pumba image use : gaiaadm/pumba:0.6.5
- name: LIB_IMAGE
value: 'gaiaadm/pumba:0.6.5'
- name: TC_IMAGE
value: 'gaiadocker/iproute2'
- name: NETWORK_PACKET_LOSS_PERCENTAGE
value: '100' # in percentage
- name: TOTAL_CHAOS_DURATION
value: '120' # in seconds
- name: LIVENESS_APP_LABEL
value: ''
- name: LIVENESS_APP_NAMESPACE
value: ''
- name: DATA_PERSISTENCE
value: ''
labels:
name: openebs-pool-network-loss
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: experiment-job
app.kubernetes.io/version: latest

View File

@@ -1,44 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
name: openebs-pool-network-loss
version: 0.1.3
annotations:
categories: OpenEBS
vendor: CNCF
support: https://slack.openebs.io/
spec:
displayName: openebs-pool-network-loss
categoryDescription: >
Network loss to pool pod belonging to a OpenEBS PVC
keywords:
- Kubernetes
- K8S
- Storage Pool
- OpenEBS
- Network
platforms:
- GKE
- EKS
maturity: alpha
chaosType: infra
maintainers:
- name: Raj Babu Das
email: raj.das@mayadata.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
labels:
app.kubernetes.io/component: chartserviceversion
app.kubernetes.io/version: latest
links:
- name: Source Code
url: https://github.com/litmuschaos/litmus-ansible/tree/master/experiments/openebs/openebs-pool-network-loss
- name: Documentation
url: https://v1-docs.litmuschaos.io/docs/openebs-pool-network-loss/
- name: Video
url:
icon:
- url: ""
mediatype: ""
chaosexpcrdlink: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/openebs/openebs-pool-network-loss/experiment.yaml

View File

@@ -1,35 +0,0 @@
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: pool-network-loss-sa
namespace: default
labels:
name: pool-network-loss-sa
---
# Source: openebs/templates/clusterrole.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: pool-network-loss-sa
labels:
name: pool-network-loss-sa
rules:
- apiGroups: ["","apps","litmuschaos.io","batch","extensions","storage.k8s.io","openebs.io"]
resources: ["pods","pods/exec","pods/log","events","jobs","configmaps","services","persistentvolumeclaims","storageclasses","persistentvolumes","chaosengines","chaosexperiments","chaosresults","cstorpools","cstorvolumereplicas","replicasets"]
verbs: ["create","list","get","patch","update","delete"]
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: pool-network-loss-sa
labels:
name: pool-network-loss-sa
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: pool-network-loss-sa
subjects:
- kind: ServiceAccount
name: pool-network-loss-sa
namespace: default

View File

@@ -1,28 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
kind: ChaosEngine
metadata:
name: pool-chaos
namespace: default
spec:
# It can be active/stop
engineState: 'active'
#ex. values: ns1:name=percona,ns2:run=nginx
auxiliaryAppInfo: ''
appinfo:
appns: 'default'
applabel: 'app=nginx'
appkind: 'deployment'
chaosServiceAccount: pool-pod-failure-sa
experiments:
- name: openebs-pool-pod-failure
spec:
components:
env:
- name: CHAOS_ITERATIONS
value: '2'
- name: APP_PVC
value: 'demo-nginx-claim'
- name: DEPLOY_TYPE
value: 'deployment'

View File

@@ -1,97 +0,0 @@
---
apiVersion: litmuschaos.io/v1alpha1
description:
message: |
Kill the pool pod and check if gets scheduled again
kind: ChaosExperiment
metadata:
labels:
litmuschaos.io/name: openebs
name: openebs-pool-pod-failure
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: chaosexperiment
app.kubernetes.io/version: latest
name: openebs-pool-pod-failure
spec:
definition:
scope: Cluster
permissions:
- apiGroups:
- ""
- "extensions"
- "apps"
- "batch"
- "litmuschaos.io"
- "openebs.io"
- "storage.k8s.io"
resources:
- "deployments"
- "replicasets"
- "jobs"
- "pods/log"
- "events"
- "pods"
- "configmaps"
- "secrets"
- "storageclasses"
- "persistentvolumeclaims"
- "cstorvolumereplicas"
- "chaosengines"
- "chaosexperiments"
- "chaosresults"
verbs:
- "create"
- "get"
- "delete"
- "list"
- "patch"
- "update"
- apiGroups:
- ""
resources:
- "nodes"
verbs:
- "get"
- "list"
image: "litmuschaos/ansible-runner:latest"
imagePullPolicy: Always
args:
- -c
- ansible-playbook ./experiments/openebs/openebs-pool-pod-failure/openebs_pool_pod_failure_ansible_logic.yml -i /etc/ansible/hosts -vv; exit 0
command:
- /bin/bash
env:
- name: ANSIBLE_STDOUT_CALLBACK
value: default
- name: OPENEBS_NS
value: 'openebs'
- name: APP_PVC
value: ''
- name: LIVENESS_APP_LABEL
value: ''
- name: LIVENESS_APP_NAMESPACE
value: ''
- name: CHAOS_ITERATIONS
value: '2'
# provide the kill count
- name: KILL_COUNT
value: ''
- name: DATA_PERSISTENCE
value: ''
labels:
name: openebs-pool-pod-failure
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: experiment-job
app.kubernetes.io/version: latest
#configmaps:
#- name: openebs-pool-pod-failure
# mountPath: /mnt

View File

@@ -1,45 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
createdAt: 2019-11-12T10:28:08Z
name: openebs-pool-pod-failure
version: 0.1.12
annotations:
categories: OpenEBS
vendor: CNCF
support: https://slack.openebs.io/
spec:
displayName: openebs-pool-pod-failure
categoryDescription: >
Kill the pool pod and check if gets scheduled again.
keywords:
- Kubernetes
- K8S
- Storage Pool
- OpenEBS
platforms:
- GKE
- EKS
- OpenShift
maturity: alpha
chaosType: infra
maintainers:
- name: shubham chaudhary
email: shubham.chaudhary@mayadata.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
labels:
app.kubernetes.io/component: chartserviceversion
app.kubernetes.io/version: latest
links:
- name: Source Code
url: https://github.com/litmuschaos/litmus-ansible/tree/master/experiments/openebs/openebs-pool-pod-failure
- name: Documentation
url: https://v1-docs.litmuschaos.io/docs/openebs-pool-pod-failure/
- name: Video
url:
icon:
- url: ""
mediatype: ""
chaosexpcrdlink: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/openebs/openebs-pool-pod-failure/experiment.yaml

View File

@@ -1,38 +0,0 @@
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: pool-pod-failure-sa
namespace: default
labels:
name: pool-pod-failure-sa
---
# Source: openebs/templates/clusterrole.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: pool-pod-failure-sa
labels:
name: pool-pod-failure-sa
rules:
- apiGroups: ["","apps","litmuschaos.io","batch","extensions","storage.k8s.io","openebs.io"]
resources: ["pods","jobs","deployments","pods/log","events","configmaps","secrets","replicasets","persistentvolumeclaims","storageclasses","cstorvolumereplicas","chaosexperiments","chaosresults","chaosengines"]
verbs: ["create","list","get","patch","update","delete"]
- apiGroups: [""]
resources: ["nodes"]
verbs: ["get","list"]
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: pool-pod-failure-sa
labels:
name: pool-pod-failure-sa
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: pool-pod-failure-sa
subjects:
- kind: ServiceAccount
name: pool-pod-failure-sa
namespace: default

View File

@@ -1,32 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
kind: ChaosEngine
metadata:
name: target-chaos
namespace: default
spec:
# It can be active/stop
engineState: 'active'
#ex. values: ns1:name=percona,ns2:run=nginx
auxiliaryAppInfo: ''
appinfo:
appns: 'default'
applabel: 'app=nginx'
appkind: 'deployment'
chaosServiceAccount: target-container-failure-sa
experiments:
- name: openebs-target-container-failure
spec:
components:
env:
# provide the total chaos duration
- name: TOTAL_CHAOS_DURATION
value: '20'
- name: TARGET_CONTAINER
value: 'cstor-istgt'
- name: APP_PVC
value: 'demo-nginx-claim'
- name: DEPLOY_TYPE
value: 'deployment'

View File

@@ -1,113 +0,0 @@
---
apiVersion: litmuschaos.io/v1alpha1
description:
message: |
Kill the cstor target/Jiva controller container and check if gets created again
kind: ChaosExperiment
metadata:
labels:
litmuschaos.io/name: openebs
name: openebs-target-container-failure
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: chaosexperiment
app.kubernetes.io/version: latest
name: openebs-target-container-failure
spec:
definition:
scope: Cluster
permissions:
- apiGroups:
- ""
- "apps"
- "batch"
- "litmuschaos.io"
- "storage.k8s.io"
resources:
- "jobs"
- "pods"
- "events"
- "pods/exec"
- "pods/log"
- "configmaps"
- "secrets"
- "persistentvolumeclaims"
- "storageclasses"
- "persistentvolumes"
- "chaosengines"
- "chaosexperiments"
- "chaosresults"
verbs:
- "create"
- "delete"
- "get"
- "list"
- "patch"
- "update"
image: "litmuschaos/ansible-runner:latest"
imagePullPolicy: Always
args:
- -c
- ansible-playbook ./experiments/openebs/openebs-target-container-failure/openebs_target_container_failure_ansible_logic.yml -i /etc/ansible/hosts -vv; exit 0
command:
- /bin/bash
env:
- name: ANSIBLE_STDOUT_CALLBACK
value: 'default'
- name: OPENEBS_NAMESPACE
value: 'openebs'
- name: APP_PVC
value: ''
# LIB_IMAGE can be - gaiaadm/pumba:0.6.5, gprasath/crictl:ci
# For pumba image use : gaiaadm/pumba:0.6.5
# For containerd image use : gprasath/crictl:ci
- name: LIB_IMAGE
value: 'gaiaadm/pumba:0.6.5'
# Specify the container runtime used , to pick the relevant chaos util
- name: CONTAINER_RUNTIME
value: 'docker'
# TARGET_CONTAINER values: cstor-volume-mgmt , cstor-istgt
# For cstor-volume-istgt container kill use : cstor-istgt
# For volume-mgmt-kill container use : cstor-volume-mgmt
- name: TARGET_CONTAINER
value: 'cstor-volume-mgmt'
- name: FORCE
value: 'true'
- name: LIVENESS_APP_LABEL
value: ''
- name: LIVENESS_APP_NAMESPACE
value: ''
- name: DATA_PERSISTENCE
value: ''
- name: DEPLOY_TYPE
value: 'deployment'
# provide the chaos interval
- name: CHAOS_INTERVAL
value: '10'
# provide the total chaos duration
- name: TOTAL_CHAOS_DURATION
value: '20'
- name: SOAK_TIME
value: '60'
labels:
name: openebs-target-container-failure
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: experiment-job
app.kubernetes.io/version: latest
#configmaps:
#- name: openebs-target-container-failure
# mountPath: /mnt

View File

@@ -1,45 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
createdAt: 2019-11-11T10:28:08Z
name: openebs-target-container-failure
version: 0.1.12
annotations:
categories: OpenEBS
vendor: CNCF
support: https://slack.openebs.io/
spec:
displayName: openebs-target-container-failure
categoryDescription: >
Kill the cstor target/Jiva controller container and check if gets created again
keywords:
- Kubernetes
- K8S
- Storage Controller
- OpenEBS
platforms:
- GKE
- EKS
- OpenShift
maturity: alpha
chaosType: infra
maintainers:
- name: shubham chaudhary
email: shubham.chaudhary@mayadata.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
labels:
app.kubernetes.io/component: chartserviceversion
app.kubernetes.io/version: latest
links:
- name: Source Code
url: https://github.com/litmuschaos/litmus-ansible/tree/master/experiments/openebs/openebs-target-container-failure
- name: Documentation
url: https://v1-docs.litmuschaos.io/docs/openebs-target-container-failure/
- name: Video
url:
icon:
- url: ""
mediatype: ""
chaosexpcrdlink: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/openebs/openebs-target-container-failure/experiment.yaml

View File

@@ -1,38 +0,0 @@
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: target-container-failure-sa
namespace: default
labels:
name: target-container-failure-sa
app.kubernetes.io/part-of: litmus
---
# Source: openebs/templates/clusterrole.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: target-container-failure-sa
labels:
name: target-container-failure-sa
app.kubernetes.io/part-of: litmus
rules:
- apiGroups: ["","litmuschaos.io","batch","apps","storage.k8s.io"]
resources: ["pods","jobs","pods/log","pods/exec","events","configmaps","secrets","persistentvolumeclaims","storageclasses","persistentvolumes","chaosengines","chaosexperiments","chaosresults"]
verbs: ["create","list","get","patch","update","delete"]
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: target-container-failure-sa
labels:
name: target-container-failure-sa
app.kubernetes.io/part-of: litmus
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: target-container-failure-sa
subjects:
- kind: ServiceAccount
name: target-container-failure-sa
namespace: default

View File

@@ -1,35 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
kind: ChaosEngine
metadata:
name: target-chaos
namespace: default
spec:
# It can be active/stop
engineState: 'active'
#ex. values: ns1:name=percona,ns2:run=nginx
auxiliaryAppInfo: ''
appinfo:
appns: 'default'
applabel: 'app=nginx'
appkind: 'deployment'
chaosServiceAccount: target-network-delay-sa
experiments:
- name: openebs-target-network-delay
spec:
components:
env:
- name: TOTAL_CHAOS_DURATION
value: '60' # in seconds
- name: TARGET_CONTAINER
value: 'cstor-istgt'
- name: APP_PVC
value: 'demo-nginx-claim'
- name: DEPLOY_TYPE
value: 'deployment'
- name: NETWORK_DELAY
value: '30000'

View File

@@ -1,94 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
description:
message: |
Network delay to target pod belonging to a deployment/statefulset/daemonset
kind: ChaosExperiment
metadata:
labels:
litmuschaos.io/name: openebs
name: openebs-target-network-delay
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: chaosexperiment
app.kubernetes.io/version: latest
name: openebs-target-network-delay
spec:
definition:
scope: Cluster
permissions:
- apiGroups:
- ""
- "extensions"
- "apps"
- "batch"
- "litmuschaos.io"
- "storage.k8s.io"
resources:
- "jobs"
- "pods"
- "services"
- "events"
- "pods/exec"
- "pods/log"
- "configmaps"
- "secrets"
- "persistentvolumeclaims"
- "storageclasses"
- "persistentvolumes"
- "chaosengines"
- "chaosexperiments"
- "chaosresults"
verbs:
- "create"
- "get"
- "delete"
- "list"
- "patch"
- "update"
image: "litmuschaos/ansible-runner:latest"
imagePullPolicy: Always
args:
- -c
- ansible-playbook ./experiments/openebs/openebs-target-network-delay/openebs_target_network_delay_ansible_logic.yml -i /etc/ansible/hosts -vv; exit 0
command:
- /bin/bash
env:
- name: ANSIBLE_STDOUT_CALLBACK
value: 'default'
- name: OPENEBS_NAMESPACE
value: 'openebs'
- name: APP_PVC
value: ''
- name: TC_IMAGE
value: 'gaiadocker/iproute2'
# only pumba supported
# For pumba image use : gaiaadm/pumba:0.6.5
- name: LIB_IMAGE
value: 'gaiaadm/pumba:0.6.5'
- name: NETWORK_DELAY
value: '60000' # in milliseconds
- name: TOTAL_CHAOS_DURATION
value: '60' # in seconds
- name: LIVENESS_APP_LABEL
value: ''
- name: LIVENESS_APP_NAMESPACE
value: ''
- name: DATA_PERSISTENCE
value: ''
labels:
name: openebs-target-network-delay
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: experiment-job
app.kubernetes.io/version: latest
#configmaps:
#- name: openebs-target-network-delay
# mountPath: /mnt

View File

@@ -1,45 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
createdAt: 2019-10-17T10:28:08Z
name: openebs-target-network-delay
version: 0.1.9
annotations:
categories: Kubernetes
vendor: CNCF
support: https://slack.kubernetes.io/
spec:
displayName: openebs-target-network-delay
categoryDescription: >
Network delay to Pod belonging to a OpenEBS PVC
keywords:
- OpenEBS
- Kubernetes
- K8S
- Network
platforms:
- GKE
- EKS
- OpenShift
maturity: alpha
chaosType: infra
maintainers:
- name: giri
email: giridhara.prasad@mayadata.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
labels:
app.kubernetes.io/component: chartserviceversion
app.kubernetes.io/version: latest
links:
- name: Source Code
url: https://github.com/litmuschaos/litmus-ansible/tree/master/experiments/openebs/openebs-target-network-delay
- name: Documentation
url: https://v1-docs.litmuschaos.io/docs/openebs-target-network-delay/
- name: Video
url:
icon:
- url:
mediatype: ""
chaosexpcrdlink: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/openebs/openebs-target-network-delay/experiment.yaml

View File

@@ -1,38 +0,0 @@
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: target-network-delay-sa
namespace: default
labels:
name: target-network-delay-sa
app.kubernetes.io/part-of: litmus
---
# Source: openebs/templates/clusterrole.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: target-network-delay-sa
labels:
name: target-network-delay-sa
app.kubernetes.io/part-of: litmus
rules:
- apiGroups: ["","apps","litmuschaos.io","batch","extensions","storage.k8s.io"]
resources: ["pods","pods/exec","pods/log","events","jobs","configmaps","secrets","services","persistentvolumeclaims","storageclasses","persistentvolumes","chaosexperiments","chaosresults","chaosengines"]
verbs: ["create","list","get","patch","update","delete"]
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: target-network-delay-sa
labels:
name: target-network-delay-sa
app.kubernetes.io/part-of: litmus
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: target-network-delay-sa
subjects:
- kind: ServiceAccount
name: target-network-delay-sa
namespace: default

View File

@@ -1,31 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
kind: ChaosEngine
metadata:
name: target-chaos
namespace: default
spec:
# It can be active/stop
engineState: 'active'
#ex. values: ns1:name=percona,ns2:run=nginx
auxiliaryAppInfo: ''
appinfo:
appns: 'default'
applabel: 'app=nginx'
appkind: 'deployment'
chaosServiceAccount: target-network-loss-sa
experiments:
- name: openebs-target-network-loss
spec:
components:
env:
- name: TOTAL_CHAOS_DURATION
value: '120' # in seconds
- name: TARGET_CONTAINER
value: 'cstor-istgt'
- name: APP_PVC
value: 'demo-nginx-claim'
- name: DEPLOY_TYPE
value: 'deployment'

View File

@@ -1,94 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
description:
message: |
Network loss to target pod belonging to a OpenEBS PVC
kind: ChaosExperiment
metadata:
labels:
litmuschaos.io/name: openebs
name: openebs-target-network-loss
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: chaosexperiment
app.kubernetes.io/version: latest
name: openebs-target-network-loss
spec:
definition:
scope: Cluster
permissions:
- apiGroups:
- ""
- "extensions"
- "apps"
- "batch"
- "litmuschaos.io"
- "storage.k8s.io"
resources:
- "jobs"
- "pods"
- "events"
- "services"
- "pods/log"
- "pods/exec"
- "configmaps"
- "secrets"
- "persistentvolumeclaims"
- "storageclasses"
- "persistentvolumes"
- "chaosengines"
- "chaosexperiments"
- "chaosresults"
verbs:
- "create"
- "get"
- "delete"
- "list"
- "patch"
- "update"
image: "litmuschaos/ansible-runner:latest"
imagePullPolicy: Always
args:
- -c
- ansible-playbook ./experiments/openebs/openebs-target-network-loss/openebs_target_network_loss_ansible_logic.yml -i /etc/ansible/hosts -vv; exit 0
command:
- /bin/bash
env:
- name: ANSIBLE_STDOUT_CALLBACK
value: 'default'
- name: OPENEBS_NAMESPACE
value: 'openebs'
- name: APP_PVC
value: ''
- name: TC_IMAGE
value: 'gaiadocker/iproute2'
# only pumba supported
# For pumba image use : gaiaadm/pumba:0.6.5
- name: LIB_IMAGE
value: 'gaiaadm/pumba:0.6.5'
- name: NETWORK_PACKET_LOSS_PERCENTAGE
value: '100' # in percentage
- name: TOTAL_CHAOS_DURATION
value: '120' # in seconds
- name: LIVENESS_APP_LABEL
value: ''
- name: LIVENESS_APP_NAMESPACE
value: ''
- name: DATA_PERSISTENCE
value: ''
labels:
name: openebs-target-network-loss
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: experiment-job
app.kubernetes.io/version: latest
#configmaps:
#- name: openebs-target-network-loss
# mountPath: /mnt

View File

@@ -1,45 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
createdAt: 2019-10-17T10:28:08Z
name: openebs-target-network-loss
version: 0.1.9
annotations:
categories: Kubernetes
vendor: CNCF
support: https://slack.kubernetes.io/
spec:
displayName: openebs-target-network-loss
categoryDescription: >
Network loss to the target pod belonging to an OpenEBS PVC
keywords:
- Kubernetes
- K8S
- OpenEBS
- Network
platforms:
- GKE
- EKS
- OpenShift
maturity: alpha
chaosType: infra
maintainers:
- name: giri
email: giridhara.prasad@mayadata.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
labels:
app.kubernetes.io/component: chartserviceversion
app.kubernetes.io/version: latest
links:
- name: Source Code
url: https://github.com/litmuschaos/litmus-ansible/tree/master/experiments/openebs/openebs-target-network-loss
- name: Documentation
url: https://v1-docs.litmuschaos.io/docs/openebs-target-network-loss/
- name: Video
url:
icon:
- url: ""
mediatype: ""
chaosexpcrdlink: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/openebs/openebs-target-network-loss/experiment.yaml

View File

@@ -1,38 +0,0 @@
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: target-network-loss-sa
namespace: default
labels:
name: target-network-loss-sa
app.kubernetes.io/part-of: litmus
---
# Source: openebs/templates/clusterrole.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: target-network-loss-sa
labels:
name: target-network-loss-sa
app.kubernetes.io/part-of: litmus
rules:
- apiGroups: ["","apps","litmuschaos.io","batch","extensions","storage.k8s.io"]
resources: ["pods","pods/exec","pods/log","events","jobs","configmaps","secrets","services","persistentvolumeclaims","storageclasses","persistentvolumes","chaosexperiments","chaosresults","chaosengines"]
verbs: ["create","list","get","patch","update","delete"]
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: target-network-loss-sa
labels:
name: target-network-loss-sa
app.kubernetes.io/part-of: litmus
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: target-network-loss-sa
subjects:
- kind: ServiceAccount
name: target-network-loss-sa
namespace: default

View File

@@ -1,31 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
kind: ChaosEngine
metadata:
name: target-chaos
namespace: default
spec:
# It can be active/stop
engineState: 'active'
#ex. values: ns1:name=percona,ns2:run=nginx
auxiliaryAppInfo: ''
appinfo:
appns: 'default'
applabel: 'app=nginx'
appkind: 'deployment'
chaosServiceAccount: target-pod-failure-sa
experiments:
- name: openebs-target-pod-failure
spec:
components:
env:
- name: TOTAL_CHAOS_DURATION
value: '60'
- name: FORCE
value: 'true'
- name: APP_PVC
value: 'demo-nginx-claim'
- name: DEPLOY_TYPE
value: 'deployment'

View File

@@ -1,105 +0,0 @@
---
apiVersion: litmuschaos.io/v1alpha1
description:
message: |
Kill the cstor target/Jiva controller pod and check if gets created again
kind: ChaosExperiment
metadata:
labels:
litmuschaos.io/name: openebs
name: openebs-target-pod-failure
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: chaosexperiment
app.kubernetes.io/version: latest
name: openebs-target-pod-failure
spec:
definition:
scope: Cluster
permissions:
- apiGroups:
- ""
- "extensions"
- "apps"
- "batch"
- "litmuschaos.io"
- "storage.k8s.io"
resources:
- "deployments"
- "jobs"
- "events"
- "pods"
- "pods/log"
- "pods/exec"
- "configmaps"
- "secrets"
- "services"
- "chaosengines"
- "chaosexperiments"
- "chaosresults"
- "persistentvolumeclaims"
- "storageclasses"
- "persistentvolumes"
verbs:
- "create"
- "get"
- "delete"
- "list"
- "patch"
- "update"
- apiGroups:
- ""
resources:
- "nodes"
verbs:
- "get"
- "list"
image: "litmuschaos/ansible-runner:latest"
imagePullPolicy: Always
args:
- -c
- ansible-playbook ./experiments/openebs/openebs-target-pod-failure/openebs_target_pod_failure_ansible_logic.yml -i /etc/ansible/hosts -vv; exit 0
command:
- /bin/bash
env:
- name: ANSIBLE_STDOUT_CALLBACK
value: 'default'
- name: OPENEBS_NAMESPACE
value: 'openebs'
- name: APP_PVC
value: ''
- name: FORCE
value: 'true'
- name: LIVENESS_APP_LABEL
value: ''
- name: LIVENESS_APP_NAMESPACE
value: ''
- name: DATA_PERSISTENCE
value: ''
- name: TOTAL_CHAOS_DURATION
value: '60'
# provide the kill count
- name: KILL_COUNT
value: ''
- name: CHAOS_INTERVAL
value: '15'
- name: DEPLOY_TYPE
value: 'deployment'
labels:
name: openebs-target-pod-failure
app.kubernetes.io/part-of: litmus
app.kubernetes.io/component: experiment-job
app.kubernetes.io/version: latest
#configmaps:
#- name: openebs-target-pod-failure
# mountPath: /mnt

View File

@@ -1,46 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
createdAt: 2019-11-11T10:28:08Z
name: openebs-target-pod-failure
version: 0.1.11
annotations:
categories: OpenEBS
vendor: CNCF
support: https://slack.openebs.io/
spec:
displayName: openebs-target-pod-failure
categoryDescription: >
Kill the cstor target/Jiva controller pod and check if gets created again
keywords:
- Kubernetes
- K8S
- Storage Controller
- OpenEBS
platforms:
- GKE
- EKS
- OpenShift
maturity: alpha
chaosType: infra
maintainers:
- name: shubham chaudhary
email: shubham.chaudhary@mayadata.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
labels:
app.kubernetes.io/component: chartserviceversion
app.kubernetes.io/version: latest
links:
- name: Source Code
url: https://github.com/litmuschaos/litmus-ansible/tree/master/experiments/openebs/openebs-target-pod-failure
- name: Documentation
url: https://v1-docs.litmuschaos.io/docs/openebs-target-pod-failure/
- name: Video
url:
icon:
- url: ""
mediatype: ""
chaosexpcrdlink: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/openebs/openebs-target-pod-failure/experiment.yaml

View File

@@ -1,42 +0,0 @@
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: target-pod-failure-sa
namespace: default
labels:
name: target-pod-failure-sa
app.kubernetes.io/part-of: litmus
---
# Source: openebs/templates/clusterrole.yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: target-pod-failure-sa
labels:
name: target-pod-failure-sa
app.kubernetes.io/part-of: litmus
rules:
- apiGroups: ["","apps","litmuschaos.io","batch","extensions","storage.k8s.io"]
resources: ["pods","jobs","pods/log","deployments","pods/exec","events","chaosexperiments","chaosresults","chaosengines","configmaps","secrets","services","persistentvolumeclaims","storageclasses","persistentvolumes"]
verbs: ["create","list","get","patch","update","delete"]
- apiGroups: [""]
resources: ["nodes"]
verbs: ["get","list"]
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: target-pod-failure-sa
labels:
name: target-pod-failure-sa
app.kubernetes.io/part-of: litmus
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: target-pod-failure-sa
subjects:
- kind: ServiceAccount
name: target-pod-failure-sa
namespace: default

View File

@@ -1,52 +0,0 @@
apiVersion: litmuchaos.io/v1alpha1
kind: ChartServiceVersion
metadata:
createdAt: 2019-08-27T10:28:08Z
name: openebs
version: 0.1.8
annotations:
categories: OpenEBS
vendor: MayaData
support: https://slack.openebs.io/
chartDescription: Injects generic openebs chaos
spec:
displayName: OpenEBS
categoryDescription: >
OpenEBS is the most active Container Attached Storage project. OpenEBS enables your DevOps teams to have their own storage policies for every workload and it is truly Kubernetes native and 100% in userspace.
experiments:
- openebs-pool-pod-failure
- openebs-pool-container-failure
- openebs-target-pod-failure
- openebs-target-container-failure
- openebs-target-network-delay
- openebs-target-network-loss
- openebs-pool-network-delay
- openebs-pool-network-loss
- openebs-control-plane-chaos
- openebs-pool-disk-loss
- openebs-nfs-provisioner-kill
keywords:
- Kubernetes
- K8S
- OpenEBS
- CAS
- Storage
- Network
- Pod
- Pool
maintainers:
- name: ksatchit
email: karthik.s@mayadata.io
minKubeVersion: 1.12.0
provider:
name: Mayadata
links:
- name: OpenEBS Website
url: https://openebs.io
- name: Source Code
url: https://github.com/openebs/openebs
icon:
- url: https://raw.githubusercontent.com/litmuschaos/charthub.litmuschaos.io/master/public/icons/openebs.png
mediatype: image/png
chaosexpcrdlink: https://raw.githubusercontent.com/litmuschaos/chaos-charts/master/charts/openebs/experiments.yaml

View File

@@ -1,35 +0,0 @@
packageName: openebs
experiments:
- name: openebs-pool-pod-failure
CSV: openebs-pool-pod-failure.chartserviceversion.yaml
desc: "openebs-pool-pod-failure"
- name: openebs-pool-container-failure
CSV: openebs-pool-container-failure.chartserviceversion.yaml
desc: "openebs-pool-container-failure"
- name: openebs-target-pod-failure
CSV: openebs-target-pod-failure.chartserviceversion.yaml
desc: "openebs-target-pod-failure"
- name: openebs-target-container-failure
CSV: openebs-target-container-failure.chartserviceversion.yaml
desc: "openebs-target-container-failure"
- name: openebs-target-network-delay
CSV: openebs-target-network-delay.chartserviceversion.yaml
desc: "openebs-target-network-delay"
- name: openebs-target-network-loss
CSV: openebs-target-network-loss.chartserviceversion.yaml
desc: "openebs-target-network-loss"
- name: openebs-control-plane-chaos
CSV: openebs-control-plane-chaos.chartserviceversion.yaml
desc: "openebs-control-plane-chaos"
- name: openebs-pool-network-delay
CSV: openebs-pool-network-delay.chartserviceversion.yaml
desc: "openebs-pool-network-delay"
- name: openebs-pool-network-loss
CSV: openebs-pool-network-loss.chartserviceversion.yaml
desc: "openebs-pool-network-loss"
- name: openebs-pool-disk-loss
CSV: openebs-pool-disk-loss.chartserviceversion.yaml
desc: "openebs-pool-disk-loss"
- name: openebs-nfs-provisioner-kill
CSV: openebs-nfs-provisioner-kill.chartserviceversion.yaml
desc: "openebs-nfs-provisioner-kill"

View File

@@ -1,30 +0,0 @@
apiVersion: litmuschaos.io/v1alpha1
kind: ChaosEngine
metadata:
name: openebs-chaos
namespace: default
spec:
# It can be active/stop
engineState: 'active'
#ex. values: ns1:name=percona,ns2:run=nginx
auxiliaryAppInfo: ''
appinfo:
appns: 'default'
applabel: 'app=nginx'
appkind: 'deployment'
chaosServiceAccount: <experiments-name>-sa
experiments:
- name: <experiments-name>
spec:
components:
env:
- name: TARGET_CONTAINER
value: 'cstor-istgt'
- name: APP_PVC
value: 'demo-nginx-claim'
configMaps:
- name: map1
mountPath: '/mnt'
secrets:
- name: secret1
mountPath: '/tmp'