5335667721 | ||
---|---|---|
.. | ||
templates | ||
Chart.yaml | ||
README.md | ||
values.yaml |
README.md
Redis
Redis is an advanced key-value cache and store. It is often referred to as a data structure server since keys can contain strings, hashes, lists, sets, sorted sets, bitmaps and hyperloglogs.
TL;DR
helm repo add dandydev https://dandydeveloper.github.io/charts
helm install dandydev/redis-ha
By default this chart install 3 pods total:
- one pod containing a redis master and sentinel container (optional prometheus metrics exporter sidecar available)
- two pods each containing a redis slave and sentinel containers (optional prometheus metrics exporter sidecars available)
Introduction
This chart bootstraps a Redis highly available master/slave statefulset in a Kubernetes cluster using the Helm package manager.
Prerequisites
- Kubernetes 1.8+ with Beta APIs enabled
- PV provisioner support in the underlying infrastructure
Upgrading the Chart
Please note that there have been a number of changes simplifying the redis management strategy (for better failover and elections) in the 3.x version of this chart. These changes allow the use of official redis images that do not require special RBAC or ServiceAccount roles. As a result when upgrading from version >=2.0.1 to >=3.0.0 of this chart, Role
, RoleBinding
, and ServiceAccount
resources should be deleted manually.
Upgrading the chart from 3.x to 4.x
Starting from version 4.x
HAProxy sidecar prometheus-exporter removed and replaced by the embedded HAProxy metrics endpoint, as a result when upgrading from version 3.x to 4.x section haproxy.exporter
should be removed and the haproxy.metrics
need to be configured for fit your needs.
Installing the Chart
To install the chart
helm repo add dandydev https://dandydeveloper.github.io/charts
helm install dandydev/redis-ha
The command deploys Redis on the Kubernetes cluster in the default configuration. By default this chart install one master pod containing redis master container and sentinel container along with 2 redis slave pods each containing their own sentinel sidecars. The configuration section lists the parameters that can be configured during installation.
Tip: List all releases using
helm list
Uninstalling the Chart
To uninstall/delete the deployment:
helm delete <chart-name>
The command removes all the Kubernetes components associated with the chart and deletes the release.
Configuration
The following table lists the configurable parameters of the Redis chart and their default values.
Parameter | Description | Default |
---|---|---|
image.repository |
Redis image repository | redis |
image.tag |
Redis image tag | 6.2.5-alpine |
image.pullPolicy |
Redis image pull policy | IfNotPresent |
imagePullSecrets |
Reference to one or more secrets to be used when pulling redis images | [] |
tag |
Redis tag | 6.2.5-alpine |
replicas |
Number of redis master/slave pods | 3 |
podManagementPolicy |
The statefulset pod management policy | OrderedReady |
ro_replicas |
Comma separated list of slaves which never get promoted to be master. Count starts with 0. Allowed values 1-9. i.e. 3,4 - 3th and 4th redis slave never make it to be master, where master is index 0. | `` |
serviceAccount.create |
Specifies whether a ServiceAccount should be created | true |
serviceAccount.name |
The name of the ServiceAccount to create | Generated using the redis-ha.fullname template |
serviceAccount.automountToken |
Opt in/out of automounting API credentials into container | false |
serviceAnnotations |
Annotations to set on Redis HA service | null |
serviceLabels |
Labels to set on Redis HA service | {} |
rbac.create |
Create and use RBAC resources | true |
redis.port |
Port to access the redis service | 6379 |
redis.tlsPort |
TLS Port to access the redis service | `` |
redis.tlsReplication |
Configures redis with tls-replication parameter, if true sets "tls-replication yes" in redis.conf | `` |
redis.authClients |
It is possible to disable client side certificates authentication when "authClients" is set to "no" | `` |
redis.livenessProbe.initialDelaySeconds |
Initial delay in seconds for liveness probe | 30 |
redis.livenessProbe.periodSeconds |
Period in seconds after which liveness probe will be repeated | 15 |
redis.livenessProbe.timeoutSeconds |
Timeout seconds for liveness probe | 15 |
redis.livenessProbe.successThreshold |
Success threshold for liveness probe | 1 |
redis.livenessProbe.failureThreshold |
Failure threshold for liveness probe | 5 |
redis.readinessProbe.initialDelaySeconds |
Initial delay in seconds for readiness probe | 30 |
redis.readinessProbe.periodSeconds |
Period in seconds after which readiness probe will be repeated | 15 |
redis.readinessProbe.timeoutSeconds |
Timeout seconds for readiness probe | 15 |
redis.readinessProbe.successThreshold |
Success threshold for readiness probe | 1 |
redis.readinessProbe.failureThreshold |
Failure threshold for readiness probe | 5 |
redis.masterGroupName |
Redis convention for naming the cluster group: must match ^[\\w-\\.]+$ and can be templated |
mymaster |
redis.disableCommands |
Array with commands to disable | ["FLUSHDB","FLUSHALL"] |
redis.config |
Any valid redis config options in this section will be applied to each server (see below) | see values.yaml |
redis.customConfig |
Allows for custom redis.conf files to be applied. If this is used then redis.config is ignored |
`` |
redis.resources |
CPU/Memory for master/slave nodes resource requests/limits | {} |
redis.lifecycle |
Container Lifecycle Hooks for redis container | see values.yaml |
redis.annotations |
Annotations for the redis statefulset | {} |
redis.updateStategy.type |
Update strategy for redis statefulSet | RollingUpdate |
redis.extraVolumeMounts |
Extra volume mounts for Redis container | [] |
sentinel.port |
Port to access the sentinel service | 26379 |
sentinel.bind |
Configure the 'bind' directive to bind to a list of network interfaces | `` |
sentinel.tlsPort |
TLS Port to access the sentinel service | `` |
sentinel.tlsReplication |
Configures sentinel with tls-replication parameter, if true sets "tls-replication yes" in sentinel.conf | `` |
sentinel.authClients |
It is possible to disable client side certificates authentication when "authClients" is set to "no" | `` |
sentinel.livenessProbe.initialDelaySeconds |
Initial delay in seconds for liveness probe | 30 |
sentinel.livenessProbe.periodSeconds |
Period in seconds after which liveness probe will be repeated | 15 |
sentinel.livenessProbe.timeoutSeconds |
Timeout seconds for liveness probe | 15 |
sentinel.livenessProbe.successThreshold |
Success threshold for liveness probe | 1 |
sentinel.livenessProbe.failureThreshold |
Failure threshold for liveness probe | 5 |
sentinel.readinessProbe.initialDelaySeconds |
Initial delay in seconds for readiness probe | 30 |
sentinel.readinessProbe.periodSeconds |
Period in seconds after which readiness probe will be repeated | 15 |
sentinel.readinessProbe.timeoutSeconds |
Timeout seconds for readiness probe | 15 |
sentinel.readinessProbe.successThreshold |
Success threshold for readiness probe | 3 |
sentinel.readinessProbe.failureThreshold |
Failure threshold for readiness probe | 5 |
sentinel.auth |
Enables or disables sentinel AUTH (Requires sentinel.password to be set) |
false |
sentinel.password |
A password that configures a requirepass in the conf parameters (Requires sentinel.auth: enabled ) |
`` |
sentinel.existingSecret |
An existing secret containing a key defined by sentinel.authKey that configures requirepass in the conf parameters (Requires sentinel.auth: enabled , cannot be used in conjunction with .Values.sentinel.password ) |
`` |
sentinel.authKey |
The key holding the sentinel password in an existing secret. | sentinel-password |
sentinel.quorum |
Minimum number of servers necessary to maintain quorum | 2 |
sentinel.config |
Valid sentinel config options in this section will be applied as config options to each sentinel (see below) | see values.yaml |
sentinel.customConfig |
Allows for custom sentinel.conf files to be applied. If this is used then sentinel.config is ignored |
`` |
sentinel.resources |
CPU/Memory for sentinel node resource requests/limits | {} |
sentinel.lifecycle |
Container Lifecycle Hooks for sentinel container | {} |
sentinel.extraVolumeMounts |
Extra volume mounts for Sentinel container | [] |
init.resources |
CPU/Memory for init Container node resource requests/limits | {} |
auth |
Enables or disables redis AUTH (Requires redisPassword to be set) |
false |
redisPassword |
A password that configures a requirepass and masterauth in the conf parameters (Requires auth: enabled ) |
`` |
authKey |
The key holding the redis password in an existing secret. | auth |
existingSecret |
An existing secret containing a key defined by authKey that configures requirepass and masterauth in the conf parameters (Requires auth: enabled , cannot be used in conjunction with .Values.redisPassword ) |
`` |
nodeSelector |
Node labels for pod assignment | {} |
tolerations |
Toleration labels for pod assignment | [] |
hardAntiAffinity |
Whether the Redis server pods should be forced to run on separate nodes. | true |
additionalAffinities |
Additional affinities to add to the Redis server pods. | {} |
securityContext |
Security context to be added to the Redis StatefulSet. | {runAsUser: 1000, fsGroup: 1000, runAsNonRoot: true} |
containerSecurityContext |
Security context to be added to the Redis containers. | { runAsNonRoot: true, allowPrivilegeEscalation: false, seccompProfile: { type: RuntimeDefault }, capabilities: { drop: [ "ALL" ] } |
affinity |
Override all other affinity settings with a string. | "" |
labels |
Labels for the Redis pod. | {} |
configmap.labels |
Labels for the Redis configmap. | {} |
configmapTest.image.repository |
Repository of the configmap shellcheck test image. | koalaman/shellcheck |
configmapTest.image.tag |
Tag of the configmap shellcheck test image. | v0.5.0 |
configmapTest.resources |
Resources for the ConfigMap tests. | {} |
persistentVolume.size |
Size for the volume | 10Gi |
persistentVolume.annotations |
Annotations for the volume | {} |
persistentVolume.labels |
Labels for the volume | {} |
emptyDir |
Configuration of emptyDir , used only if persistentVolume is disabled and no hostPath specified |
{} |
exporter.enabled |
If true , the prometheus exporter sidecar is enabled |
false |
exporter.image |
Exporter image | oliver006/redis_exporter |
exporter.tag |
Exporter tag | v1.27.0 |
exporter.port |
Exporter port | 9121 |
exporter.portName |
Exporter port name | exporter-port |
exporter.address |
Redis instance Hostname/Address Exists to circumvent some issues with issues in IPv6 hostname resolution | localhost |
exporter.annotations |
Prometheus scrape annotations | {prometheus.io/path: /metrics, prometheus.io/port: "9121", prometheus.io/scrape: "true"} |
exporter.extraArgs |
Additional args for the exporter | {} |
exporter.script |
A custom custom Lua script that will be mounted to exporter for collection of custom metrics. Creates a ConfigMap and sets env var REDIS_EXPORTER_SCRIPT . |
|
exporter.serviceMonitor.enabled |
Use servicemonitor from prometheus operator | false |
exporter.serviceMonitor.namespace |
Namespace the service monitor is created in | default |
exporter.serviceMonitor.interval |
Scrape interval, If not set, the Prometheus default scrape interval is used | nil |
exporter.serviceMonitor.telemetryPath |
Path to redis-exporter telemetry-path | /metrics |
exporter.serviceMonitor.labels |
Labels for the servicemonitor passed to Prometheus Operator | {} |
exporter.serviceMonitor.timeout |
How long until a scrape request times out. If not set, the Prometheus default scape timeout is used | nil |
haproxy.enabled |
Enabled HAProxy LoadBalancing/Proxy | false |
haproxy.replicas |
Number of HAProxy instances | 3 |
haproxy.servicePort |
Modify HAProxy service port | 6379 |
haproxy.containerPort |
Modify HAProxy deployment container port | 6379 |
haproxy.image.repository |
HAProxy Image Repository | haproxy |
haproxy.image.tag |
HAProxy Image Tag | 2.4.2 |
haproxy.image.pullPolicy |
HAProxy Image PullPolicy | IfNotPresent |
haproxy.imagePullSecrets |
Reference to one or more secrets to be used when pulling haproxy images | [] |
haproxy.tls.enabled |
If "true" this will enable TLS termination on haproxy | false |
haproxy.tls.secretName |
Secret containing the .pem file | "" |
haproxy.tls.certMountPath |
Path to mount the secret that contains the certificates. haproxy | false |
haproxy.tls.secretName |
Secret containing the .pem file | "" |
haproxy.annotations |
HAProxy template annotations | {} |
haproxy.customConfig |
Allows for custom config-haproxy.cfg file to be applied. If this is used then default config will be overwriten | `` |
haproxy.extraConfig |
Allows to place any additional configuration section to add to the default config-haproxy.cfg | `` |
haproxy.resources |
HAProxy resources | {} |
haproxy.emptyDir |
Configuration of emptyDir |
{} |
haproxy.labels |
Labels for the HAProxy pod | {} |
haproxy.service.type |
HAProxy service type "ClusterIP", "LoadBalancer" or "NodePort" | ClusterIP |
haproxy.service.nodePort |
HAProxy service nodePort value (haproxy.service.type must be NodePort) | not set |
haproxy.image.serviceAccountName |
HAProxy serviceAccountName | default |
haproxy.service.externalTrafficPolicy |
HAProxy service externalTrafficPolicy value (haproxy.service.type must be LoadBalancer) | not set |
haproxy.service.annotations |
HAProxy service annotations | {} |
haproxy.service.labels |
HAProxy service labels | {} |
haproxy.stickyBalancing |
HAProxy sticky load balancing to Redis nodes. Helps with connections shutdown. | false |
haproxy.hapreadport.enable |
Enable a read only port for redis slaves | false |
haproxy.hapreadport.port |
Haproxy port for read only redis slaves | 6380 |
haproxy.metrics.enabled |
HAProxy enable prometheus metric scraping | false |
haproxy.metrics.port |
HAProxy prometheus metrics scraping port | 9101 |
haproxy.metrics.portName |
HAProxy metrics scraping port name | http-exporter-port |
haproxy.metrics.scrapePath |
HAProxy prometheus metrics scraping port | /metrics |
haproxy.metrics.serviceMonitor.enabled |
Use servicemonitor from prometheus operator for HAProxy metrics | false |
haproxy.metrics.serviceMonitor.namespace |
Namespace the service monitor for HAProxy metrics is created in | default |
haproxy.metrics.serviceMonitor.interval |
Scrape interval, If not set, the Prometheus default scrape interval is used | nil |
haproxy.metrics.serviceMonitor.telemetryPath |
Path to HAProxy metrics telemetry-path | /metrics |
haproxy.metrics.serviceMonitor.labels |
Labels for the HAProxy metrics servicemonitor passed to Prometheus Operator | {} |
haproxy.metrics.serviceMonitor.timeout |
How long until a scrape request times out. If not set, the Prometheus default scape timeout is used | nil |
haproxy.init.resources |
Extra init resources | {} |
haproxy.timeout.connect |
haproxy.cfg timeout connect setting |
4s |
haproxy.timeout.server |
haproxy.cfg timeout server setting |
30s |
haproxy.timeout.client |
haproxy.cfg timeout client setting |
30s |
haproxy.timeout.check |
haproxy.cfg timeout check setting |
2s |
haproxy.checkInterval |
haproxy.cfg check inter setting |
1s |
haproxy.checkFall |
haproxy.cfg check fall setting |
1 |
haproxy.priorityClassName |
priorityClassName for haproxy deployment |
not set |
haproxy.securityContext |
Security context to be added to the HAProxy deployment. | {runAsUser: 99, fsGroup: 99, runAsNonRoot: true} |
haproxy.containerSecurityContext |
Security context to be added to the HAProxy containers. | { runAsNonRoot: true, allowPrivilegeEscalation: false, seccompProfile: { type: RuntimeDefault }, capabilities: { drop: [ "ALL" ] } |
haproxy.hardAntiAffinity |
Whether the haproxy pods should be forced to run on separate nodes. | true |
haproxy.affinity |
Override all other haproxy affinity settings with a string. | "" |
haproxy.additionalAffinities |
Additional affinities to add to the haproxy server pods. | {} |
haproxy.tests.resources |
Pod resources for the tests against HAProxy. | {} |
haproxy.IPv6.enabled |
Disables certain binding options to support non-IPv6 environments. | true |
podDisruptionBudget |
Pod Disruption Budget rules | {} |
nameOverride |
Override the chart name | "" |
fullnameOverride |
Fully override the release name and chart name | "" |
priorityClassName |
priorityClassName for redis-ha-statefulset |
not set |
hostPath.path |
Use this path on the host for data storage | not set |
hostPath.chown |
Run an init-container as root to set ownership on the hostPath | true |
sysctlImage.enabled |
Enable an init container to modify Kernel settings | false |
sysctlImage.command |
sysctlImage command to execute | [] |
sysctlImage.registry |
sysctlImage Init container registry | docker.io |
sysctlImage.repository |
sysctlImage Init container name | busybox |
sysctlImage.tag |
sysctlImage Init container tag | 1.31.1 |
sysctlImage.pullPolicy |
sysctlImage Init container pull policy | Always |
sysctlImage.mountHostSys |
Mount the host /sys folder to /host-sys |
false |
sysctlImage.resources |
sysctlImage resources | {} |
schedulerName |
Alternate scheduler name | nil |
tls.secretName |
The name of secret if you want to use your own TLS certificates. The secret should contains keys named by "tls.certFile" - the certificate, "tls.keyFile" - the private key, "tls.caCertFile" - the certificate of CA and "tls.dhParamsFile" - the dh parameter file | `` |
tls.certFile |
Name of certificate file | redis.crt |
tls.keyFile |
Name of key file | redis.key |
tls.dhParamsFile |
Name of Diffie-Hellman (DH) key exchange parameters file | `` |
tls.caCertFile |
Name of CA certificate file | ca.crt |
restore.s3.source |
Restore init container - AWS S3 location of dump - i.e. s3://bucket/dump.rdb | false |
restore.existingSecret |
Set to true to use existingSecret for the AWS S3 or SSH credentials | false |
topologySpreadConstraints.enabled |
Enable topology spread constraints | false |
topologySpreadConstraints.maxSkew |
Max skew of pods tolerated | 1 |
topologySpreadConstraints.topologyKey |
Topology key for spread | topology.kubernetes.io/zone |
topologySpreadConstraints.whenUnsatisfiable |
Enforcement policy, hard or soft | ScheduleAnyway |
restore.s3.access_key |
Restore init container - AWS AWS_ACCESS_KEY_ID to access restore.s3.source | `` |
restore.s3.secret_key |
Restore init container - AWS AWS_SECRET_ACCESS_KEY to access restore.s3.source | `` |
restore.s3.region |
Restore init container - AWS AWS_REGION to access restore.s3.source | `` |
restore.ssh.source |
Restore init container - SSH scp location of dump - i.e. user@server:/path/dump.rdb | false |
restore.ssh.key |
Restore init container - SSH private key to scp restore.ssh.source to init container. Key should be in one line separated with \n. i.e. -----BEGIN RSA PRIVATE KEY-----\n...\n...\n-----END RSA PRIVATE KEY----- | `` |
extraContainers |
Extra containers to include in StatefulSet | [] |
extraInitContainers |
Extra init containers to include in StatefulSet | [] |
extraVolumes |
Extra volumes to include in StatefulSet | [] |
extraLabels |
Labels that should be applied to all created resources | {} |
networkPolicy.enabled |
Create NetworkPolicy for Redis StatefulSet pods | false |
networkPolicy.labels |
Labels for NetworkPolicy | {} |
networkPolicy.annotations |
Annotations for NetworkPolicy | {} |
networkPolicy.ingressRules[].selectors |
Label selector query to define resources for this ingress rule | [] |
networkPolicy.ingressRules[].ports |
The destination ports for the ingress rule | [{port: redis.port, protocol: TCP}, {port: sentinel.port, protocol: TCP}] |
networkPolicy.egressRules[].selectors |
Label selector query to define resources for this egress rule | [] |
networkPolicy.egressRules[].ports |
The destination ports for the egress rule | `` |
splitBrainDetection.interval |
Interval between redis sentinel and server split brain checks (in seconds) | 60 |
splitBrainDetection.resources |
splitBrainDetection resources | {} |
Specify each parameter using the --set key=value[,key=value]
argument to helm install
. For example,
$ helm repo add dandydev https://dandydeveloper.github.io/charts
$ helm install \
--set image=redis \
--set tag=5.0.5-alpine \
dandydev/redis-ha
The above command sets the Redis server within default
namespace.
Alternatively, a YAML file that specifies the values for the parameters can be provided while installing the chart. For example,
helm install -f values.yaml dandydev/redis-ha
Tip: You can use the default values.yaml
Custom Redis and Sentinel config options
This chart allows for most redis or sentinel config options to be passed as a key value pair through the values.yaml
under redis.config
and sentinel.config
. See links below for all available options.
Example redis.conf Example sentinel.conf
For example repl-timeout 60
would be added to the redis.config
section of the values.yaml
as:
repl-timeout: "60"
Note:
-
Some config options should be renamed by redis version,e.g.:
# In redis 5.x,see https://raw.githubusercontent.com/antirez/redis/5.0/redis.conf min-replicas-to-write: 1 min-replicas-max-lag: 5 # In redis 4.x and redis 3.x,see https://raw.githubusercontent.com/antirez/redis/4.0/redis.conf and https://raw.githubusercontent.com/antirez/redis/3.0/redis.conf min-slaves-to-write 1 min-slaves-max-lag 5
Sentinel options supported must be in the the sentinel <option> <master-group-name> <value>
format. For example, sentinel down-after-milliseconds 30000
would be added to the sentinel.config
section of the values.yaml
as:
down-after-milliseconds: 30000
If more control is needed from either the redis or sentinel config then an entire config can be defined under redis.customConfig
or sentinel.customConfig
. Please note that these values will override any configuration options under their respective section. For example, if you define sentinel.customConfig
then the sentinel.config
is ignored.
Host Kernel Settings
Redis may require some changes in the kernel of the host machine to work as expected, in particular increasing the somaxconn
value and disabling transparent huge pages.
To do so, you can set up a privileged initContainer with the sysctlImage
config values, for example:
sysctlImage:
enabled: true
mountHostSys: true
command:
- /bin/sh
- -xc
- |-
sysctl -w net.core.somaxconn=10000
echo never > /host-sys/kernel/mm/transparent_hugepage/enabled
HAProxy startup
When HAProxy is enabled, it will attempt to connect to each announce-service of each redis replica instance in its init container before starting.
It will fail if announce-service IP is not available fast enough (10 seconds max by announce-service).
A such case could happen if the orchestator is pending the nomination of redis pods.
Risk is limited because announce-service is using publishNotReadyAddresses: true
, although, in such case, HAProxy pod will be rescheduled afterward by the orchestrator.
PodDisruptionBudgets are not configured by default, you may need to set the haproxy.podDisruptionBudget
parameter in values.yaml to enable it.
Network policies
If networkPolicy.enabled
is set to true
, then a NetworkPolicy
resource is created with default rules to allow inter-Redis and Sentinel connectivity.
This is a requirement for Redis Pods to come up successfully.
You will need to define ingressRules
to permit your application connectivity to Redis.
The selectors
block should be in the format of a label selector.
Templating is also supported in the selectors.
See such a configuration below.
networkPolicy: true
ingressRules:
- selectors:
- namespaceSelector:
matchLabels:
name: my-redis-client-namespace
podSelector:
matchLabels:
# template example
app: |-
{{- .App.Name }}
## ports block is optional (defaults to below), define the block to override the defaults
# ports:
# - port: 6379
# protocol: TCP
# - port: 26379
# protocol: TCP
Should your Pod require additional egress rules, define them in a egressRules
key which is structured identically to an ingressRules
key.
Sentinel and redis server split brain detection
Under not entirely known yet circumstances redis sentinel and its corresponding redis server reach a condition that this chart authors call "split brain" (for short). The observed behaviour is the following: the sentinel switches to the new re-elected master, but does not switch its redis server. Majority of original discussion on the problem has happened at the https://github.com/DandyDeveloper/charts/issues/121.
The proposed solution is currently implemented as a sidecar container that runs a bash script with the following logic:
- Every
splitBrainDetection.interval
seconds a master (as known by sentinel) is determined - If it is the current node: ensure the redis server's role is master as well.
- If it is not the current node: ensure the redis server also replicates from the same node.
If any of the checks above fails - the redis server reinitialisation happens (it regenerates configs the same way it's done during the pod init), and then the redis server is instructed to shutdown. Then kubernetes restarts the container immediately.
Change Log
4.14.9 - ** POTENTIAL BREAKING CHANGE. **
Introduced the ability to change the Haproxy Deployment container pod
- Container port in redis-haproxy-deployment.yam has been changed. Was redis.port To haproxy.containerPort. Default value is 6379.
- Port in redis-haproxy-service.yaml has been changed. Was redis.port To haproxy.servicePort. Default value is 6379.
4.21.0 - BREAKING CHANGES (Kubernetes Deprecation)
This version introduced the deprecation of the PSP and subsequently added fields to the securityContexts that were introduced in Kubernetes v1.19:
https://kubernetes.io/docs/tutorials/security/seccomp/
As a result, from this version onwards Kubernetes versions older than 1.19 will fail to install without the removal of .Values.containerSeucrityContext.seccompProfile
and .Values.haproxy.containerSeucrityContext.seccompProfile
(If HAProxy is enabled)