mirror of https://git.rancher.io/charts
(dev-v2.6-archive) Update rancher-monitoring to new base c6208979d494156a3869d2e5faab669ce4301c68, upstream chart kube-prometheus-stack 19.0.3
(partially cherry picked from commit d6f6f208df
)
pull/1680/head
parent
980d7e2009
commit
ac13d0fd26
|
@ -1,12 +1,12 @@
|
|||
dependencies:
|
||||
- name: kube-state-metrics
|
||||
repository: https://prometheus-community.github.io/helm-charts
|
||||
version: 3.4.2
|
||||
version: 3.5.2
|
||||
- name: prometheus-node-exporter
|
||||
repository: https://prometheus-community.github.io/helm-charts
|
||||
version: 2.0.4
|
||||
- name: grafana
|
||||
repository: https://grafana.github.io/helm-charts
|
||||
version: 6.16.9
|
||||
digest: sha256:cb65f7087c26bba381fdf1478dd8d710b5257855b5c9de67a48997a932112838
|
||||
generated: "2021-09-23T16:19:49.942745722+03:00"
|
||||
version: 6.16.10
|
||||
digest: sha256:94dad976ca1630e9e3cd006fadb255783387b53bd9d0d19e105bd39d8e8e34be
|
||||
generated: "2021-09-28T10:26:46.319411+07:00"
|
||||
|
|
|
@ -0,0 +1,97 @@
|
|||
suite: test ingress
|
||||
templates:
|
||||
- alertmanager/ingress.yaml
|
||||
tests:
|
||||
- it: should be empty if alertmanager is not enabled
|
||||
set:
|
||||
alertmanager.enabled: false
|
||||
alertmanager.ingress.enabled: true
|
||||
asserts:
|
||||
- hasDocuments:
|
||||
count: 0
|
||||
- it: should be empty if ingress is not enabled
|
||||
set:
|
||||
alertmanager.enabled: true
|
||||
alertmanager.ingress.enabled: false
|
||||
asserts:
|
||||
- hasDocuments:
|
||||
count: 0
|
||||
- it: should have apiVersion extensions/v1beta1 for k8s < 1.16
|
||||
set:
|
||||
alertmanager.enabled: true
|
||||
alertmanager.ingress.enabled: true
|
||||
capabilities:
|
||||
majorVersion: 1
|
||||
minorVersion: 15
|
||||
asserts:
|
||||
- hasDocuments:
|
||||
count: 1
|
||||
- isKind:
|
||||
of: Ingress
|
||||
- isAPIVersion:
|
||||
of: extensions/v1beta1
|
||||
- it: should have apiVersion networking.k8s.io/v1beta1 for k8s >= 1.16 < 1.19
|
||||
set:
|
||||
alertmanager.enabled: true
|
||||
alertmanager.ingress.enabled: true
|
||||
capabilities:
|
||||
majorVersion: 1
|
||||
minorVersion: 16
|
||||
apiVersions:
|
||||
- networking.k8s.io/v1beta1
|
||||
asserts:
|
||||
- hasDocuments:
|
||||
count: 1
|
||||
- isKind:
|
||||
of: Ingress
|
||||
- isAPIVersion:
|
||||
of: networking.k8s.io/v1beta1
|
||||
- it: should have apiVersion networking.k8s.io/v1 for k8s >= 1.19 < 1.22
|
||||
set:
|
||||
alertmanager.enabled: true
|
||||
alertmanager.ingress.enabled: true
|
||||
capabilities:
|
||||
majorVersion: 1
|
||||
minorVersion: 10
|
||||
apiVersions:
|
||||
- networking.k8s.io/v1
|
||||
- networking.k8s.io/v1beta1
|
||||
asserts:
|
||||
- hasDocuments:
|
||||
count: 1
|
||||
- isKind:
|
||||
of: Ingress
|
||||
- isAPIVersion:
|
||||
of: networking.k8s.io/v1beta1
|
||||
- it: should have apiVersion networking.k8s.io/v1 for k8s >= 1.22
|
||||
set:
|
||||
alertmanager.enabled: true
|
||||
alertmanager.ingress.enabled: true
|
||||
capabilities:
|
||||
majorVersion: 1
|
||||
minorVersion: 22
|
||||
apiVersions:
|
||||
- networking.k8s.io/v1
|
||||
asserts:
|
||||
- hasDocuments:
|
||||
count: 1
|
||||
- isKind:
|
||||
of: Ingress
|
||||
- isAPIVersion:
|
||||
of: networking.k8s.io/v1
|
||||
|
||||
- it: should have explicit pathType ImplementationSpecific for networking.k8s.io/v1 by default
|
||||
set:
|
||||
alertmanager.enabled: true
|
||||
alertmanager.ingress.enabled: true
|
||||
capabilities:
|
||||
majorVersion: 1
|
||||
minorVersion: 19
|
||||
apiVersions:
|
||||
- networking.k8s.io/v1
|
||||
asserts:
|
||||
- hasDocuments:
|
||||
count: 1
|
||||
- equal:
|
||||
path: spec.rules[0].http.paths[0].pathType
|
||||
value: ImplementationSpecific
|
|
@ -1,12 +0,0 @@
|
|||
{{- if .Values.prometheus.prometheusSpec.thanos.secretProviderClass }}
|
||||
---
|
||||
apiVersion: secrets-store.csi.x-k8s.io/v1alpha1
|
||||
kind: SecretProviderClass
|
||||
metadata:
|
||||
name: {{ template "kube-prometheus-stack.fullname" . }}-prometheus
|
||||
namespace: {{ template "kube-prometheus-stack.namespace" . }}
|
||||
labels:
|
||||
app: {{ template "kube-prometheus-stack.name" . }}-prometheus
|
||||
spec:
|
||||
{{ toYaml .Values.prometheus.prometheusSpec.thanos.secretProviderClass | indent 2 }}
|
||||
{{- end }}
|
|
@ -90,4 +90,4 @@
|
|||
-- https://github.com/prometheus-community/helm-charts
|
||||
-- https://github.com/prometheus-operator/kube-prometheus
|
||||
-type: application
|
||||
-version: 19.0.0
|
||||
-version: 19.0.3
|
||||
|
|
|
@ -9,11 +9,3 @@
|
|||
labels:
|
||||
app: {{ template "kube-prometheus-stack.name" . }}-coredns
|
||||
{{ include "kube-prometheus-stack.labels" . | indent 4 }}
|
||||
@@ -31,6 +31,6 @@
|
||||
{{- end }}
|
||||
{{- if .Values.coreDns.serviceMonitor.relabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.coreDns.serviceMonitor.relabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.coreDns.serviceMonitor.relabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- end }}
|
||||
|
|
|
@ -9,12 +9,3 @@
|
|||
labels:
|
||||
app: {{ template "kube-prometheus-stack.name" . }}-apiserver
|
||||
{{ include "kube-prometheus-stack.labels" . | indent 4 }}
|
||||
@@ -24,7 +24,7 @@
|
||||
{{- end }}
|
||||
{{- if .Values.kubeApiServer.serviceMonitor.relabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.kubeApiServer.serviceMonitor.relabelings | indent 6 }}
|
||||
+{{ tpl (toYaml .Values.kubeApiServer.serviceMonitor.relabelings | indent 6) . }}
|
||||
{{- end }}
|
||||
tlsConfig:
|
||||
caFile: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt
|
||||
|
|
|
@ -9,11 +9,3 @@
|
|||
labels:
|
||||
app: {{ template "kube-prometheus-stack.name" . }}-kube-controller-manager
|
||||
{{ include "kube-prometheus-stack.labels" . | indent 4 }}
|
||||
@@ -42,6 +42,6 @@
|
||||
{{- end }}
|
||||
{{- if .Values.kubeControllerManager.serviceMonitor.relabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.kubeControllerManager.serviceMonitor.relabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.kubeControllerManager.serviceMonitor.relabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- end }}
|
||||
|
|
|
@ -9,11 +9,3 @@
|
|||
labels:
|
||||
app: {{ template "kube-prometheus-stack.name" . }}-kube-dns
|
||||
{{ include "kube-prometheus-stack.labels" . | indent 4 }}
|
||||
@@ -44,6 +44,6 @@
|
||||
{{- end }}
|
||||
{{- if .Values.kubeDns.serviceMonitor.relabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.kubeDns.serviceMonitor.relabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.kubeDns.serviceMonitor.relabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- end }}
|
||||
|
|
|
@ -1,10 +0,0 @@
|
|||
--- charts-original/templates/exporters/kube-etcd/servicemonitor.yaml
|
||||
+++ charts/templates/exporters/kube-etcd/servicemonitor.yaml
|
||||
@@ -48,6 +48,6 @@
|
||||
{{- end }}
|
||||
{{- if .Values.kubeEtcd.serviceMonitor.relabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.kubeEtcd.serviceMonitor.relabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.kubeEtcd.serviceMonitor.relabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- end }}
|
|
@ -1,15 +0,0 @@
|
|||
--- charts-original/templates/exporters/kube-proxy/servicemonitor.yaml
|
||||
+++ charts/templates/exporters/kube-proxy/servicemonitor.yaml
|
||||
@@ -32,10 +32,10 @@
|
||||
{{- end}}
|
||||
{{- if .Values.kubeProxy.serviceMonitor.metricRelabelings }}
|
||||
metricRelabelings:
|
||||
-{{ toYaml .Values.kubeProxy.serviceMonitor.metricRelabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.kubeProxy.serviceMonitor.metricRelabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- if .Values.kubeProxy.serviceMonitor.relabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.kubeProxy.serviceMonitor.relabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.kubeProxy.serviceMonitor.relabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- end }}
|
|
@ -1,10 +0,0 @@
|
|||
--- charts-original/templates/exporters/kube-scheduler/servicemonitor.yaml
|
||||
+++ charts/templates/exporters/kube-scheduler/servicemonitor.yaml
|
||||
@@ -42,6 +42,6 @@
|
||||
{{- end }}
|
||||
{{- if .Values.kubeScheduler.serviceMonitor.relabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.kubeScheduler.serviceMonitor.relabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.kubeScheduler.serviceMonitor.relabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- end }}
|
|
@ -1,29 +0,0 @@
|
|||
--- charts-original/templates/exporters/kube-state-metrics/serviceMonitor.yaml
|
||||
+++ charts/templates/exporters/kube-state-metrics/serviceMonitor.yaml
|
||||
@@ -10,7 +10,7 @@
|
||||
spec:
|
||||
jobLabel: app.kubernetes.io/name
|
||||
endpoints:
|
||||
- - port: http
|
||||
+ - port: metrics
|
||||
{{- if .Values.kubeStateMetrics.serviceMonitor.interval }}
|
||||
interval: {{ .Values.kubeStateMetrics.serviceMonitor.interval }}
|
||||
{{- end }}
|
||||
@@ -27,7 +27,7 @@
|
||||
{{- end }}
|
||||
{{- if .Values.kubeStateMetrics.serviceMonitor.relabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.kubeStateMetrics.serviceMonitor.relabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.kubeStateMetrics.serviceMonitor.relabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- if .Values.kubeStateMetrics.serviceMonitor.selfMonitor.enabled }}
|
||||
- port: metrics
|
||||
@@ -44,7 +44,7 @@
|
||||
{{- end }}
|
||||
{{- if .Values.kubeStateMetrics.serviceMonitor.relabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.kubeStateMetrics.serviceMonitor.relabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.kubeStateMetrics.serviceMonitor.relabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- end }}
|
||||
namespaceSelector:
|
|
@ -15,32 +15,7 @@
|
|||
labels:
|
||||
app: {{ template "kube-prometheus-stack.name" . }}-kubelet
|
||||
{{- include "kube-prometheus-stack.labels" . | indent 4 }}
|
||||
@@ -29,7 +32,7 @@
|
||||
{{- end }}
|
||||
{{- if .Values.kubelet.serviceMonitor.relabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.kubelet.serviceMonitor.relabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.kubelet.serviceMonitor.relabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- if .Values.kubelet.serviceMonitor.cAdvisor }}
|
||||
- port: https-metrics
|
||||
@@ -55,7 +58,7 @@
|
||||
{{- end }}
|
||||
{{- if .Values.kubelet.serviceMonitor.cAdvisorRelabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.kubelet.serviceMonitor.cAdvisorRelabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.kubelet.serviceMonitor.cAdvisorRelabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- end }}
|
||||
{{- if .Values.kubelet.serviceMonitor.probes }}
|
||||
@@ -79,13 +82,13 @@
|
||||
{{- end }}
|
||||
{{- if .Values.kubelet.serviceMonitor.probesRelabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.kubelet.serviceMonitor.probesRelabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.kubelet.serviceMonitor.probesRelabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- end }}
|
||||
@@ -85,7 +88,7 @@
|
||||
{{- if .Values.kubelet.serviceMonitor.resource }}
|
||||
- port: https-metrics
|
||||
scheme: https
|
||||
|
@ -49,30 +24,7 @@
|
|||
{{- if .Values.kubelet.serviceMonitor.interval }}
|
||||
interval: {{ .Values.kubelet.serviceMonitor.interval }}
|
||||
{{- end }}
|
||||
@@ -103,7 +106,7 @@
|
||||
{{- end }}
|
||||
{{- if .Values.kubelet.serviceMonitor.resourceRelabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.kubelet.serviceMonitor.resourceRelabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.kubelet.serviceMonitor.resourceRelabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- end }}
|
||||
{{- else }}
|
||||
@@ -121,7 +124,7 @@
|
||||
{{- end }}
|
||||
{{- if .Values.kubelet.serviceMonitor.relabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.kubelet.serviceMonitor.relabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.kubelet.serviceMonitor.relabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- if .Values.kubelet.serviceMonitor.cAdvisor }}
|
||||
- port: http-metrics
|
||||
@@ -139,11 +142,11 @@
|
||||
{{- end }}
|
||||
{{- if .Values.kubelet.serviceMonitor.cAdvisorRelabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.kubelet.serviceMonitor.cAdvisorRelabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.kubelet.serviceMonitor.cAdvisorRelabelings | indent 4) . }}
|
||||
@@ -143,7 +146,7 @@
|
||||
{{- end }}
|
||||
{{- if .Values.kubelet.serviceMonitor.resource }}
|
||||
- port: http-metrics
|
||||
|
@ -81,12 +33,3 @@
|
|||
{{- if .Values.kubelet.serviceMonitor.interval }}
|
||||
interval: {{ .Values.kubelet.serviceMonitor.interval }}
|
||||
{{- end }}
|
||||
@@ -157,7 +160,7 @@
|
||||
{{- end }}
|
||||
{{- if .Values.kubelet.serviceMonitor.resourceRelabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.kubelet.serviceMonitor.resourceRelabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.kubelet.serviceMonitor.resourceRelabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- end }}
|
||||
{{- end }}
|
||||
|
|
|
@ -9,11 +9,3 @@
|
|||
apiVersion: monitoring.coreos.com/v1
|
||||
kind: ServiceMonitor
|
||||
metadata:
|
||||
@@ -33,6 +36,6 @@
|
||||
{{- end }}
|
||||
{{- if .Values.nodeExporter.serviceMonitor.relabelings }}
|
||||
relabelings:
|
||||
-{{ toYaml .Values.nodeExporter.serviceMonitor.relabelings | indent 4 }}
|
||||
+{{ tpl (toYaml .Values.nodeExporter.serviceMonitor.relabelings | indent 4) . }}
|
||||
{{- end }}
|
||||
{{- end }}
|
||||
|
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/configmap-dashboards.yaml
|
||||
+++ charts/templates/grafana/configmap-dashboards.yaml
|
||||
@@ -10,7 +10,7 @@
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) $dashboardName | trunc 63 | trimSuffix "-" }}
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" $ }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
labels:
|
||||
{{- if $.Values.grafana.sidecar.dashboards.label }}
|
||||
{{ $.Values.grafana.sidecar.dashboards.label }}: "1"
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/configmaps-datasources.yaml
|
||||
+++ charts/templates/grafana/configmaps-datasources.yaml
|
||||
@@ -3,7 +3,7 @@
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
name: {{ template "kube-prometheus-stack.fullname" . }}-grafana-datasource
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ default .Values.grafana.sidecar.datasources.searchNamespace (include "kube-prometheus-stack.namespace" .) }}
|
||||
{{- if .Values.grafana.sidecar.datasources.annotations }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.datasources.annotations | indent 4 }}
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/apiserver.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/apiserver.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "apiserver" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/cluster-total.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/cluster-total.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "cluster-total" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -1,6 +1,6 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/controller-manager.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/controller-manager.yaml
|
||||
@@ -4,7 +4,8 @@
|
||||
@@ -4,11 +4,12 @@
|
||||
https://github.com/prometheus-community/helm-charts/tree/main/charts/kube-prometheus-stack/hack
|
||||
*/ -}}
|
||||
{{- $kubeTargetVersion := default .Capabilities.KubeVersion.GitVersion .Values.kubeTargetVersionOverride }}
|
||||
|
@ -10,6 +10,11 @@
|
|||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "controller-manager" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
||||
@@ -107,7 +108,11 @@
|
||||
"tableColumn": "",
|
||||
"targets": [
|
||||
|
|
|
@ -1,6 +1,6 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/etcd.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/etcd.yaml
|
||||
@@ -4,7 +4,8 @@
|
||||
@@ -4,11 +4,12 @@
|
||||
https://github.com/prometheus-community/helm-charts/tree/main/charts/kube-prometheus-stack/hack
|
||||
*/ -}}
|
||||
{{- $kubeTargetVersion := default .Capabilities.KubeVersion.GitVersion .Values.kubeTargetVersionOverride }}
|
||||
|
@ -10,6 +10,11 @@
|
|||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "etcd" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
||||
@@ -1113,4 +1114,5 @@
|
||||
"title": "etcd",
|
||||
"version": 215
|
||||
|
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/k8s-coredns.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/k8s-coredns.yaml
|
||||
@@ -4,7 +4,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "k8s-coredns" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/k8s-resources-cluster.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/k8s-resources-cluster.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "k8s-resources-cluster" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/k8s-resources-namespace.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/k8s-resources-namespace.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "k8s-resources-namespace" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/k8s-resources-node.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/k8s-resources-node.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "k8s-resources-node" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/k8s-resources-pod.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/k8s-resources-pod.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "k8s-resources-pod" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/k8s-resources-workload.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/k8s-resources-workload.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "k8s-resources-workload" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/k8s-resources-workloads-namespace.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/k8s-resources-workloads-namespace.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "k8s-resources-workloads-namespace" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -1,6 +1,6 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/kubelet.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/kubelet.yaml
|
||||
@@ -4,7 +4,8 @@
|
||||
@@ -4,11 +4,12 @@
|
||||
https://github.com/prometheus-community/helm-charts/tree/main/charts/kube-prometheus-stack/hack
|
||||
*/ -}}
|
||||
{{- $kubeTargetVersion := default .Capabilities.KubeVersion.GitVersion .Values.kubeTargetVersionOverride }}
|
||||
|
@ -10,6 +10,276 @@
|
|||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "kubelet" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
||||
@@ -87,7 +88,7 @@
|
||||
"pluginVersion": "7",
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "sum(kubelet_node_name{cluster=\"$cluster\", job=\"kubelet\", metrics_path=\"/metrics\"})",
|
||||
+ "expr": "sum(kubelet_node_name{cluster=\"$cluster\", job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\"})",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "",
|
||||
@@ -144,7 +145,7 @@
|
||||
"pluginVersion": "7",
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "sum(kubelet_running_pods{cluster=\"$cluster\", job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\"}) OR sum(kubelet_running_pod_count{cluster=\"$cluster\", job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\"})",
|
||||
+ "expr": "sum(kubelet_running_pods{cluster=\"$cluster\", job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\"}) OR sum(kubelet_running_pod_count{cluster=\"$cluster\", job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\"})",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}}",
|
||||
@@ -201,7 +202,7 @@
|
||||
"pluginVersion": "7",
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "sum(kubelet_running_containers{cluster=\"$cluster\", job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\"}) OR sum(kubelet_running_container_count{cluster=\"$cluster\", job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\"})",
|
||||
+ "expr": "sum(kubelet_running_containers{cluster=\"$cluster\", job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\"}) OR sum(kubelet_running_container_count{cluster=\"$cluster\", job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\"})",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}}",
|
||||
@@ -258,7 +259,7 @@
|
||||
"pluginVersion": "7",
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "sum(volume_manager_total_volumes{cluster=\"$cluster\", job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\", state=\"actual_state_of_world\"})",
|
||||
+ "expr": "sum(volume_manager_total_volumes{cluster=\"$cluster\", job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\", state=\"actual_state_of_world\"})",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}}",
|
||||
@@ -315,7 +316,7 @@
|
||||
"pluginVersion": "7",
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "sum(volume_manager_total_volumes{cluster=\"$cluster\", job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\",state=\"desired_state_of_world\"})",
|
||||
+ "expr": "sum(volume_manager_total_volumes{cluster=\"$cluster\", job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\",state=\"desired_state_of_world\"})",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}}",
|
||||
@@ -372,7 +373,7 @@
|
||||
"pluginVersion": "7",
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "sum(rate(kubelet_node_config_error{cluster=\"$cluster\", job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\"}[5m]))",
|
||||
+ "expr": "sum(rate(kubelet_node_config_error{cluster=\"$cluster\", job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\"}[5m]))",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}}",
|
||||
@@ -431,7 +432,7 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "sum(rate(kubelet_runtime_operations_total{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (operation_type, instance)",
|
||||
+ "expr": "sum(rate(kubelet_runtime_operations_total{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (operation_type, instance)",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}} {{`{{`}}operation_type{{`}}`}}",
|
||||
@@ -526,7 +527,7 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "sum(rate(kubelet_runtime_operations_errors_total{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, operation_type)",
|
||||
+ "expr": "sum(rate(kubelet_runtime_operations_errors_total{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, operation_type)",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}} {{`{{`}}operation_type{{`}}`}}",
|
||||
@@ -621,7 +622,7 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "histogram_quantile(0.99, sum(rate(kubelet_runtime_operations_duration_seconds_bucket{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, operation_type, le))",
|
||||
+ "expr": "histogram_quantile(0.99, sum(rate(kubelet_runtime_operations_duration_seconds_bucket{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, operation_type, le))",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}} {{`{{`}}operation_type{{`}}`}}",
|
||||
@@ -716,14 +717,14 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "sum(rate(kubelet_pod_start_duration_seconds_count{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance)",
|
||||
+ "expr": "sum(rate(kubelet_pod_start_duration_seconds_count{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance)",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}} pod",
|
||||
"refId": "A"
|
||||
},
|
||||
{
|
||||
- "expr": "sum(rate(kubelet_pod_worker_duration_seconds_count{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance)",
|
||||
+ "expr": "sum(rate(kubelet_pod_worker_duration_seconds_count{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance)",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}} worker",
|
||||
@@ -818,14 +819,14 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "histogram_quantile(0.99, sum(rate(kubelet_pod_start_duration_seconds_count{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, le))",
|
||||
+ "expr": "histogram_quantile(0.99, sum(rate(kubelet_pod_start_duration_seconds_count{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, le))",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}} pod",
|
||||
"refId": "A"
|
||||
},
|
||||
{
|
||||
- "expr": "histogram_quantile(0.99, sum(rate(kubelet_pod_worker_duration_seconds_bucket{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, le))",
|
||||
+ "expr": "histogram_quantile(0.99, sum(rate(kubelet_pod_worker_duration_seconds_bucket{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, le))",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}} worker",
|
||||
@@ -922,7 +923,7 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "sum(rate(storage_operation_duration_seconds_count{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, operation_name, volume_plugin)",
|
||||
+ "expr": "sum(rate(storage_operation_duration_seconds_count{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, operation_name, volume_plugin)",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}} {{`{{`}}operation_name{{`}}`}} {{`{{`}}volume_plugin{{`}}`}}",
|
||||
@@ -1019,7 +1020,7 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "sum(rate(storage_operation_errors_total{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, operation_name, volume_plugin)",
|
||||
+ "expr": "sum(rate(storage_operation_errors_total{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, operation_name, volume_plugin)",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}} {{`{{`}}operation_name{{`}}`}} {{`{{`}}volume_plugin{{`}}`}}",
|
||||
@@ -1116,7 +1117,7 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "histogram_quantile(0.99, sum(rate(storage_operation_duration_seconds_bucket{cluster=\"$cluster\", job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\"}[5m])) by (instance, operation_name, volume_plugin, le))",
|
||||
+ "expr": "histogram_quantile(0.99, sum(rate(storage_operation_duration_seconds_bucket{cluster=\"$cluster\", job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\"}[5m])) by (instance, operation_name, volume_plugin, le))",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}} {{`{{`}}operation_name{{`}}`}} {{`{{`}}volume_plugin{{`}}`}}",
|
||||
@@ -1211,7 +1212,7 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "sum(rate(kubelet_cgroup_manager_duration_seconds_count{cluster=\"$cluster\", job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\"}[5m])) by (instance, operation_type)",
|
||||
+ "expr": "sum(rate(kubelet_cgroup_manager_duration_seconds_count{cluster=\"$cluster\", job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\"}[5m])) by (instance, operation_type)",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}operation_type{{`}}`}}",
|
||||
@@ -1306,7 +1307,7 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "histogram_quantile(0.99, sum(rate(kubelet_cgroup_manager_duration_seconds_bucket{cluster=\"$cluster\", job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\"}[5m])) by (instance, operation_type, le))",
|
||||
+ "expr": "histogram_quantile(0.99, sum(rate(kubelet_cgroup_manager_duration_seconds_bucket{cluster=\"$cluster\", job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\"}[5m])) by (instance, operation_type, le))",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}} {{`{{`}}operation_type{{`}}`}}",
|
||||
@@ -1402,7 +1403,7 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "sum(rate(kubelet_pleg_relist_duration_seconds_count{cluster=\"$cluster\", job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\"}[5m])) by (instance)",
|
||||
+ "expr": "sum(rate(kubelet_pleg_relist_duration_seconds_count{cluster=\"$cluster\", job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\"}[5m])) by (instance)",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}}",
|
||||
@@ -1497,7 +1498,7 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "histogram_quantile(0.99, sum(rate(kubelet_pleg_relist_interval_seconds_bucket{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, le))",
|
||||
+ "expr": "histogram_quantile(0.99, sum(rate(kubelet_pleg_relist_interval_seconds_bucket{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, le))",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}}",
|
||||
@@ -1592,7 +1593,7 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "histogram_quantile(0.99, sum(rate(kubelet_pleg_relist_duration_seconds_bucket{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, le))",
|
||||
+ "expr": "histogram_quantile(0.99, sum(rate(kubelet_pleg_relist_duration_seconds_bucket{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])) by (instance, le))",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}}",
|
||||
@@ -1687,28 +1688,28 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "sum(rate(rest_client_requests_total{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\",code=~\"2..\"}[5m]))",
|
||||
+ "expr": "sum(rate(rest_client_requests_total{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\",code=~\"2..\"}[5m]))",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "2xx",
|
||||
"refId": "A"
|
||||
},
|
||||
{
|
||||
- "expr": "sum(rate(rest_client_requests_total{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\",code=~\"3..\"}[5m]))",
|
||||
+ "expr": "sum(rate(rest_client_requests_total{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\",code=~\"3..\"}[5m]))",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "3xx",
|
||||
"refId": "B"
|
||||
},
|
||||
{
|
||||
- "expr": "sum(rate(rest_client_requests_total{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\",code=~\"4..\"}[5m]))",
|
||||
+ "expr": "sum(rate(rest_client_requests_total{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\",code=~\"4..\"}[5m]))",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "4xx",
|
||||
"refId": "C"
|
||||
},
|
||||
{
|
||||
- "expr": "sum(rate(rest_client_requests_total{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\",code=~\"5..\"}[5m]))",
|
||||
+ "expr": "sum(rate(rest_client_requests_total{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\",code=~\"5..\"}[5m]))",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "5xx",
|
||||
@@ -1803,7 +1804,7 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "histogram_quantile(0.99, sum(rate(rest_client_request_duration_seconds_bucket{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\", instance=~\"$instance\"}[5m])) by (instance, verb, url, le))",
|
||||
+ "expr": "histogram_quantile(0.99, sum(rate(rest_client_request_duration_seconds_bucket{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\", instance=~\"$instance\"}[5m])) by (instance, verb, url, le))",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}} {{`{{`}}verb{{`}}`}} {{`{{`}}url{{`}}`}}",
|
||||
@@ -1898,7 +1899,7 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "process_resident_memory_bytes{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\",instance=~\"$instance\"}",
|
||||
+ "expr": "process_resident_memory_bytes{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\",instance=~\"$instance\"}",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}}",
|
||||
@@ -1993,7 +1994,7 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "rate(process_cpu_seconds_total{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])",
|
||||
+ "expr": "rate(process_cpu_seconds_total{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\",instance=~\"$instance\"}[5m])",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}}",
|
||||
@@ -2088,7 +2089,7 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
{
|
||||
- "expr": "go_goroutines{cluster=\"$cluster\",job=\"kubelet\", metrics_path=\"/metrics\",instance=~\"$instance\"}",
|
||||
+ "expr": "go_goroutines{cluster=\"$cluster\",job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\",instance=~\"$instance\"}",
|
||||
"format": "time_series",
|
||||
"intervalFactor": 2,
|
||||
"legendFormat": "{{`{{`}}instance{{`}}`}}",
|
||||
@@ -2177,7 +2178,7 @@
|
||||
"options": [
|
||||
|
||||
],
|
||||
- "query": "label_values(up{job=\"kubelet\", metrics_path=\"/metrics\"}, cluster)",
|
||||
+ "query": "label_values(up{job=\"{{ include "exporter.kubelet.jobName" . }}\", metrics_path=\"/metrics\"}, cluster)",
|
||||
"refresh": 2,
|
||||
"regex": "",
|
||||
"sort": 1,
|
||||
@@ -2203,7 +2204,7 @@
|
||||
"options": [
|
||||
|
||||
|
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/namespace-by-pod.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/namespace-by-pod.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "namespace-by-pod" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/namespace-by-workload.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/namespace-by-workload.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "namespace-by-workload" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/node-cluster-rsrc-use.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/node-cluster-rsrc-use.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "node-cluster-rsrc-use" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/node-rsrc-use.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/node-rsrc-use.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "node-rsrc-use" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/nodes.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/nodes.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "nodes" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -1,5 +1,14 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/persistentvolumesusage.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/persistentvolumesusage.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "persistentvolumesusage" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
||||
@@ -92,14 +92,14 @@
|
||||
"steppedLine": false,
|
||||
"targets": [
|
||||
|
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/pod-total.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/pod-total.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "pod-total" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/prometheus-remote-write.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/prometheus-remote-write.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "prometheus-remote-write" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/prometheus.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/prometheus.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "prometheus" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -1,6 +1,6 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/proxy.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/proxy.yaml
|
||||
@@ -4,7 +4,8 @@
|
||||
@@ -4,11 +4,12 @@
|
||||
https://github.com/prometheus-community/helm-charts/tree/main/charts/kube-prometheus-stack/hack
|
||||
*/ -}}
|
||||
{{- $kubeTargetVersion := default .Capabilities.KubeVersion.GitVersion .Values.kubeTargetVersionOverride }}
|
||||
|
@ -10,6 +10,11 @@
|
|||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "proxy" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
||||
@@ -107,7 +108,11 @@
|
||||
"tableColumn": "",
|
||||
"targets": [
|
||||
|
|
|
@ -1,6 +1,6 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/scheduler.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/scheduler.yaml
|
||||
@@ -4,7 +4,8 @@
|
||||
@@ -4,11 +4,12 @@
|
||||
https://github.com/prometheus-community/helm-charts/tree/main/charts/kube-prometheus-stack/hack
|
||||
*/ -}}
|
||||
{{- $kubeTargetVersion := default .Capabilities.KubeVersion.GitVersion .Values.kubeTargetVersionOverride }}
|
||||
|
@ -10,6 +10,11 @@
|
|||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "scheduler" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
||||
@@ -107,7 +108,11 @@
|
||||
"tableColumn": "",
|
||||
"targets": [
|
||||
|
|
|
@ -1,5 +1,14 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/statefulset.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/statefulset.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "statefulset" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
||||
@@ -106,7 +106,7 @@
|
||||
"tableColumn": "",
|
||||
"targets": [
|
||||
|
|
|
@ -0,0 +1,11 @@
|
|||
--- charts-original/templates/grafana/dashboards-1.14/workload-total.yaml
|
||||
+++ charts/templates/grafana/dashboards-1.14/workload-total.yaml
|
||||
@@ -8,7 +8,7 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
- namespace: {{ template "kube-prometheus-stack-grafana.namespace" . }}
|
||||
+ namespace: {{ .Values.grafana.defaultDashboards.namespace }}
|
||||
name: {{ printf "%s-%s" (include "kube-prometheus-stack.fullname" $) "workload-total" | trunc 63 | trimSuffix "-" }}
|
||||
annotations:
|
||||
{{ toYaml .Values.grafana.sidecar.dashboards.annotations | indent 4 }}
|
|
@ -1,5 +1,23 @@
|
|||
--- charts-original/templates/prometheus/prometheus.yaml
|
||||
+++ charts/templates/prometheus/prometheus.yaml
|
||||
@@ -33,13 +33,13 @@
|
||||
{{- end }}
|
||||
{{- if .Values.prometheus.prometheusSpec.image }}
|
||||
{{- if and .Values.prometheus.prometheusSpec.image.tag .Values.prometheus.prometheusSpec.image.sha }}
|
||||
- image: "{{ .Values.prometheus.prometheusSpec.image.repository }}:{{ .Values.prometheus.prometheusSpec.image.tag }}@sha256:{{ .Values.prometheus.prometheusSpec.image.sha }}"
|
||||
+ image: "{{ template "system_default_registry" . }}{{ .Values.prometheus.prometheusSpec.image.repository }}:{{ .Values.prometheus.prometheusSpec.image.tag }}@sha256:{{ .Values.prometheus.prometheusSpec.image.sha }}"
|
||||
{{- else if .Values.prometheus.prometheusSpec.image.sha }}
|
||||
- image: "{{ .Values.prometheus.prometheusSpec.image.repository }}@sha256:{{ .Values.prometheus.prometheusSpec.image.sha }}"
|
||||
+ image: "{{ template "system_default_registry" . }}{{ .Values.prometheus.prometheusSpec.image.repository }}@sha256:{{ .Values.prometheus.prometheusSpec.image.sha }}"
|
||||
{{- else if .Values.prometheus.prometheusSpec.image.tag }}
|
||||
- image: "{{ .Values.prometheus.prometheusSpec.image.repository }}:{{ .Values.prometheus.prometheusSpec.image.tag }}"
|
||||
+ image: "{{ template "system_default_registry" . }}{{ .Values.prometheus.prometheusSpec.image.repository }}:{{ .Values.prometheus.prometheusSpec.image.tag }}"
|
||||
{{- else }}
|
||||
- image: "{{ .Values.prometheus.prometheusSpec.image.repository }}"
|
||||
+ image: "{{ template "system_default_registry" . }}{{ .Values.prometheus.prometheusSpec.image.repository }}"
|
||||
{{- end }}
|
||||
version: {{ .Values.prometheus.prometheusSpec.image.tag }}
|
||||
{{- if .Values.prometheus.prometheusSpec.image.sha }}
|
||||
@@ -64,11 +64,13 @@
|
||||
externalUrl: "{{ tpl .Values.prometheus.prometheusSpec.externalUrl . }}"
|
||||
{{- else if and .Values.prometheus.ingress.enabled .Values.prometheus.ingress.hosts }}
|
||||
|
|
|
@ -463,34 +463,104 @@
|
|||
pspEnabled: true
|
||||
pspAnnotations: {}
|
||||
## Specify pod annotations
|
||||
@@ -187,6 +625,7 @@
|
||||
@@ -187,25 +625,77 @@
|
||||
## ref: https://prometheus.io/docs/alerting/notifications/
|
||||
## https://prometheus.io/docs/alerting/notification_examples/
|
||||
##
|
||||
- templateFiles: {}
|
||||
- #
|
||||
- ## An example template:
|
||||
- # template_1.tmpl: |-
|
||||
- # {{ define "cluster" }}{{ .ExternalURL | reReplaceAll ".*alertmanager\\.(.*)" "$1" }}{{ end }}
|
||||
- #
|
||||
- # {{ define "slack.myorg.text" }}
|
||||
- # {{- $root := . -}}
|
||||
- # {{ range .Alerts }}
|
||||
- # *Alert:* {{ .Annotations.summary }} - `{{ .Labels.severity }}`
|
||||
- # *Cluster:* {{ template "cluster" $root }}
|
||||
- # *Description:* {{ .Annotations.description }}
|
||||
- # *Graph:* <{{ .GeneratorURL }}|:chart_with_upwards_trend:>
|
||||
- # *Runbook:* <{{ .Annotations.runbook }}|:spiral_note_pad:>
|
||||
- # *Details:*
|
||||
- # {{ range .Labels.SortedPairs }} - *{{ .Name }}:* `{{ .Value }}`
|
||||
- # {{ end }}
|
||||
- # {{ end }}
|
||||
- # {{ end }}
|
||||
+
|
||||
templateFiles: {}
|
||||
#
|
||||
## An example template:
|
||||
@@ -366,14 +805,16 @@
|
||||
+ templateFiles:
|
||||
+ rancher_defaults.tmpl: |-
|
||||
+ {{- define "slack.rancher.text" -}}
|
||||
+ {{ template "rancher.text_multiple" . }}
|
||||
+ {{- end -}}
|
||||
+
|
||||
+ {{- define "rancher.text_multiple" -}}
|
||||
+ *[GROUP - Details]*
|
||||
+ One or more alarms in this group have triggered a notification.
|
||||
+
|
||||
+ {{- if gt (len .GroupLabels.Values) 0 }}
|
||||
+ *Group Labels:*
|
||||
+ {{- range .GroupLabels.SortedPairs }}
|
||||
+ • *{{ .Name }}:* `{{ .Value }}`
|
||||
+ {{- end }}
|
||||
+ {{- end }}
|
||||
+ {{- if .ExternalURL }}
|
||||
+ *Link to AlertManager:* {{ .ExternalURL }}
|
||||
+ {{- end }}
|
||||
+
|
||||
+ {{- range .Alerts }}
|
||||
+ {{ template "rancher.text_single" . }}
|
||||
+ {{- end }}
|
||||
+ {{- end -}}
|
||||
+
|
||||
+ {{- define "rancher.text_single" -}}
|
||||
+ {{- if .Labels.alertname }}
|
||||
+ *[ALERT - {{ .Labels.alertname }}]*
|
||||
+ {{- else }}
|
||||
+ *[ALERT]*
|
||||
+ {{- end }}
|
||||
+ {{- if .Labels.severity }}
|
||||
+ *Severity:* `{{ .Labels.severity }}`
|
||||
+ {{- end }}
|
||||
+ {{- if .Labels.cluster }}
|
||||
+ *Cluster:* {{ .Labels.cluster }}
|
||||
+ {{- end }}
|
||||
+ {{- if .Annotations.summary }}
|
||||
+ *Summary:* {{ .Annotations.summary }}
|
||||
+ {{- end }}
|
||||
+ {{- if .Annotations.message }}
|
||||
+ *Message:* {{ .Annotations.message }}
|
||||
+ {{- end }}
|
||||
+ {{- if .Annotations.description }}
|
||||
+ *Description:* {{ .Annotations.description }}
|
||||
+ {{- end }}
|
||||
+ {{- if .Annotations.runbook_url }}
|
||||
+ *Runbook URL:* <{{ .Annotations.runbook_url }}|:spiral_note_pad:>
|
||||
+ {{- end }}
|
||||
+ {{- with .Labels }}
|
||||
+ {{- with .Remove (stringSlice "alertname" "severity" "cluster") }}
|
||||
+ {{- if gt (len .) 0 }}
|
||||
+ *Additional Labels:*
|
||||
+ {{- range .SortedPairs }}
|
||||
+ • *{{ .Name }}:* `{{ .Value }}`
|
||||
+ {{- end }}
|
||||
+ {{- end }}
|
||||
+ {{- end }}
|
||||
+ {{- end }}
|
||||
+ {{- with .Annotations }}
|
||||
+ {{- with .Remove (stringSlice "summary" "message" "description" "runbook_url") }}
|
||||
+ {{- if gt (len .) 0 }}
|
||||
+ *Additional Annotations:*
|
||||
+ {{- range .SortedPairs }}
|
||||
+ • *{{ .Name }}:* `{{ .Value }}`
|
||||
+ {{- end }}
|
||||
+ {{- end }}
|
||||
+ {{- end }}
|
||||
+ {{- end }}
|
||||
+ {{- end -}}
|
||||
|
||||
bearerTokenFile:
|
||||
|
||||
- ## Metric relabel configs to apply to samples before ingestion.
|
||||
+ ## MetricRelabelConfigs to apply to samples after scraping, but before ingestion.
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
metricRelabelings: []
|
||||
# - action: keep
|
||||
# regex: 'kube_(daemonset|deployment|pod|namespace|node|statefulset).+'
|
||||
# sourceLabels: [__name__]
|
||||
|
||||
- # relabel configs to apply to samples before ingestion.
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
relabelings: []
|
||||
# - sourceLabels: [__meta_kubernetes_pod_node_name]
|
||||
@@ -395,7 +836,7 @@
|
||||
ingress:
|
||||
enabled: false
|
||||
@@ -397,7 +887,7 @@
|
||||
## Image of Alertmanager
|
||||
##
|
||||
image:
|
||||
|
@ -499,7 +569,7 @@
|
|||
tag: v0.22.2
|
||||
sha: ""
|
||||
|
||||
@@ -507,9 +948,13 @@
|
||||
@@ -509,9 +999,13 @@
|
||||
## Define resources requests and limits for single Pods.
|
||||
## ref: https://kubernetes.io/docs/user-guide/compute-resources/
|
||||
##
|
||||
|
@ -516,7 +586,7 @@
|
|||
|
||||
## Pod anti-affinity can prevent the scheduler from placing Prometheus replicas on the same node.
|
||||
## The default value "soft" means that the scheduler should *prefer* to not schedule two replica pods onto the same node but no guarantee is provided.
|
||||
@@ -623,6 +1068,30 @@
|
||||
@@ -625,6 +1119,30 @@
|
||||
enabled: true
|
||||
namespaceOverride: ""
|
||||
|
||||
|
@ -547,7 +617,26 @@
|
|||
## ForceDeployDatasources Create datasource configmap even if grafana deployment has been disabled
|
||||
##
|
||||
forceDeployDatasources: false
|
||||
@@ -679,6 +1148,7 @@
|
||||
@@ -637,6 +1155,18 @@
|
||||
##
|
||||
defaultDashboardsEnabled: true
|
||||
|
||||
+ # Additional options for defaultDashboards
|
||||
+ defaultDashboards:
|
||||
+ # The default namespace to place defaultDashboards within
|
||||
+ namespace: cattle-dashboards
|
||||
+ # Whether to create the default namespace as a Helm managed namespace or use an existing namespace
|
||||
+ # If false, the defaultDashboards.namespace will be created as a Helm managed namespace
|
||||
+ useExistingNamespace: false
|
||||
+ # Whether the Helm managed namespace created by this chart should be left behind on a Helm uninstall
|
||||
+ # If you place other dashboards in this namespace, then they will be deleted on a helm uninstall
|
||||
+ # Ignore if useExistingNamespace is true
|
||||
+ cleanupOnUninstall: false
|
||||
+
|
||||
## Timezone for the default dashboards
|
||||
## Other options are: browser or a specific timezone, i.e. Europe/Luxembourg
|
||||
##
|
||||
@@ -681,6 +1211,7 @@
|
||||
dashboards:
|
||||
enabled: true
|
||||
label: grafana_dashboard
|
||||
|
@ -555,7 +644,7 @@
|
|||
|
||||
## Annotations for Grafana dashboard configmaps
|
||||
##
|
||||
@@ -737,7 +1207,60 @@
|
||||
@@ -739,7 +1270,60 @@
|
||||
## Passed to grafana subchart and used by servicemonitor below
|
||||
##
|
||||
service:
|
||||
|
@ -617,31 +706,10 @@
|
|||
|
||||
## If true, create a serviceMonitor for grafana
|
||||
##
|
||||
@@ -751,14 +1274,17 @@
|
||||
# in grafana.ini
|
||||
path: "/metrics"
|
||||
|
||||
- ## Metric relabel configs to apply to samples before ingestion.
|
||||
+
|
||||
+ ## MetricRelabelConfigs to apply to samples after scraping, but before ingestion.
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
metricRelabelings: []
|
||||
# - action: keep
|
||||
# regex: 'kube_(daemonset|deployment|pod|namespace|node|statefulset).+'
|
||||
# sourceLabels: [__name__]
|
||||
|
||||
- # relabel configs to apply to samples before ingestion.
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
relabelings: []
|
||||
# - sourceLabels: [__meta_kubernetes_pod_node_name]
|
||||
@@ -767,6 +1293,14 @@
|
||||
# targetLabel: nodename
|
||||
@@ -773,6 +1357,14 @@
|
||||
# replacement: $1
|
||||
# action: replace
|
||||
+
|
||||
|
||||
+ resources:
|
||||
+ limits:
|
||||
+ memory: 200Mi
|
||||
|
@ -649,106 +717,11 @@
|
|||
+ requests:
|
||||
+ memory: 100Mi
|
||||
+ cpu: 100m
|
||||
|
||||
+
|
||||
## Component scraping the kube api server
|
||||
##
|
||||
@@ -789,12 +1323,17 @@
|
||||
component: apiserver
|
||||
provider: kubernetes
|
||||
|
||||
- ## Metric relabel configs to apply to samples before ingestion.
|
||||
+ ## MetricRelabelConfigs to apply to samples after scraping, but before ingestion.
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
metricRelabelings: []
|
||||
# - action: keep
|
||||
# regex: 'kube_(daemonset|deployment|pod|namespace|node|statefulset).+'
|
||||
# sourceLabels: [__name__]
|
||||
+
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
+ ##
|
||||
relabelings: []
|
||||
# - sourceLabels:
|
||||
# - __meta_kubernetes_namespace
|
||||
@@ -839,7 +1378,9 @@
|
||||
resource: false
|
||||
# From kubernetes 1.18, /metrics/resource/v1alpha1 renamed to /metrics/resource
|
||||
resourcePath: "/metrics/resource/v1alpha1"
|
||||
- ## Metric relabellings to apply to samples before ingestion
|
||||
+
|
||||
+ ## MetricRelabelConfigs to apply to samples after scraping, but before ingestion.
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
cAdvisorMetricRelabelings: []
|
||||
# - sourceLabels: [__name__, image]
|
||||
@@ -853,7 +1394,8 @@
|
||||
# replacement: $1
|
||||
# action: drop
|
||||
|
||||
- ## Metric relabellings to apply to samples before ingestion
|
||||
+ ## MetricRelabelConfigs to apply to samples after scraping, but before ingestion.
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
probesMetricRelabelings: []
|
||||
# - sourceLabels: [__name__, image]
|
||||
@@ -867,9 +1409,10 @@
|
||||
# replacement: $1
|
||||
# action: drop
|
||||
|
||||
- # relabel configs to apply to samples before ingestion.
|
||||
- # metrics_path is required to match upstream rules and charts
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
+ ## metrics_path is required to match upstream rules and charts
|
||||
cAdvisorRelabelings:
|
||||
- sourceLabels: [__metrics_path__]
|
||||
targetLabel: metrics_path
|
||||
@@ -880,6 +1423,9 @@
|
||||
# replacement: $1
|
||||
# action: replace
|
||||
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
+ ##
|
||||
probesRelabelings:
|
||||
- sourceLabels: [__metrics_path__]
|
||||
targetLabel: metrics_path
|
||||
@@ -890,6 +1436,9 @@
|
||||
# replacement: $1
|
||||
# action: replace
|
||||
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
+ ##
|
||||
resourceRelabelings:
|
||||
- sourceLabels: [__metrics_path__]
|
||||
targetLabel: metrics_path
|
||||
@@ -900,6 +1449,9 @@
|
||||
# replacement: $1
|
||||
# action: replace
|
||||
|
||||
+ ## MetricRelabelConfigs to apply to samples after scraping, but before ingestion.
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
+ ##
|
||||
metricRelabelings: []
|
||||
# - sourceLabels: [__name__, image]
|
||||
# separator: ;
|
||||
@@ -912,9 +1464,10 @@
|
||||
# replacement: $1
|
||||
# action: drop
|
||||
|
||||
- # relabel configs to apply to samples before ingestion.
|
||||
- # metrics_path is required to match upstream rules and charts
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
+ ## metrics_path is required to match upstream rules and charts
|
||||
relabelings:
|
||||
- sourceLabels: [__metrics_path__]
|
||||
targetLabel: metrics_path
|
||||
@@ -928,7 +1481,7 @@
|
||||
kubeApiServer:
|
||||
@@ -952,7 +1544,7 @@
|
||||
## Component scraping the kube controller manager
|
||||
##
|
||||
kubeControllerManager:
|
||||
|
@ -757,83 +730,7 @@
|
|||
|
||||
## If your kube controller manager is not deployed as a pod, specify IPs it can be found on
|
||||
##
|
||||
@@ -967,14 +1520,16 @@
|
||||
# Name of the server to use when validating TLS certificate
|
||||
serverName: null
|
||||
|
||||
- ## Metric relabel configs to apply to samples before ingestion.
|
||||
+ ## MetricRelabelConfigs to apply to samples after scraping, but before ingestion.
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
metricRelabelings: []
|
||||
# - action: keep
|
||||
# regex: 'kube_(daemonset|deployment|pod|namespace|node|statefulset).+'
|
||||
# sourceLabels: [__name__]
|
||||
|
||||
- # relabel configs to apply to samples before ingestion.
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
relabelings: []
|
||||
# - sourceLabels: [__meta_kubernetes_pod_node_name]
|
||||
@@ -1002,14 +1557,16 @@
|
||||
##
|
||||
proxyUrl: ""
|
||||
|
||||
- ## Metric relabel configs to apply to samples before ingestion.
|
||||
+ ## MetricRelabelConfigs to apply to samples after scraping, but before ingestion.
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
metricRelabelings: []
|
||||
# - action: keep
|
||||
# regex: 'kube_(daemonset|deployment|pod|namespace|node|statefulset).+'
|
||||
# sourceLabels: [__name__]
|
||||
|
||||
- # relabel configs to apply to samples before ingestion.
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
relabelings: []
|
||||
# - sourceLabels: [__meta_kubernetes_pod_node_name]
|
||||
@@ -1041,14 +1598,16 @@
|
||||
##
|
||||
proxyUrl: ""
|
||||
|
||||
- ## Metric relabel configs to apply to samples before ingestion.
|
||||
+ ## MetricRelabelConfigs to apply to samples after scraping, but before ingestion.
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
metricRelabelings: []
|
||||
# - action: keep
|
||||
# regex: 'kube_(daemonset|deployment|pod|namespace|node|statefulset).+'
|
||||
# sourceLabels: [__name__]
|
||||
|
||||
- # relabel configs to apply to samples before ingestion.
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
relabelings: []
|
||||
# - sourceLabels: [__meta_kubernetes_pod_node_name]
|
||||
@@ -1057,12 +1616,17 @@
|
||||
# targetLabel: nodename
|
||||
# replacement: $1
|
||||
# action: replace
|
||||
+
|
||||
+ ## MetricRelabelConfigs to apply to samples after scraping, but before ingestion.
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
+ ##
|
||||
dnsmasqMetricRelabelings: []
|
||||
# - action: keep
|
||||
# regex: 'kube_(daemonset|deployment|pod|namespace|node|statefulset).+'
|
||||
# sourceLabels: [__name__]
|
||||
|
||||
- # relabel configs to apply to samples before ingestion.
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
dnsmasqRelabelings: []
|
||||
# - sourceLabels: [__meta_kubernetes_pod_node_name]
|
||||
@@ -1075,7 +1639,7 @@
|
||||
@@ -1110,7 +1702,7 @@
|
||||
## Component scraping etcd
|
||||
##
|
||||
kubeEtcd:
|
||||
|
@ -842,26 +739,7 @@
|
|||
|
||||
## If your etcd is not deployed as a pod, specify IPs it can be found on
|
||||
##
|
||||
@@ -1119,14 +1683,16 @@
|
||||
certFile: ""
|
||||
keyFile: ""
|
||||
|
||||
- ## Metric relabel configs to apply to samples before ingestion.
|
||||
+ ## MetricRelabelConfigs to apply to samples after scraping, but before ingestion.
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
metricRelabelings: []
|
||||
# - action: keep
|
||||
# regex: 'kube_(daemonset|deployment|pod|namespace|node|statefulset).+'
|
||||
# sourceLabels: [__name__]
|
||||
|
||||
- # relabel configs to apply to samples before ingestion.
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
relabelings: []
|
||||
# - sourceLabels: [__meta_kubernetes_pod_node_name]
|
||||
@@ -1140,7 +1706,7 @@
|
||||
@@ -1177,7 +1769,7 @@
|
||||
## Component scraping kube scheduler
|
||||
##
|
||||
kubeScheduler:
|
||||
|
@ -870,26 +748,7 @@
|
|||
|
||||
## If your kube scheduler is not deployed as a pod, specify IPs it can be found on
|
||||
##
|
||||
@@ -1177,14 +1743,16 @@
|
||||
## Name of the server to use when validating TLS certificate
|
||||
serverName: null
|
||||
|
||||
- ## Metric relabel configs to apply to samples before ingestion.
|
||||
+ ## MetricRelabelConfigs to apply to samples after scraping, but before ingestion.
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
metricRelabelings: []
|
||||
# - action: keep
|
||||
# regex: 'kube_(daemonset|deployment|pod|namespace|node|statefulset).+'
|
||||
# sourceLabels: [__name__]
|
||||
|
||||
- # relabel configs to apply to samples before ingestion.
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
relabelings: []
|
||||
# - sourceLabels: [__meta_kubernetes_pod_node_name]
|
||||
@@ -1198,7 +1766,7 @@
|
||||
@@ -1237,7 +1829,7 @@
|
||||
## Component scraping kube proxy
|
||||
##
|
||||
kubeProxy:
|
||||
|
@ -898,45 +757,7 @@
|
|||
|
||||
## If your kube proxy is not deployed as a pod, specify IPs it can be found on
|
||||
##
|
||||
@@ -1229,14 +1797,16 @@
|
||||
##
|
||||
https: false
|
||||
|
||||
- ## Metric relabel configs to apply to samples before ingestion.
|
||||
+ ## MetricRelabelConfigs to apply to samples after scraping, but before ingestion.
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
metricRelabelings: []
|
||||
# - action: keep
|
||||
# regex: 'kube_(daemonset|deployment|pod|namespace|node|statefulset).+'
|
||||
# sourceLabels: [__name__]
|
||||
|
||||
- # relabel configs to apply to samples before ingestion.
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
relabelings: []
|
||||
# - action: keep
|
||||
@@ -1262,14 +1832,16 @@
|
||||
##
|
||||
selectorOverride: {}
|
||||
|
||||
- ## Metric relabel configs to apply to samples before ingestion.
|
||||
+ ## MetricRelabelConfigs to apply to samples after scraping, but before ingestion.
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
metricRelabelings: []
|
||||
# - action: keep
|
||||
# regex: 'kube_(daemonset|deployment|pod|namespace|node|statefulset).+'
|
||||
# sourceLabels: [__name__]
|
||||
|
||||
- # relabel configs to apply to samples before ingestion.
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
relabelings: []
|
||||
# - sourceLabels: [__meta_kubernetes_pod_node_name]
|
||||
@@ -1294,6 +1866,13 @@
|
||||
@@ -1337,6 +1929,13 @@
|
||||
create: true
|
||||
podSecurityPolicy:
|
||||
enabled: true
|
||||
|
@ -950,27 +771,7 @@
|
|||
|
||||
## Deploy node exporter as a daemonset to all nodes
|
||||
##
|
||||
@@ -1317,7 +1896,8 @@
|
||||
##
|
||||
scrapeTimeout: ""
|
||||
|
||||
- ## Metric relabel configs to apply to samples before ingestion.
|
||||
+ ## MetricRelabelConfigs to apply to samples after scraping, but before ingestion.
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
metricRelabelings: []
|
||||
# - sourceLabels: [__name__]
|
||||
@@ -1326,7 +1906,8 @@
|
||||
# replacement: $1
|
||||
# action: drop
|
||||
|
||||
- ## relabel configs to apply to samples before ingestion.
|
||||
+ ## RelabelConfigs to apply to samples before scraping
|
||||
+ ## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#relabelconfig
|
||||
##
|
||||
relabelings: []
|
||||
# - sourceLabels: [__meta_kubernetes_pod_node_name]
|
||||
@@ -1347,6 +1928,16 @@
|
||||
@@ -1392,6 +1991,16 @@
|
||||
extraArgs:
|
||||
- --collector.filesystem.ignored-mount-points=^/(dev|proc|sys|var/lib/docker/.+|var/lib/kubelet/.+)($|/)
|
||||
- --collector.filesystem.ignored-fs-types=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$
|
||||
|
@ -987,7 +788,7 @@
|
|||
|
||||
## Manages Prometheus and Alertmanager components
|
||||
##
|
||||
@@ -1359,8 +1950,8 @@
|
||||
@@ -1404,8 +2013,8 @@
|
||||
enabled: true
|
||||
# Value must match version names from https://golang.org/pkg/crypto/tls/#pkg-constants
|
||||
tlsMinVersion: VersionTLS13
|
||||
|
@ -998,16 +799,19 @@
|
|||
|
||||
## Admission webhook support for PrometheusRules resources added in Prometheus Operator 0.30 can be enabled to prevent incorrectly formatted
|
||||
## rules from making their way into prometheus and potentially preventing the container from starting
|
||||
@@ -1377,7 +1968,7 @@
|
||||
@@ -1422,9 +2031,9 @@
|
||||
patch:
|
||||
enabled: true
|
||||
image:
|
||||
- repository: k8s.gcr.io/ingress-nginx/kube-webhook-certgen
|
||||
+ repository: rancher-mirrored-k8s.gcr.io-ingress-nginx-kube-webhook-certgen
|
||||
+ repository: rancher/mirrored-ingress-nginx-kube-webhook-certgen
|
||||
tag: v1.0
|
||||
sha: "f3b6b39a6062328c095337b4cadcefd1612348fdd5190b1dcbcb9b9e90bd8068"
|
||||
- sha: "f3b6b39a6062328c095337b4cadcefd1612348fdd5190b1dcbcb9b9e90bd8068"
|
||||
+ sha: ""
|
||||
pullPolicy: IfNotPresent
|
||||
@@ -1526,13 +2117,13 @@
|
||||
resources: {}
|
||||
## Provide a priority class name to the webhook patching job
|
||||
@@ -1571,13 +2180,13 @@
|
||||
|
||||
## Resource limits & requests
|
||||
##
|
||||
|
@ -1028,7 +832,7 @@
|
|||
|
||||
# Required for use in managed kubernetes clusters (such as AWS EKS) with custom CNI (such as calico),
|
||||
# because control-plane managed by AWS cannot communicate with pods' IP CIDR and admission webhooks are not working
|
||||
@@ -1585,7 +2176,7 @@
|
||||
@@ -1630,7 +2239,7 @@
|
||||
## Prometheus-operator image
|
||||
##
|
||||
image:
|
||||
|
@ -1037,7 +841,7 @@
|
|||
tag: v0.50.0
|
||||
sha: ""
|
||||
pullPolicy: IfNotPresent
|
||||
@@ -1601,7 +2192,7 @@
|
||||
@@ -1646,7 +2255,7 @@
|
||||
## Prometheus-config-reloader image to use for config and rule reloading
|
||||
##
|
||||
prometheusConfigReloaderImage:
|
||||
|
@ -1046,7 +850,7 @@
|
|||
tag: v0.50.0
|
||||
sha: ""
|
||||
|
||||
@@ -1736,7 +2327,7 @@
|
||||
@@ -1781,7 +2390,7 @@
|
||||
port: 9090
|
||||
|
||||
## To be used with a proxy extraContainer port
|
||||
|
@ -1055,7 +859,7 @@
|
|||
|
||||
## List of IP addresses at which the Prometheus server service is available
|
||||
## Ref: https://kubernetes.io/docs/user-guide/services/#external-ips
|
||||
@@ -2009,7 +2600,7 @@
|
||||
@@ -2054,7 +2663,7 @@
|
||||
## Image of Prometheus.
|
||||
##
|
||||
image:
|
||||
|
@ -1064,7 +868,7 @@
|
|||
tag: v2.28.1
|
||||
sha: ""
|
||||
|
||||
@@ -2072,6 +2663,11 @@
|
||||
@@ -2117,6 +2726,11 @@
|
||||
##
|
||||
externalUrl: ""
|
||||
|
||||
|
@ -1076,7 +880,7 @@
|
|||
## Define which Nodes the Pods are scheduled on.
|
||||
## ref: https://kubernetes.io/docs/user-guide/node-selection/
|
||||
##
|
||||
@@ -2104,7 +2700,7 @@
|
||||
@@ -2149,7 +2763,7 @@
|
||||
## prometheus resource to be created with selectors based on values in the helm deployment,
|
||||
## which will also match the PrometheusRule resources created
|
||||
##
|
||||
|
@ -1085,7 +889,7 @@
|
|||
|
||||
## PrometheusRules to be selected for target discovery.
|
||||
## If {}, select all PrometheusRules
|
||||
@@ -2129,7 +2725,7 @@
|
||||
@@ -2174,7 +2788,7 @@
|
||||
## prometheus resource to be created with selectors based on values in the helm deployment,
|
||||
## which will also match the servicemonitors created
|
||||
##
|
||||
|
@ -1094,7 +898,7 @@
|
|||
|
||||
## ServiceMonitors to be selected for target discovery.
|
||||
## If {}, select all ServiceMonitors
|
||||
@@ -2152,7 +2748,7 @@
|
||||
@@ -2197,7 +2811,7 @@
|
||||
## prometheus resource to be created with selectors based on values in the helm deployment,
|
||||
## which will also match the podmonitors created
|
||||
##
|
||||
|
@ -1103,7 +907,7 @@
|
|||
|
||||
## PodMonitors to be selected for target discovery.
|
||||
## If {}, select all PodMonitors
|
||||
@@ -2283,9 +2879,13 @@
|
||||
@@ -2328,9 +2942,13 @@
|
||||
|
||||
## Resource limits & requests
|
||||
##
|
||||
|
@ -1120,7 +924,7 @@
|
|||
|
||||
## Prometheus StorageSpec for persistent data
|
||||
## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/storage.md
|
||||
@@ -2308,7 +2908,13 @@
|
||||
@@ -2353,7 +2971,13 @@
|
||||
# medium: Memory
|
||||
|
||||
# Additional volumes on the output StatefulSet definition.
|
||||
|
@ -1135,26 +939,43 @@
|
|||
|
||||
# Additional VolumeMounts on the output StatefulSet definition.
|
||||
volumeMounts: []
|
||||
@@ -2422,6 +3028,18 @@
|
||||
## ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/api.md#thanosspec
|
||||
##
|
||||
thanos: {}
|
||||
+ # secretProviderClass:
|
||||
+ # provider: gcp
|
||||
+ # parameters:
|
||||
+ # secrets: |
|
||||
+ # - resourceName: "projects/$PROJECT_ID/secrets/testsecret/versions/latest"
|
||||
+ # fileName: "objstore.yaml"
|
||||
+ # objectStorageConfigFile: /var/secrets/object-store.yaml
|
||||
+
|
||||
@@ -2475,9 +3099,34 @@
|
||||
# fileName: "objstore.yaml"
|
||||
# objectStorageConfigFile: /var/secrets/object-store.yaml
|
||||
|
||||
+ proxy:
|
||||
+ image:
|
||||
+ repository: rancher/mirrored-library-nginx
|
||||
+ tag: 1.21.1-alpine
|
||||
|
||||
+
|
||||
## Containers allows injecting additional containers. This is meant to allow adding an authentication proxy to a Prometheus pod.
|
||||
## if using proxy extraContainer update targetPort with proxy container port
|
||||
@@ -2433,7 +3051,7 @@
|
||||
- containers: []
|
||||
+ containers: |
|
||||
+ - name: prometheus-proxy
|
||||
+ args:
|
||||
+ - nginx
|
||||
+ - -g
|
||||
+ - daemon off;
|
||||
+ - -c
|
||||
+ - /nginx/nginx.conf
|
||||
+ image: "{{ template "system_default_registry" . }}{{ .Values.prometheus.prometheusSpec.proxy.image.repository }}:{{ .Values.prometheus.prometheusSpec.proxy.image.tag }}"
|
||||
+ ports:
|
||||
+ - containerPort: 8081
|
||||
+ name: nginx-http
|
||||
+ protocol: TCP
|
||||
+ volumeMounts:
|
||||
+ - mountPath: /nginx
|
||||
+ name: prometheus-nginx
|
||||
+ - mountPath: /var/cache/nginx
|
||||
+ name: nginx-home
|
||||
+ securityContext:
|
||||
+ runAsUser: 101
|
||||
+ runAsGroup: 101
|
||||
|
||||
## InitContainers allows injecting additional initContainers. This is meant to allow doing some changes
|
||||
## (permissions, dir tree) on mounted volumes before starting prometheus
|
||||
@@ -2485,7 +3134,7 @@
|
||||
|
||||
## PortName to use for Prometheus.
|
||||
##
|
||||
|
|
|
@ -1,6 +1,6 @@
|
|||
url: https://github.com/prometheus-community/helm-charts.git
|
||||
subdirectory: charts/kube-prometheus-stack
|
||||
commit: 12699d2f2d299902c38e80c911ab4b6901bdc686
|
||||
commit: c6208979d494156a3869d2e5faab669ce4301c68
|
||||
version: 100.0.0
|
||||
additionalCharts:
|
||||
- workingDir: charts-crd
|
||||
|
|
Loading…
Reference in New Issue