summaryrefslogtreecommitdiff
path: root/config/prometheus/common_metrics.yml
diff options
context:
space:
mode:
Diffstat (limited to 'config/prometheus/common_metrics.yml')
-rw-r--r--config/prometheus/common_metrics.yml28
1 files changed, 0 insertions, 28 deletions
diff --git a/config/prometheus/common_metrics.yml b/config/prometheus/common_metrics.yml
index d9aaff12a4d..67994740ff1 100644
--- a/config/prometheus/common_metrics.yml
+++ b/config/prometheus/common_metrics.yml
@@ -2,12 +2,10 @@ dashboard: 'Environment metrics'
priority: 1
panel_groups:
- group: System metrics (Kubernetes)
- priority: 15
panels:
- title: "Memory Usage (Total)"
type: "area-chart"
y_label: "Total Memory Used (GB)"
- weight: 4
metrics:
- id: system_metrics_kubernetes_container_memory_total
# Remove the second metric (after OR) when we drop support for K8s 1.13
@@ -18,7 +16,6 @@ panel_groups:
- title: "Core Usage (Total)"
type: "area-chart"
y_label: "Total Cores"
- weight: 3
metrics:
- id: system_metrics_kubernetes_container_cores_total
# Remove the second metric (after OR) when we drop support for K8s 1.13
@@ -29,7 +26,6 @@ panel_groups:
- title: "Memory Usage (Pod average)"
type: "line-chart"
y_label: "Memory Used per Pod (MB)"
- weight: 2
metrics:
- id: system_metrics_kubernetes_container_memory_average
# Remove the second metric (after OR) when we drop support for K8s 1.13
@@ -40,7 +36,6 @@ panel_groups:
- title: "Canary: Memory Usage (Pod Average)"
type: "line-chart"
y_label: "Memory Used per Pod (MB)"
- weight: 2
metrics:
- id: system_metrics_kubernetes_container_memory_average_canary
# Remove the second metric (after OR) when we drop support for K8s 1.13
@@ -52,7 +47,6 @@ panel_groups:
- title: "Core Usage (Pod Average)"
type: "line-chart"
y_label: "Cores per Pod"
- weight: 1
metrics:
- id: system_metrics_kubernetes_container_core_usage
# Remove the second metric (after OR) when we drop support for K8s 1.13
@@ -63,7 +57,6 @@ panel_groups:
- title: "Canary: Core Usage (Pod Average)"
type: "line-chart"
y_label: "Cores per Pod"
- weight: 1
metrics:
- id: system_metrics_kubernetes_container_core_usage_canary
# Remove the second metric (after OR) when we drop support for K8s 1.13
@@ -75,7 +68,6 @@ panel_groups:
- title: "Knative function invocations"
type: "area-chart"
y_label: "Invocations"
- weight: 1
metrics:
- id: system_metrics_knative_function_invocation_count
query_range: 'sum(ceil(rate(istio_requests_total{destination_service_namespace="{{kube_namespace}}", destination_service=~"{{function_name}}.*"}[1m])*60))'
@@ -83,12 +75,10 @@ panel_groups:
unit: requests
# NGINX Ingress metrics for pre-0.16.0 versions
- group: Response metrics (NGINX Ingress VTS)
- priority: 10
panels:
- title: "Throughput"
type: "area-chart"
y_label: "Requests / Sec"
- weight: 1
metrics:
- id: response_metrics_nginx_ingress_throughput_status_code
query_range: 'sum(rate(nginx_upstream_responses_total{upstream=~"{{kube_namespace}}-{{ci_environment_slug}}-.*"}[2m])) by (status_code)'
@@ -99,7 +89,6 @@ panel_groups:
y_label: "Latency (ms)"
y_axis:
format: milliseconds
- weight: 1
metrics:
- id: response_metrics_nginx_ingress_latency_pod_average
query_range: 'avg(nginx_upstream_response_msecs_avg{upstream=~"{{kube_namespace}}-{{ci_environment_slug}}-.*"})'
@@ -110,7 +99,6 @@ panel_groups:
y_label: "HTTP Errors (%)"
y_axis:
format: percentHundred
- weight: 1
metrics:
- id: response_metrics_nginx_ingress_http_error_rate
query_range: 'sum(rate(nginx_upstream_responses_total{status_code="5xx", upstream=~"{{kube_namespace}}-{{ci_environment_slug}}-.*"}[2m])) / sum(rate(nginx_upstream_responses_total{upstream=~"{{kube_namespace}}-{{ci_environment_slug}}-.*"}[2m])) * 100'
@@ -118,12 +106,10 @@ panel_groups:
unit: "%"
# NGINX Ingress metrics for post-0.16.0 versions
- group: Response metrics (NGINX Ingress)
- priority: 10
panels:
- title: "Throughput"
type: "area-chart"
y_label: "Requests / Sec"
- weight: 1
metrics:
- id: response_metrics_nginx_ingress_16_throughput_status_code
query_range: 'sum(label_replace(rate(nginx_ingress_controller_requests{namespace="{{kube_namespace}}",ingress=~".*{{ci_environment_slug}}.*"}[2m]), "status_code", "${1}xx", "status", "(.)..")) by (status_code)'
@@ -132,7 +118,6 @@ panel_groups:
- title: "Latency"
type: "area-chart"
y_label: "Latency (ms)"
- weight: 1
metrics:
- id: response_metrics_nginx_ingress_16_latency_pod_average
query_range: 'sum(rate(nginx_ingress_controller_ingress_upstream_latency_seconds_sum{namespace="{{kube_namespace}}",ingress=~".*{{ci_environment_slug}}.*"}[2m])) / sum(rate(nginx_ingress_controller_ingress_upstream_latency_seconds_count{namespace="{{kube_namespace}}",ingress=~".*{{ci_environment_slug}}.*"}[2m])) * 1000'
@@ -141,19 +126,16 @@ panel_groups:
- title: "HTTP Error Rate"
type: "area-chart"
y_label: "HTTP Errors (%)"
- weight: 1
metrics:
- id: response_metrics_nginx_ingress_16_http_error_rate
query_range: 'sum(rate(nginx_ingress_controller_requests{status=~"5.*",namespace="{{kube_namespace}}",ingress=~".*{{ci_environment_slug}}.*"}[2m])) / sum(rate(nginx_ingress_controller_requests{namespace="{{kube_namespace}}",ingress=~".*{{ci_environment_slug}}.*"}[2m])) * 100'
label: 5xx Errors (%)
unit: "%"
- group: Response metrics (HA Proxy)
- priority: 10
panels:
- title: "Throughput"
type: "area-chart"
y_label: "Requests / Sec"
- weight: 1
metrics:
- id: response_metrics_ha_proxy_throughput_status_code
query_range: 'sum(rate(haproxy_frontend_http_requests_total{ {{environment_filter}} }[2m])) by (code)'
@@ -162,19 +144,16 @@ panel_groups:
- title: "HTTP Error Rate"
type: "area-chart"
y_label: "Error Rate (%)"
- weight: 1
metrics:
- id: response_metrics_ha_proxy_http_error_rate
query_range: 'sum(rate(haproxy_frontend_http_responses_total{code="5xx",{{environment_filter}} }[2m])) / sum(rate(haproxy_frontend_http_responses_total{ {{environment_filter}} }[2m]))'
label: HTTP Errors (%)
unit: "%"
- group: Response metrics (AWS ELB)
- priority: 10
panels:
- title: "Throughput"
type: "area-chart"
y_label: "Requests / Sec"
- weight: 1
metrics:
- id: response_metrics_aws_elb_throughput_requests
query_range: 'sum(aws_elb_request_count_sum{ {{environment_filter}} }) / 60'
@@ -183,7 +162,6 @@ panel_groups:
- title: "Latency"
type: "area-chart"
y_label: "Latency (ms)"
- weight: 1
metrics:
- id: response_metrics_aws_elb_latency_average
query_range: 'avg(aws_elb_latency_average{ {{environment_filter}} }) * 1000'
@@ -192,19 +170,16 @@ panel_groups:
- title: "HTTP Error Rate"
type: "area-chart"
y_label: "Error Rate (%)"
- weight: 1
metrics:
- id: response_metrics_aws_elb_http_error_rate
query_range: 'sum(aws_elb_httpcode_backend_5_xx_sum{ {{environment_filter}} }) / sum(aws_elb_request_count_sum{ {{environment_filter}} })'
label: HTTP Errors (%)
unit: "%"
- group: Response metrics (NGINX)
- priority: 10
panels:
- title: "Throughput"
type: "area-chart"
y_label: "Requests / Sec"
- weight: 1
metrics:
- id: response_metrics_nginx_throughput_status_code
query_range: 'sum(rate(nginx_server_requests{server_zone!="*", server_zone!="_", {{environment_filter}} }[2m])) by (code)'
@@ -213,7 +188,6 @@ panel_groups:
- title: "Latency"
type: "area-chart"
y_label: "Latency (ms)"
- weight: 1
metrics:
- id: response_metrics_nginx_latency
query_range: 'avg(nginx_server_requestMsec{ {{environment_filter}} })'
@@ -224,7 +198,6 @@ panel_groups:
y_label: "HTTP 500 Errors / Sec"
y_axis:
precision: 0
- weight: 1
metrics:
- id: response_metrics_nginx_http_error_rate
query_range: 'sum(rate(nginx_server_requests{code="5xx", {{environment_filter}} }[2m]))'
@@ -233,7 +206,6 @@ panel_groups:
- title: "HTTP Error Rate"
type: "area-chart"
y_label: "HTTP Errors (%)"
- weight: 1
metrics:
- id: response_metrics_nginx_http_error_percentage
query_range: 'sum(rate(nginx_server_requests{code=~"5.*", host="*", {{environment_filter}} }[2m])) / sum(rate(nginx_server_requests{code="total", host="*", {{environment_filter}} }[2m])) * 100'