Skip to content
Switch branches/tags

Latest commit


Git stats


Failed to load latest commit information.
Latest commit message
Commit time

Try it: a minimal monitoring stack for Kubernetes

This repository contains configuration files to stand up a minimal monitoring stack on a Kubernetes cluster, as well as a number of sample queries you can test against your cluster. I tested these configurations and queries against the latest 1.12 stable release on GKE (1.12.5-gke.10 as of the time of writing).

This README assumes that you already have access to a Kubernetes cluster running the 1.12 release. You can spin up a new cluster pretty quickly (it took me about 10 minutes) using a free GKE trial. See the Google Cloud Platform console and developer docs for how to do that; there are also docs available for how to install the gcloud command line client and kubectl tool.

Set up kube-state-metrics (KSM)

kube-state-metrics is like a Prometheus adapter for your current cluster state. It provides many useful metrics that can help you understand your cluster's quality of service.

You can deploy the configurations provided by upstream directly from the master branch. I tested using the 1.5.0 release of KSM. A copy of the working 1.5.0 configurations is included in this repository for reference under the kube-state-metrics/ folder.

# Apply KSM configurations to the cluster
kubectl apply -f kube-state-metrics/
# created
# created
# deployment.apps/kube-state-metrics created
# created
# created
# serviceaccount/kube-state-metrics created
# service/kube-state-metrics created

# Start a proxy to verify metrics
kubectl proxy &

# Curl the KSM metrics endpoint via the API server proxy
curl http://localhost:8001/api/v1/namespaces/kube-system/services/kube-state-metrics:8080/proxy/metrics
# Output:
# # HELP kube_configmap_info Information about configmap.
# # TYPE kube_configmap_info gauge
# kube_configmap_info{namespace="kube-system",configmap="ingress-uid"} 1
# kube_configmap_info{namespace="kube-system",configmap="extension-apiserver-authentication"} 1
# ...

Set up Prometheus

Prometheus is a time-series based monitoring system that scrapes metrics on an interval from instrumented jobs. Most Kubernetes components export metrics in Prometheus format by default.

While prometheus-operator is a convenient way to launch Prometheus in a Kubernetes cluster, it is quite complex. Here, I've included some simple, minimal configuration files to launch a Prometheus instance on our cluster. These are available in the prometheus/ folder in this repo.

# Apply Prometheus configurations to the cluster
kubectl apply -f prometheus/
# serviceaccount/prometheus created
# created
# created
# configmap/prom-server-config created
# service/prom-ss created
# service/prometheus-server created
# statefulset.apps/prom-ss created

# Port-forward the Prometheus service for web access
kubectl port-forward -n kube-system svc/prometheus-server 8080:80 &
# Forwarding from -> 9090
# Forwarding from [::1]:8080 -> 9090

Now you can visit the Prometheus web UI in your browser at and try out some queries. Perhaps you could check if your API server is up:

Query Prometheus

Here are a number of PromQL queries you can try out against your Prometheus instance. You can browse all available timeseries in the Prometheus graph UI by clicking on the dropdown list.

The timeseries names tell us the source of the various metrics:

  • apiserver_* metrics are provided by the API servers
  • kube_* metrics are provided by kube-state-metrics
  • kubelet_* metrics are provided by the kubelets
  • container_* metrics are provided by cadvisor

There are also some built-in metrics that are generated at Prometheus scrape time, such as up (whether a scrape target is up or not).

General pod info

kube_pod_info                      # pod IP, node pod is scheduled on, controller type for pod
kube_pod_container_info            # container names in a pod, ID, images
kube_pod_container_restarts_total  # number of times a container has restarted
sum(kubelet_running_pod_count)     # all running pods

CPU usage (cpu cores)

sum(rate(container_cpu_usage_seconds_total{container_name!="POD"}[1m])) by (container_name,pod_name)  # actual CPU use per container
kube_pod_container_resource_requests_cpu_cores  # CPU requested by each container
kube_pod_container_resource_limits_cpu_cores    # CPU limits for each container

Memory usage (bytes)

sum(container_memory_working_set_bytes{container_name!="POD"}) by (container_name,pod_name)  # actual memory use per container
sum(container_memory_usage_bytes{container_name!="POD"}) by (container_name,pod_name)  # reserved memory in use per container
kube_pod_container_resource_requests_memory_bytes  # memory requested by each container
kube_pod_container_resource_limits_memory_bytes    # memory limits for each container

Network usage (b/s)

sum(rate(container_network_receive_bytes_total{container_name!="POD"}[1m])) by (container_name,pod_name)  # bytes received per second per container
-sum(rate(container_network_transmit_bytes_total{container_name!="POD"}[1m])) by (container_name)  # bytes transmitted per second per container

General cluster info

sum(up{job="kubernetes-nodes"})                           # number of online nodes
sum(kube_node_spec_unschedulable)                         # number of unavailable nodes
max(avg_over_time(up{job="kubernetes-apiservers"}[1d]))   # control plane uptime
count(kube_service_info)                                  # running services
sum(machine_cpu_cores)                                    # total cluster CPUs
sum(rate(container_cpu_usage_seconds_total{id="/"}[1m]))  # total used CPUs
sum(machine_memory_bytes)                                 # total cluster memory
sum(container_memory_working_set_bytes{id="/"})           # total used memory

Suppose some of our machines are unschedulable (i.e. the kubelet is posting an issue or someone ran kubectl cordon on the node). How might we calculate what capacity in our cluster is actually available for scheduling?

We could do some label joins with PromQL to figure this out:

sum(machine_cpu_cores) - sum(label_join(machine_cpu_cores, "node", "", "kubernetes_io_hostname") * ON(node) kube_node_spec_unschedulable)  # total available CPUs
sum(machine_memory_bytes) - sum(label_join(machine_memory_bytes, "node", "", "kubernetes_io_hostname") * ON(node) kube_node_spec_unschedulable)  # total available memory

Control plane info

sum(rate(apiserver_request_count[1m])) by (verb)  # control plane throughput by HTTP verb)
histogram_quantile(0.99, sum(rate(apiserver_request_latencies_bucket{verb!="WATCH",verb!="CONNECT"}[1m])) by (le, verb))  # p99 control plane latencies by verb


Copyright (c) 2019 Two Sigma Investments, LP.

Distributed under the Apache License 2.0. See the LICENSE file.

Contents of the kube-state-metrics/ folder are copyright (c) 2016-2019 The Linux Foundation via the Cloud Native Computing Foundation project and distributed under the Apache License 2.0.

These files have been copied, unmodified, from the kubernetes/kube-state-metrics repository. For full source history, you can view the upstream files at the corresponding VCS commit.

The file prometheus/server-config.yml is a derivative work of the example Kubernetes configurations provided by the Prometheus project. It copies and modifies the provided global and Kubernetes-specific configuration files.


Kubernetes Monitoring Prometheus PoC and sample configs for SREcon19 Americas





No releases published


No packages published