Skip to content
Kubernetes Monitoring resources for SREcon19 Americas
Branch: master
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
kube-state-metrics
prometheus
LICENSE
README.md

README.md

Try it: a minimal monitoring stack for Kubernetes

This repository contains configuration files to stand up a minimal monitoring stack on a Kubernetes cluster, as well as a number of sample queries you can test against your cluster. I tested these configurations and queries against the latest 1.12 stable release on GKE (1.12.5-gke.10 as of the time of writing).

This README assumes that you already have access to a Kubernetes cluster running the 1.12 release. You can spin up a new cluster pretty quickly (it took me about 10 minutes) using a free GKE trial. See the Google Cloud Platform console and developer docs for how to do that; there are also docs available for how to install the gcloud command line client and kubectl tool.

Set up kube-state-metrics (KSM)

kube-state-metrics is like a Prometheus adapter for your current cluster state. It provides many useful metrics that can help you understand your cluster's quality of service.

You can deploy the configurations provided by upstream directly from the master branch. I tested using the 1.5.0 release of KSM. A copy of the working 1.5.0 configurations is included in this repository for reference under the kube-state-metrics/ folder.

# Apply KSM configurations to the cluster
kubectl apply -f kube-state-metrics/
# clusterrolebinding.rbac.authorization.k8s.io/kube-state-metrics created
# clusterrole.rbac.authorization.k8s.io/kube-state-metrics created
# deployment.apps/kube-state-metrics created
# rolebinding.rbac.authorization.k8s.io/kube-state-metrics created
# role.rbac.authorization.k8s.io/kube-state-metrics-resizer created
# serviceaccount/kube-state-metrics created
# service/kube-state-metrics created

# Start a proxy to verify metrics
kubectl proxy &

# Curl the KSM metrics endpoint via the API server proxy
curl http://localhost:8001/api/v1/namespaces/kube-system/services/kube-state-metrics:8080/proxy/metrics
# Output:
# # HELP kube_configmap_info Information about configmap.
# # TYPE kube_configmap_info gauge
# kube_configmap_info{namespace="kube-system",configmap="ingress-uid"} 1
# kube_configmap_info{namespace="kube-system",configmap="extension-apiserver-authentication"} 1
# ...

Set up Prometheus

Prometheus is a time-series based monitoring system that scrapes metrics on an interval from instrumented jobs. Most Kubernetes components export metrics in Prometheus format by default.

While prometheus-operator is a convenient way to launch Prometheus in a Kubernetes cluster, it is quite complex. Here, I've included some simple, minimal configuration files to launch a Prometheus instance on our cluster. These are available in the prometheus/ folder in this repo.

# Apply Prometheus configurations to the cluster
kubectl apply -f prometheus/
# serviceaccount/prometheus created
# clusterrole.rbac.authorization.k8s.io/prometheus created
# clusterrolebinding.rbac.authorization.k8s.io/prometheus created
# configmap/prom-server-config created
# service/prom-ss created
# service/prometheus-server created
# statefulset.apps/prom-ss created

# Port-forward the Prometheus service for web access
kubectl port-forward -n kube-system svc/prometheus-server 8080:80 &
# Forwarding from 127.0.0.1:8080 -> 9090
# Forwarding from [::1]:8080 -> 9090

Now you can visit the Prometheus web UI in your browser at http://127.0.0.1:8080/graph and try out some queries. Perhaps you could check if your API server is up: http://127.0.0.1:8080/graph?g0.range_input=1h&g0.expr=up%7Bjob%3D%22kubernetes-apiservers%22%7D&g0.tab=0

Query Prometheus

Here are a number of PromQL queries you can try out against your Prometheus instance. You can browse all available timeseries in the Prometheus graph UI by clicking on the dropdown list.

The timeseries names tell us the source of the various metrics:

  • apiserver_* metrics are provided by the API servers
  • kube_* metrics are provided by kube-state-metrics
  • kubelet_* metrics are provided by the kubelets
  • container_* metrics are provided by cadvisor

There are also some built-in metrics that are generated at Prometheus scrape time, such as up (whether a scrape target is up or not).

General pod info

kube_pod_info                      # pod IP, node pod is scheduled on, controller type for pod
kube_pod_container_info            # container names in a pod, ID, images
kube_pod_container_restarts_total  # number of times a container has restarted
sum(kubelet_running_pod_count)     # all running pods

CPU usage (cpu cores)

sum(rate(container_cpu_usage_seconds_total{container_name!="POD"}[1m])) by (container_name,pod_name)  # actual CPU use per container
kube_pod_container_resource_requests_cpu_cores  # CPU requested by each container
kube_pod_container_resource_limits_cpu_cores    # CPU limits for each container

Memory usage (bytes)

sum(container_memory_working_set_bytes{container_name!="POD"}) by (container_name,pod_name)  # actual memory use per container
sum(container_memory_usage_bytes{container_name!="POD"}) by (container_name,pod_name)  # reserved memory in use per container
kube_pod_container_resource_requests_memory_bytes  # memory requested by each container
kube_pod_container_resource_limits_memory_bytes    # memory limits for each container

Network usage (b/s)

sum(rate(container_network_receive_bytes_total{container_name!="POD"}[1m])) by (container_name,pod_name)  # bytes received per second per container
-sum(rate(container_network_transmit_bytes_total{container_name!="POD"}[1m])) by (container_name)  # bytes transmitted per second per container

General cluster info

sum(up{job="kubernetes-nodes"})                           # number of online nodes
sum(kube_node_spec_unschedulable)                         # number of unavailable nodes
max(avg_over_time(up{job="kubernetes-apiservers"}[1d]))   # control plane uptime
count(kube_service_info)                                  # running services
sum(machine_cpu_cores)                                    # total cluster CPUs
sum(rate(container_cpu_usage_seconds_total{id="/"}[1m]))  # total used CPUs
sum(machine_memory_bytes)                                 # total cluster memory
sum(container_memory_working_set_bytes{id="/"})           # total used memory

Suppose some of our machines are unschedulable (i.e. the kubelet is posting an issue or someone ran kubectl cordon on the node). How might we calculate what capacity in our cluster is actually available for scheduling?

We could do some label joins with PromQL to figure this out:

sum(machine_cpu_cores) - sum(label_join(machine_cpu_cores, "node", "", "kubernetes_io_hostname") * ON(node) kube_node_spec_unschedulable)  # total available CPUs
sum(machine_memory_bytes) - sum(label_join(machine_memory_bytes, "node", "", "kubernetes_io_hostname") * ON(node) kube_node_spec_unschedulable)  # total available memory

Control plane info

sum(rate(apiserver_request_count[1m])) by (verb)  # control plane throughput by HTTP verb)
histogram_quantile(0.99, sum(rate(apiserver_request_latencies_bucket{verb!="WATCH",verb!="CONNECT"}[1m])) by (le, verb))  # p99 control plane latencies by verb

License

Copyright (c) 2019 Two Sigma Investments, LP.

Distributed under the Apache License 2.0. See the LICENSE file.

Contents of the kube-state-metrics/ folder are copyright (c) 2016-2019 The Linux Foundation via the Cloud Native Computing Foundation project and distributed under the Apache License 2.0.

These files have been copied, unmodified, from the kubernetes/kube-state-metrics repository. For full source history, you can view the upstream files at the corresponding VCS commit.

The file prometheus/server-config.yml is a derivative work of the example Kubernetes configurations provided by the Prometheus project. It copies and modifies the provided global and Kubernetes-specific configuration files.

You can’t perform that action at this time.