Skip to content

Latest commit

 

History

History
178 lines (129 loc) · 6.19 KB

scheduler.md

File metadata and controls

178 lines (129 loc) · 6.19 KB

Kubernetes-Mesos Scheduler

Kubernetes on Mesos does not use the upstream scheduler binary, but replaces it with its own Mesos framework scheduler. The following gives an overview of the differences.

Labels and Mesos Agent Attributes

The scheduler of Kubernetes-Mesos takes labels into account: it matches specified labels in pod specs with defined labels of nodes.

In addition to user defined labels, attributes of Mesos agents are converted into node labels by the scheduler, following the pattern

k8s.mesosphere.io/attribute-<name>: value

As an example, a Mesos agent attribute of generation:2015 will result in the node label

k8s.mesosphere.io/attribute-generation: 2015

and can be used to schedule pods onto nodes which are of generation 2015.

Note: Node labels prefixed by k8s.mesosphere.io are managed by Kubernetes-Mesos and should not be modified manually by the user or admin. For example, the Kubernetes-Mesos executor manages k8s.mesosphere.io/attribute labels and will auto-detect and update modified attributes when the mesos-slave is restarted.

Resource Roles

A Mesos cluster can be statically partitioned using resources roles. Each resource is assigned such a role (* is the default role, if none is explicitly assigned in the mesos-slave command line). The Mesos master will send offers to frameworks for * resources and – optionally – one additional role that a framework is assigned to. Right now only one such additional role for a framework is supported.

Configuring Roles for the Scheduler

Every Mesos framework scheduler can choose among offered * resources and optionally one additional role. The Kubernetes-Mesos scheduler supports this by setting the framework roles in the scheduler command line, e.g.

$ km scheduler ... --mesos-framework-roles="*,role1" ...

This permits the Kubernetes-Mesos scheduler to accept offered resources for the * and role1 roles. By default pods may be assigned any combination of resources for the roles accepted by the scheduler. This default role assignment behavior may be overridden using the --mesos-default-pod-roles flag or else by annotating the pod (as described later).

One can configure default pod roles, e.g.

$ km scheduler ... --mesos-default-pod-roles="role1" ...

This will tell the Kubernetes-Mesos scheduler to default to role1 resource offers. The configured default pod roles must be a subset of the configured framework roles.

The order of configured default pod roles is relevant, --mesos-default-pod-roles=role1,* will first try to consume role1 resources from an offer and, once depleted, fall back to * resources.

The configuration --mesos-default-pod-roles=*,role1 has the reverse behavior. It first tries to consume * resources from an offer and, once depleted, falls back to role1 resources.

Due to restrictions of Mesos, currently only one additional role next to * can be configured for both framework and default pod roles.

Specifying Roles for Pods

By default a pod is scheduled using resources as specified using the --mesos-default-pod-roles configuration.

A pod can override of this default behaviour using a k8s.mesosphere.io/roles annotation:

k8s.mesosphere.io/roles: "*,role1"

The format is a comma separated list of allowed resource roles. The scheduler will try to schedule the pod with * resources first, using role1 resources if the former are not available or are depleted.

Note: An empty list will mean that no resource roles are allowed which is equivalent to a pod which is unschedulable.

For example:

apiVersion: v1
kind: Pod
metadata:
  name: backend
  annotations:
    k8s.mesosphere.io/roles: "*,public"
  namespace: prod
spec:
  ...

This */public pod will be scheduled using resources from both roles, preferably using * resources, followed by public. If none of those roles provides enough resources, the scheduling fails.

Note: The scheduler will also allow to mix different roles in the following sense: if a node provides cpu resources for the * role, but mem resources only for the public role, the above pod will be scheduled using cpu(*) and mem(public) resources.

Note: The scheduler might also mix within one resource type, i.e. it will use as many cpus of the * role as possible. If a pod requires even more cpu resources (defined using the pod.spec.resources.limits property) for successful scheduling, the scheduler will add resources from the public role until the pod resource requirements are satisfied. E.g. a pod might be scheduled with 0.5 cpu(*), 1.5 cpu(public) resources plus e.g. 2 GB mem(public) resources.

Tuning

The scheduler configuration can be fine-tuned using an ini-style configuration file. The filename is passed via --scheduler-config to the km scheduler command.

Be warned though that some them are pretty low-level and one has to know the inner workings of k8sm to find sensible values. Moreover, these settings may change or even disappear from version to version without further notice.

The following settings are the default:

[scheduler]
; duration an offer is viable, prior to being expired
offer-ttl = 5s

; duration an expired offer lingers in history
offer-linger-ttl = 2m

; duration between offer listener notifications
listener-delay = 1s

; size of the pod updates channel
updates-backlog = 2048

; interval we update the frameworkId stored in etcd
framework-id-refresh-interval = 30s

; wait this amount of time after initial registration before attempting
; implicit reconciliation
initial-implicit-reconciliation-delay = 15s

; interval in between internal task status checks/updates
explicit-reconciliation-max-backoff = 2m

; waiting period after attempting to cancel an ongoing reconciliation
explicit-reconciliation-abort-timeout = 30s

initial-pod-backoff = 1s
max-pod-backoff = 60s
http-handler-timeout = 10s
http-bind-interval = 5s

Low-Level Scheduler Architecture

Scheduler Structure

Analytics