Application Layer Policy for Project Calico enforces network and application layer authorization policies using Istio.
Istio mints and distributes cryptographic identities and uses them to establish mutually authenticated TLS connections between pods. Calico enforces authorization policy on this communication integrating cryptographic identities and network layer attributes.
The envoy.ext_authz
filter inserted into the proxy, which calls out to Dikastes when service requests are
processed. We compute policy based on a global store which is distributed to Dikastes by its local Felix.
This guide explains how to install Calico & Istio into your cluster, and use it to enforce authorization policies in a simple demo application.
This demo will run on a Calico-enabled Kubernetes cluster. You will need a Kubernetes cluster running v1.8 or later with RBAC and Initializers enabled.
If you have Calico or Istio installed, remove them from the cluster. This preview relies on the latest Calico build and several custom-built Istio components which will be installed in the demo.
If you do not have a test cluster running Kubernetes 1.8 or later with RBAC and Initializers, this section will walk you through creating one on your local machine using Vagrant.
If you already have a test cluster, you can skip to installing calicoctl
.
Install Vagrant and VirtualBox, then from the root directory of this repo:
cd config/cluster
vagrant up
This will create a 3-node Kubernetes cluster in 3 VirtualBox VMs.
DO NOT USE THIS IN PRODUCTION. The API server is loaded with a certificate and keypair checked into this repository. If you put this in production anyone will be able to impersonate your API server.
Open VirtualBox and click on one of the created VMs, then click Network and go to the tab for Adapter 2. You should see it "Attached to: Host-only Adapter". Make a note of the entry in the "Name:" box. This is the name of the host-only network adapter you will use to communicate with your cluster.
Add an IP address to this network adapter and bring it up.
On Linux:
sudo ip addr add 172.18.18.1/24 dev <adaptername>
sudo ip link set <adaptername> up
Verify you can ping the master
ping 172.18.18.101
Finally, add the cluster to your kubeconfig and activate the context
kubectl config set-cluster vagrant-cluster --server=https://172.18.18.101:6443 --certificate-authority=$(pwd)/apiserver.crt
kubectl config set-credentials vagrant-admin --username=admin --password=admin
kubectl config set-context vagrant-admin --cluster=vagrant-cluster --user=vagrant-admin
kubectl config use-context vagrant-admin
Verify your kubeconfig is working, for example:
kubectl get pods
Since we are using the Kubernetes API server as the Calico datastore in this demo cluster (KDD mode), we need to configure calicoctl to use that datastore as well. This can be done by setting the following environment variables
export CALICO_DATASTORE_TYPE=kubernetes CALICO_KUBECONFIG=<your kube config file>
You will need an updated version of calicoctl
.
wget https://www.projectcalico.org/builds/calicoctl
chmod +x calicoctl
Configure calicoctl to connect to your Calico datastore by following the instructions appropriate for your cluster. If you followed the directions for installing a Vagrant cluster above, you have already completed this configuration.
From the main project directory:
kubectl apply -f config/install/05-calico.yaml
Install the Istio roles, bindings, and components.
kubectl apply -f config/install/10-istio.yaml
When all components have started, you should see the pods in the istio-system
namespace similar to the following.
kubectl get pods --namespace=istio-system
NAME READY STATUS RESTARTS AGE
istio-citadel-55bbf4ddff-hdbqb 1/1 Running 0 48m
istio-cleanup-old-ca-c97g4 0/1 Completed 0 48m
istio-cleanup-secrets-2jvjx 0/1 Completed 0 48m
istio-egressgateway-6864b4f8cf-hcrtr 1/1 Running 0 48m
istio-galley-68559fd97f-sglj2 1/1 Running 0 48m
istio-ingress-598f66ccbf-scgj2 1/1 Running 0 48m
istio-ingressgateway-59f87688f-6txhr 1/1 Running 0 48m
istio-mixer-post-install-v6fn4 0/1 Completed 0 48m
istio-pilot-796444c567-sf89w 2/2 Running 0 48m
istio-policy-77b65686bb-rzxp4 2/2 Running 0 48m
istio-security-post-install-hn5g2 0/1 Completed 0 48m
istio-sidecar-injector-896658cbd-h6xk5 1/1 Running 0 48m
istio-statsd-prom-bridge-6dbb7dcc7f-75q7g 1/1 Running 0 48m
istio-telemetry-65968c56f8-qg672 2/2 Running 0 48m
prometheus-586d95b8d9-ww8jq 1/1 Running 0 48m
Dikastes is a Calico component that computes authorization policy for the Istio proxies on each host. Calico utilizes Istio's automatic sidecar injection to inject Dikastes container into a pod at pod creation time.
Enable sidecar injection for the default
namespace where the rest of this demo is run.
kubectl label namespace default istio-injection=enabled
If you want to experiment with applications in other namespaces, label them using the same command.
Apply the manifest to configure Istio to use Calico Application Layer Policy for authorization.
kubectl apply -f 20-app-policy.yaml
We will use a simple microservice application to demonstrate Calico application layer policy. The YAO Bank application creates a customer-facing web application, a microservice that serves up account summaries, and an etcd database.
kubectl apply -f config/demo/10-yaobank.yaml
When the demo application has come up, you will see 3 pods.
kubectl get pods
NAME READY STATUS RESTARTS AGE
customer-2809159614-qqfnx 3/3 Running 0 21h
database-1601951801-m4w70 3/3 Running 0 21h
summary-2817688950-g1b3n 3/3 Running 0 21h
There is a Kubernetes ServiceAccount for each microservice in the application (in addition to the default
account).
kubectl get serviceaccount
NAME SECRETS AGE
customer 1 21h
database 1 21h
default 1 21h
summary 1 21h
Notice also that Istio CA will have created a secret of type istio.io/key-and-cert
for each service account. These
keys and X.509 certificates are used to cryptographically authenticate traffic in the Istio service mesh, and the
corresponding service account identities are used by Dikastes in authorization policy.
kubectl get secret
NAME TYPE DATA AGE
customer-token-mgb8w kubernetes.io/service-account-token 3 21h
database-token-nb5xp kubernetes.io/service-account-token 3 21h
default-token-wwml6 kubernetes.io/service-account-token 3 21h
istio.customer istio.io/key-and-cert 3 21h
istio.database istio.io/key-and-cert 3 21h
istio.default istio.io/key-and-cert 3 21h
istio.summary istio.io/key-and-cert 3 21h
summary-token-8kpt1 kubernetes.io/service-account-token 3 21h
You will use the istio-ingressgateway
service to access the YAO Bank application.
External load balancers are not supported in the Vagrant test cluster. You can use the host IP of the ingress service, along with the NodePort, to access the ingress.
export GATEWAY_URL=$(kubectl get po -n istio-system -l istio=ingressgateway -o 'jsonpath={.items[0].status.hostIP}'):$(kubectl get svc istio-ingress -n istio-system -o 'jsonpath={.spec.ports[0].nodePort}')
Point your browser to http://$GATEWAY_URL/
to confirm the YAO Bank application is functioning correctly. It may take
several minutes for all the services to come up and respond, during which time you may see 404 or 500 errors.
Although Calico & Istio are running in the cluster, we have not defined any authorization policy. Istio was configured to mutually authenticate traffic between the pods in your application, so only connections with Istio-issued certificates are allowed, and all inter-pod traffic is encrypted with TLS. That's already a big step in the right direction.
But, let's consider some deficiencies in this security architecture:
- All incoming connections from workloads in the Istio mesh are equally trusted
- Possession of a key & certificate pair is the only access credential considered.
To understand why these might be a problem, let's take them one at a time.
Trusting connections from any workload in the Istio mesh is a poor security architecture because, like Kubernetes, Istio is designed to host multiple applications. Some of those applications may not be as trusted as others. They may be operated by different users or teams with wildly different security requirements. We don't want our secure financial application microservices accessible from some hacky prototype another developer is cooking up.
Even within our own application, the best practice is to limit access as much as possible. Only pods that need access to a service should get it. Consider the YAO Bank application. The customer web service does not need, and should not have direct access to the backend database. The customer web service needs to directly interact with clients outside the cluster, some of whom may be malicious. Unfortunately, vulnerabilities in web applications are all too common. For example, an unpatched vulnerabiltiy in Apache Struts is what allowed attackers their initial access into the Equifax network where they then launched a devastating attack to steal millions of people's financial information.
Imagine what would happen if an attacker were to gain control of the customer web pod in our application.
Let's simulate this by exec
'ing into that pod.
kubectl exec -ti customer-<fill in pod ID> -c customer bash
You should get a bash shell inside the customer pod. Notice that from here, we get direct access to the backend database. For example, we can list all the entries in the database like this:
curl http://database:2379/v2/keys?recursive=true | python -m json.tool
(Piping to python -m json.tool
nicely formats the output.)
The possession of a key and certificate pair is a very strong assertion that a connection is authentic because it is based on cryptographic proofs that are believed to be nearly impossible to forge. When we authenticate connections this way we can say with extremely high confidence that the party on the other end is in possession of the corresponding key. However, this is only a proxy for what we actually want to be confident of: that the party on the other end really is the authorized workload we want to communicate with. Keeping the private key a secret is vital to this confidence, and occasionally attackers can find ways to trick applications into giving up secrets they should not. For example, the Heartbleed vulnerability in OpenSSL allowed attackers to trick an affected application into reading out portions of its memory, compromising private keys (among other confidential information).
Let's simulate an attacker who has stolen the private keys of another pod. Since the keys are stored as Kubernetes secrets, we won't exploit a vulnerability in a service, but instead just mount the secret in a pod that will simulate an attacker.
If you are still exec
'd into the customer pod, exit out or open a new terminal tab (we will return the to the
customer pod later).
kubectl apply -f config/demo/20-attack-pod.yaml
Take a look at the 20-attack-pod.yaml
file in an editor. It creates an ubuntu
pod and mounts istio.summary
secret. This will allow us to masquerade as if we were the summary
service, even though this pod is not run as that
service account. Let's try this out. First, exec
into the pod.
kubectl exec -ti attack-<fill in pod ID> bash
Next, install the curl
utility to initiate HTTP connections from the command line.
apt update && apt install -y curl
Now, we will attack the database. Instead of listing the contents like we did before, let's try something more
malicious, like changing the account balance with a PUT
command.
curl -k https://database:2379/v2/keys/accounts/519940/balance -d value="10000.00" -XPUT --key /etc/certs/key.pem --cert /etc/certs/cert-chain.pem
Unlike when we did this with the customer web pod, we do not have the Istio Proxy to handle encryption, so we have to
pass an https
URL, the --key
and --cert
parameters to curl
to do the cryptography.
Return to your web browser and refresh to confirm the new balance.
We can mitigate both of the above deficiencies with a Calico policy.
Apply the sample policy.
./calicoctl create -f config/demo/30-policy.yaml
Let's examine this policy piece by piece. It consists of 3 policy objects, one for each microservice.
apiVersion: projectcalico.org/v3
kind: GlobalNetworkPolicy
metadata:
name: customer
spec:
selector: app == 'customer'
ingress:
- action: Allow
http:
methods: ["GET"]
egress:
- action: Allow
This policy protects the customer web app. Since this application is customer facing, we do not restrict what can
communicate with it. We do, however, restrict that only HTTP GET
requests are allowed.
apiVersion: projectcalico.org/v3
kind: GlobalNetworkPolicy
metadata:
name: summary
spec:
selector: app == 'summary'
ingress:
- action: Allow
source:
serviceAccounts:
names: ["customer"]
egress:
- action: Allow
The second policy protects the account summary microservice. We know the only consumer of this service is the customer web app, so we restrict the source of incoming connections to the service account for the customer web app.
apiVersion: projectcalico.org/v3
kind: GlobalNetworkPolicy
metadata:
name: database
spec:
selector: app == 'database'
ingress:
- action: Allow
source:
serviceAccounts:
names: ["summary"]
egress:
- action: Allow
The third policy protects the database. Only the summary microservice should have direct access to the database.
Let's verify our policy is working as intended. First, return to your browser and refresh, to ensure policy enforcement has not broken the application.
Next, return to the customer web app. Recall that we simulated an attacker gaining control of that pod by exec
ing
into it.
kubectl exec -ti customer-<fill in pod ID> -c customer bash
Repeat our attempt to access the database.
curl http://database:2379/v2/keys?recursive=true
This time we should get a 403 Forbidden response (we have left out the JSON formatting because we do not expect to get a valid JSON response). Only the account summary microservice has database access according to our policy.
Finally, let's return to the attack pod that simulated stealing secret keys.
kubectl exec -ti attack-<fill in pod ID> bash
Let's repeat our attack with stolen keys (we'll further increase the account balance to highlight whether it succeeds).
curl -k https://database:2379/v2/keys/account/519940/balance -d value="99999.99" -XPUT --key /etc/certs/key.pem --cert /etc/certs/cert-chain.pem
If things are working correctly, you should get no response, and refreshing your browser should not show an increased balance.
You might wonder how Calico was able to detect and prevent this attack---the attacker was able to steal the keys which prove identity in our system. This highlights the value of multi-layer authorization checks. Although our attack pod had the keys to fool the X.509 certificate check, Calico also monitors the Kubernetes API Server for which IP addresses are associated with which service accounts. Since our attack pod has an IP not associated with the account summary service account we disallow the connection.
- Only
Allow
rules can be used with Application Layer Policy selectors. If you useDeny
orPass
rules they must be restricted to network layer selectors. - Application Layer Policy selectors are only supported on ingress policy. If you use egress policy, you must restrict to network layer selectors.
Yes, you can use Dikastes with Mixer. Since Dikastes handles authorization checks, we expect most people will want to
use Mixer primarily for reporting telemetry. Simply disable the Mixer checks (disablePolicyChecks: true
), but keep
report functionality on. If you decide to use both Dikastes and Mixer for authorization checks, keep in mind that
requests must pass both checks in order to be allowed.