Skip to content
This repository has been archived by the owner on Jan 11, 2023. It is now read-only.

Latest commit

 

History

History
102 lines (76 loc) · 6.19 KB

walkthrough.md

File metadata and controls

102 lines (76 loc) · 6.19 KB

Walkthrough

Once your Kubernetes cluster has been created you will have a resource group containing:

  1. 1 master accessible by SSH on port 22 or kubectl on port 443

  2. a set of nodes in an availability set. The nodes can be accessed through a master. See agent forwarding for an example of how to do this.

The following image shows the architecture of a container service cluster with 1 master, and 2 agents:

Image of Kubernetes cluster on azure

In the image above, you can see the following parts:

  1. Master Components - The master runs the Kubernetes scheduler, api server, and controller manager. Port 443 is exposed for remote management with the kubectl cli.
  2. Nodes - the Kubernetes nodes run in an availability set. Azure load balancers are dynamically added to the cluster depending on exposed services.
  3. Common Components - All VMs run a kubelet, Docker, and a Proxy.
  4. Networking - All VMs are assigned an ip address in the 10.240.0.0/16 network. Each VM is assigned a /24 subnet for their pod CIDR enabling IP per pod. The proxy running on each VM implements the service network 10.0.0.0/16.

All VMs are in the same private VNET and are fully accessible to each other.

Create your First Kubernetes Service

After completing this walkthrough you will know how to:

  • access Kubernetes cluster via SSH,
  • deploy a simple Docker application and expose to the world,
  • the location of the Kube config file and how to access the Kubernetes cluster remotely,
  • use kubectl exec to run commands in a container,
  • and finally access the Kubernetes dashboard.
  1. After successfully deploying the template write down the master FQDNs (Fully Qualified Domain Name).

    1. If using Powershell or CLI, the output parameter is in the OutputsString section named 'masterFQDN'
    2. If using Portal, to get the output you need to:
      1. navigate to "resource group"
      2. click on the resource group you just created
      3. then click on "Succeeded" under last deployment
      4. then click on the "Microsoft.Template"
      5. now you can copy the output FQDNs and sample SSH commands Image of docker scaling
  2. SSH to the master FQDN obtained in step 1.

  3. Explore your nodes and running pods:

  4. to see a list of your nodes type kubectl get nodes. If you want full detail of the nodes, add -o yaml to become kubectl get nodes -o yaml.

  5. to see a list of running pods type kubectl get pods --all-namespaces.

  6. Start your first Docker image by typing kubectl run nginx --image nginx. This will start the nginx Docker container in a pod on one of the nodes.

  7. Type kubectl get pods -o yaml to see the full details of the nginx deployment. You can see the host IP and the podIP. The pod IP is assigned from the pod CIDR on the host. Run curl to the pod ip to see the nginx output, eg. curl 10.244.1.4

Image of curl to podIP

  1. The next step is to expose the nginx deployment as a Kubernetes service on the private service network 10.0.0.0/16:
  2. expose the service with command kubectl expose deployment nginx --port=80.
  3. get the service IP kubectl get service
  4. run curl to the IP, eg. curl 10.0.105.199

Image of curl to service IP

  1. The final step is to expose the service to the world. This is done by changing the service type from ClusterIP to LoadBalancer:
  2. edit the service: kubectl edit svc/nginx
  3. change type from ClusterIP to LoadBalancer and save it. This will now cause Kubernetes to create an Azure Load Balancer with a public IP.
  4. the change will take about 2-3 minutes. To watch the service change from "pending" to an external ip type watch 'kubectl get svc'

Image of watching the transition from pending to external ip

  1. once you see the external IP, you can browse to it in your browser:

Image of browsing to nginx

  1. The next step in this walkthrough is to show you how to remotely manage your Kubernetes cluster. First download Kubectl to your machine and put it in your path:
  1. The Kubernetes master contains the kube config file for remote access under the home directory ~/.kube/config. Download this file to your machine, set the KUBECONFIG environment variable, and run kubectl to verify you can connect to cluster:
  • Windows to use pscp from putty. Ensure you have your certificate exposed through pageant:
# MASTERFQDN is obtained in step1
pscp -P 22 azureuser@MASTERFQDN:.kube/config .
SET KUBECONFIG=%CD%\config
kubectl get nodes
  • OS X or Linux:
# MASTERFQDN is obtained in step1
scp azureuser@MASTERFQDN:.kube/config .
export KUBECONFIG=`pwd`/config
kubectl get nodes
  1. The next step is to show you how to remotely run commands in a remote Docker container:
  2. Run kubectl get pods to show the name of your nginx pod
  3. using your pod name, you can run a remote command on your pod. eg. kubectl exec nginx-701339712-retbj date
  4. try running a remote bash session. eg. kubectl exec nginx-701339712-retbj -it bash. The following screen shot shows these commands:

Image of curl to podIP

  1. The final step of this tutorial is to show you the dashboard:
  2. run kubectl proxy to directly connect to the proxy
  3. in your browser browse to the dashboard
  4. browse around and explore your pods and services. Image of Kubernetes dashboard