Detailed information about the working of Google Kubernetes Engine (GKE)

Siddhantpradhan
4 min readJul 5, 2022

Overview

Google Kubernetes Engine (GKE) provides a managed environment for deploying, managing, and scaling your containerized applications using Google infrastructure. The Kubernetes Engine environment consists of multiple machines (specifically Compute Engine instances) grouped to form a container cluster. In this lab, you get hands-on practice with container creation and application deployment with GKE.

Cluster orchestration with Google Kubernetes Engine

Google Kubernetes Engine (GKE) clusters are powered by the Kubernetes open-source cluster management system. Kubernetes provides the mechanisms through which you interact with your container cluster. You use Kubernetes commands and resources to deploy and manage your applications, perform administrative tasks, set policies, and monitor the health of your deployed workloads.

Kubernetes draws on the same design principles that run popular Google services and provides the same benefits: automatic management, monitoring and liveness probes for application containers, automatic scaling, rolling updates, and more. When you run your applications on a container cluster, you’re using technology based on Google’s 10+ years of experience with running production workloads in containers.

Kubernetes on Google Cloud

When you run a GKE cluster, you also gain the benefit of advanced cluster management features that Google Cloud provides. These include:

Create a GKE cluster

A cluster consists of at least one cluster master machine and multiple worker machines called nodes. Nodes are Compute Engine virtual machine (VM) instances that run the Kubernetes processes necessary to make them part of the cluster.

Note: Cluster names must start with a letter and end with an alphanumeric, and cannot be longer than 40 characters.

To create a cluster, run the following command, replacing [CLUSTER-NAME] it with the name you choose for the cluster (for example:my-cluster).

gcloud container clusters create [CLUSTER-NAME]

You can ignore any warnings in the output. It might take several minutes to finish creating the cluster.

Expected output:

NAME: my-cluster
LOCATION: us-central1-a
MASTER_VERSION: 1.22.8-gke.202
MASTER_IP: 34.67.240.12
MACHINE_TYPE: e2-medium
NODE_VERSION: 1.22.8-gke.202
NUM_NODES: 3
STATUS: RUNNING

Get authentication credentials for the cluster

After creating your cluster, you need authentication credentials to interact with it.

To authenticate the cluster, run the following command, replacing [CLUSTER-NAME] with the name of your cluster:

gcloud container clusters get-credentials [CLUSTER-NAME]

Expected output:

Fetching cluster endpoint and auth data.
kubeconfig entry generated for my-cluster.

Deploy an application to the cluster

You can now deploy a containerized application to the cluster. For this lab, you’ll run hello-app in your cluster.

GKE uses Kubernetes objects to create and manage your cluster’s resources. Kubernetes provides the Deployment object for deploying stateless applications like web servers. Service objects define rules and load balancing for accessing your application from the internet.

  1. To create a new Deployment hello-server from the hello-app container image, run the following kubectl create command:
  • kubectl create deployment hello-server — image=gcr.io/google-samples/hello-app:1.0
  1. Expected output:
  • deployment.apps/hello-server created
  1. This Kubernetes command creates a Deployment object that represents hello-server. In this case, --image specifies a container image to deploy. The command pulls the example image from a Container Registry bucket. gcr.io/google-samples/hello-app:1.0 indicates the specific image version to pull. If a version is not specified, the latest version is used.
  2. To create a Kubernetes Service, which is a Kubernetes resource that lets you expose your application to external traffic, run the following kubectl expose command:
  • kubectl expose deployment hello-server — type=LoadBalancer — port 8080
  1. In this command:
  • --port specifies the port that the container exposes.
  • type="LoadBalancer" creates a Compute Engine load balancer for your container.
  1. Expected output:
  • service/hello-server exposed
  1. To inspect the hello-server Service, run kubectl get:
  • kubectl get service
  1. Expected output:
  • NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE hello-server load balancer 10.39.244.36 35.202.234.26 8080:31991/TCP 65s Kubernetes ClusterIP 10.39.240.1 433/TCP 5m13s
  1. Note: It might take a minute for an external IP address to be generated. Run the previous command again if the EXTERNAL-IP column status is pending.
  2. To view the application from your web browser, open a new tab and enter the following address, replacing [EXTERNAL IP] with the EXTERNAL-IP for hello-server.
  1. Expected output: The browser tab displays the message Hello, world! as well as the version and hostname.

Deleting the cluster

  1. To delete the cluster, run the following command:
  • gcloud container clusters delete [CLUSTER-NAME]
  1. When prompted, type Y to confirm.
  2. Deleting the cluster can take a few minutes. For more information on deleted GKE clusters from the Google Kubernetes Engine (GKE) article, Deleting a cluster.

--

--