Kubernetes - Basic Concepts

Kubernetes - Basic Concepts

Week 5 #90Days of DevOps challenge #devops #trainwithshubham

What is Kubernetes?

Kubernetes is an open-source container orchestration platform that automates the deployment, scaling, and management of containerized applications. It was originally developed by Google and is now maintained by the Cloud Native Computing Foundation (CNCF).

With Kubernetes, you can deploy containerized applications across multiple nodes, and manage and scale them more easily. Kubernetes provides a range of features such as load balancing, auto-scaling, self-healing, and rolling updates, which allow you to manage and maintain containerized applications with minimal downtime and manual intervention.

Kubernetes is designed to work with a variety of container runtimes, including Docker, containerd, and CRI-O. It also provides an extensive set of APIs and tools for configuring, monitoring, and managing containerized applications.

Kubernetes has become increasingly popular in recent years due to the rise of containerization and microservices-based architectures. It is used by many organizations, including Google, Microsoft, IBM, and Red Hat, to manage their containerized workloads.

Why use Kubernetes?

  1. Automated Scheduling:

    Kubernetes provides an advanced scheduler to launch containers on cluster nodes. It performs resource optimization.

  2. Self-Healing Capabilities:

    It provides rescheduling, replacing, and restarting the containers which are dead.

  3. Automated Rollouts and Rollbacks:

    It supports rollouts and rollbacks for the desired state of the containerized application.

  4. Horizontal Scaling and Load Balancing:

    Kubernetes can scale up and scale down the application as per the requirements.

  5. Resource Utilization:

    Kubernetes provides resource utilization monitoring and optimization, ensuring containers are using their resources efficiently.

  6. Support for multiple clouds and hybrid clouds:

    Kubernetes can be deployed on different cloud platforms and run containerized applications across multiple clouds.

  7. Extensibility:

    Kubernetes is very extensible and can be extended with custom plugins and controllers.

  8. Community Support:

    Kubernetes has a large and active community with frequent updates, bug fixes, and new features being added.

Kubernetes - Cluster Architecture

The architecture of Kubernetes is divided into several components that work together to provide a scalable and resilient platform for managing containerized applications. Here are the main components:

  1. Master node/Control Plane:

    The master node is the control plane of Kubernetes and manages the overall state of the cluster. It includes several components:

    • etcd: A distributed key-value store that stores the configuration and state of the cluster.

    • API server: The main endpoint for managing the cluster, which provides a RESTful API that clients can use to interact with Kubernetes.

    • Controller manager: A collection of controllers that are responsible for maintaining the desired state of the cluster, such as ensuring that the correct number of replicas of a service are running.

    • Scheduler: A component that schedules pods to run on worker nodes based on their resource requirements and availability.

  2. Worker node:

    • The worker nodes are the compute nodes in the cluster where the containers run. Each worker node includes several components:
  • Kubelet: An agent that runs on each worker node and is responsible for managing the containers on that node.

  • Container runtime: The software that runs the containers, such as Docker or containerd.

  • kube-proxy: A network proxy that forwards traffic to the correct container.

  1. Pods:

    A pod is the smallest deployable unit in Kubernetes and represents a single instance of a running process in the cluster. A pod can contain one or more containers, which share the same network namespace and can communicate with each other through localhost.

  2. Services:

    A service is an abstraction that represents a set of pods and provides a stable IP address and DNS name for clients to access them. Services can also provide load balancing and automatic failover.

  3. Volume:

    A volume is a directory that can be mounted inside a container, allowing it to access persistent storage.

  4. Namespace:

    A namespace provides a way to partition the resources in a cluster and is used to separate different applications or environments.

Working of Kubernetes

It is a collection of various components that help us in managing the overall health of a cluster. For example, if you want to set up new pods, destroy pods, scale pods, etc. 4 services run on Control Plane:

  1. Kube-API server : The API server is a component of the Kubernetes control plane that exposes the Kubernetes API. It is like an initial gateway to the cluster that listens to updates or queries via CLI like Kubectl. Kubectl communicates with API Server to inform what needs to be done like creating pods or deleting pods etc. It also works as a gatekeeper. It generally validates requests received and then forwards them to other processes. No request can be directly passed to the cluster, it has to be passed through the API Server.

  2. Kube-Scheduler : When API Server receives a request for Scheduling Pods then the request is passed on to the Scheduler. It intelligently decides on which node to schedule the pod for better efficiency of the cluster.

  3. Kube-Controller-Manager : The kube-controller-manager is responsible for running the controllers that handle the various aspects of the cluster’s control loop. These controllers include the replication controller, which ensures that the desired number of replicas of a given application is running, and the node controller, which ensures that nodes are correctly marked as “ready” or “not ready” based on their current state.

  4. etcd : It is a key-value store of a Cluster. The Cluster State Changes get stored in the etcd. It acts as the Cluster brain because it tells the Scheduler and other processes about which resources are available and about cluster state changes.

Also following are the nodes where the actual work happens. Each Node can have multiple pods and pods have containers running inside them. There are 3 processes in every Node that are used to Schedule and manage those pods.

  1. Container runtime : A container runtime is needed to run the application containers running on pods inside a pod. Example-> Docker

  2. kubelet : kubelet interacts with both the container runtime as well as the Node. It is the process responsible for starting a pod with a container inside.

  3. kube-proxy : It is the process responsible for forwarding the request from Services to the pods. It has intelligent logic to forward the request to the right pod in the worker node.

To be continued....