Home Docker and Kubernetes
Post
Cancel

Docker and Kubernetes

About Docker

  • docker is a tool for managing containers
  • container is a package of our code along with the dependencies and libraries to run that code
  • docker follows a client server architecture
    • we issue commands via cli to the docker client
    • all tasks like creating containers, pulling images, etc. is done by docker daemon (dockerd)
  • docker can be run natively in linux, so for macOS and windows, a virtualization layer is needed
  • docker engine - dockerd, docker client
  • docker desktop - docker engine, docker cli, kubernetes, docker compose, etc

Why use Docker

  • the same piece of code will always yield the same application i.e. doesn’t rely on host environment
  • having similar development, staging and production environments
  • easily manage different projects running different versions of dependencies
  • easily switch between versions of dependencies
  • virtual machines are not as easily reproducible as containers since they have their own dedicated OS
  • sharing and distributing is very convenient using Dockerfile, image, etc

docker vs vm

Images and Containers

  • images are templates for containers, and a container is a running instance of an image
  • containers are lightweight, isolated and run independent of each other
  • we can use official prebuilt images, the most common source is docker hub
  • note: while issuing docker commands
    • container name and container id can be used interchangeably, same for image
    • first few characters of the image_id are enough to reference the image if they can uniquely identify it
  • docker container run image_name to create a container from an image
  • if the image is not available locally, it is downloaded from dockerhub by docker
  • docker container ls to list all running containers
    • docker container ls -a to list all running as well as stopped containers
  • Dockerfile is a special file name, as it is the default file docker looks for when we build an image
  • Dockerfile contains the instructions for creating our own image
  • example of a Dockerfile
    1
    2
    3
    4
    5
    6
    
    FROM node:14-alpine
    WORKDIR /app
    COPY . .
    RUN npm install
    EXPOSE 80
    CMD npm run start
    
  • all commands except the last instruction CMD are used to build the image, CMD is used to run the container
  • so basically CMD is used for docker container run...
  • EXPOSE is only for documentation purpose
  • docker image build . - used to build an image using the Dockerfile, . here is the build context
    • -t flag to specify an image tag
  • images have layers i.e. docker caches result after every instruction in the Dockerfile
  • this means docker can reuse layers if possible - e.g. two different react applications use the same base image - node layer
  • so, to optimize i.e. make building of images faster, in the Dockerfile example shown earlier, we can first install dependencies and then copy the source code, as rebuilding of image will be triggered more frequently by a change in the source code than it will be by a change in the dependencies
    1
    2
    3
    4
    5
    6
    7
    
    FROM node:14-alpine
    WORKDIR /app
    COPY package.json .
    RUN npm install
    COPY . .
    EXPOSE 80
    CMD npm run start
    
  • docker container start container_id - start a stopped container
  • we can reattach our terminal to the container using docker container attach container_id
  • we can view logs using docker container logs container_id
    • add -f flag for following the logs
  • flags for docker container run -
    • -it can be used to enter interactive mode
    • --rm flag to delete the container when we stop it
    • --name to specify the name of container
    • -d to run in detached mode i.e. to not block our current terminal and run the container in foreground
    • -p flag means publish, i.e. map host port to a container port
  • docker image ls - lists the downloaded images
  • docker image rm image_id - remove the image with id image_id
  • docker container stop container_id - stop the container
  • docker container prune to delete all stopped containers
  • to get more information on images and containers, use docker container inspect container_id and docker image inspect image_id
  • docker container cp host_folder container_id:folder to copy folder from the host to the container
    • we can also reverse the order of arguments to copy folders and files from the container to the host
  • we can share images, by sharing the Dockerfile or by hosting it on an image registry like docker hub
  • docker image push image_name:tag to push images to the registry
  • docker image pull image_name:tag to pull images from the registry
  • we can also tag images using docker image tag new_image_name old_image_name
  • a full example of running a container - docker container run -d -p 3000:80 --name=backend --rm backend
  • docker login to login to docker hub
  • note: i had to generate and use a personal access token in the docker hub ui and use that instead of the docker hub password in the cli
  • docker image rm -f $(docker image ls -a -q) - deletes all locally downloaded images
    • -q helps list only image ids
    • -a helps list intermediate images as well
    • -f force removal, e.g. if image is referenced by another image
  • we can use a file .dockerignore to prevent copying files when using command COPY inside the Dockerfile e.g.
    1
    2
    3
    
    node_modules
    Dockerfile
    .git
    

Tags

  • an image tag has two parts - the name / repository of the image and the tag
  • tag is like a version, so we can generate different versions of our image
  • the default tag if not specified is latest
  • why tags are important -
    • rollback to previous versions in the production environment if newer versions have a bug
    • newer versions of other images which are used by our images might have breaking changes in future
  • suppose we always push and pull using tag latest. when we run docker container run..., it looks for the image locally and if it doesn’t find it, it goes online to fetch it. but it will find the image with the tag latest, and docker doesn’t understand that someone else has pushed a newer version online

Layered Architecture

  • all the docker related data like images, containers, etc. can be seen in /var/lib/docker
  • the docker image we build contains of layers, and these layers are shared across various images
  • e.g. if two images use the same base image, the layer of the base image is shared
  • the image layers are read only
  • when we create a container, a new layer is created on top of the existing layers of the image
  • thus, all writes that we perform during runtime, log files, etc. get written onto this layer
  • the persistence during the container’s lifetime happens through this writable layer
  • this mechanism is called copy on write, and the changes we made are lost unless we use volumes

Volumes

  • containers should be stateless as they can be easily created and destroyed, scaled up and down
  • we can have data that we want to persist even if containers are killed
  • this data shouldn’t be stored inside containers, or we may lose that data
  • volumes - mapping a persistent storage to docker containers. the persistent storage can be cloud storage, e.g. s3 of aws or our host directory system
  • this way, every time a container tries to persist changes, they go to the persistent storage and don’t get lost irrespective of how many times the container is started or stopped
  • volumes can be of three types -
    • anonymous volumes
    • named volumes
    • bind mounts
  • docker volume ls shows all volumes
  • anonymous volumes are managed by docker
  • the reference to anonymous volumes are lost after the container shuts down
  • if we use --rm flag while running the container, the anonymous volume is deleted as well
  • we can create anonymous volume by using VOLUME ./feedback inside the Dockerfile
  • we can also create anonymous volume by using flag -v /a/b during docker container run where /a/b is the path inside the container
  • named volumes are managed by docker too
  • unlike anonymous volumes, we don’t lose the reference to named volumes after the container is deleted
  • use flag -v to create named volumes, e.g. -v feedback:/app/feedback, where the name of the volume is feedback and the directory of the container it maps to is /app/feedback
  • bind mounts are managed by us. it can be used for source code, so that the changes that we make to the code get reflected in the container
  • in case of bind mounts, we have access to the folder which gets mapped to the container’s folder
  • in case of clashes, the more specific paths win e.g. if we are using bind mounts for /app of container and anonymous volumes for /app/node_modules of container, /app/node_modules relies on anonymous volumes
  • using nodemon with bind mounts prevents us from rebuilding images repeatedly i.e. our changes in source code are accounted for in the running container
  • we can use suffix :ro so that it specifies to the container that the volume is read only e.g. -v $(pwd)/app:ro, so that only hosts and not containers can edit the source code
  • note: docker volume ls will not list bind mount volumes, since it doesn’t manage them
  • docker volume rm volume_name to remove volumes
  • docker volume prune to remove volumes not being used by any containers
  • docker volume inspect volume_name to get details of the volume

Arguments and Environment Variables

  • docker supports build time arguments and runtime environment variables
  • runtime environment variables can be provided using ENV PORT 80 inside the Dockerfile
  • we can also provide it dynamically using -e PORT=80, which makes the earlier method a default
  • for situations like api keys where security is a concern, the method suggested above is better
  • we can also use build arguments, i.e. dynamic variables used when building an image
  • can be done using ARG PORT=80 in the Dockerfile
  • my understanding - so basically, arg is used by all commands above cmd and env is used by cmd?

Example

1
2
3
ARG DEFAULT_PORT=80
ENV PORT $DEFAULT_PORT
EXPOSE $PORT
  • we are giving the value of the build argument to the environment variable
  • if we don’t provide a port, the port used by container is 80
  • now, we can change the default port while building an image using docker image build ... --build-arg DEFAULT_PORT=9999 ...
  • we can also receive a dynamic port using docker container run ... -e PORT=9545 ...
  • if we don’t provide a port dynamically, the port specified for building of images gets used

Networks

  • there are three kinds of entities with which containers can communicate -
    • internet
    • host
    • other containers
  • containers can by default talk to the internet e.g. a public api
  • for containers to talk to the host, we can replace localhost by host.docker.internal
  • e.g. for containers to talk to mongodb running on our host machine, we can use mongodb://host.docker.internal:27017/favorites
  • for containers to talk to other containers, we can use docker container inspect ... to get the container’s ip address (available in the key IPAddress) and then use it. e.g. with a mongodb container running, we run docker container inspect mongodb and then use mongodb://the_ip_address:27017/favorites
  • this is not ideal, as this IP could change after a new container replaces the old one
  • we can create a docker network, and all containers placed inside the network can reference each other directly using the container names, e.g. mongodb://mongodb_container_name:27017/favorites
  • docker network create network_name to create a network
  • docker container run ... --network=network_name ... to create a container inside a specific network
  • also, we don’t need -p for the container to which another container connects, i.e. -p is only needed when we want our host port to map to the container port, not when another container wants to communicate with it
  • docker networks support different kinds of drivers. the default driver is bridge, which we saw above
  • there can be other types of drivers and third party plugins for drivers as well
  • we can use driver as “host” so that isolation between the container’s network and localhost is removed
  • examples of usage - docker network create --driver bridge or docker container run --network host
  • we can clean up unused networks using docker network prune
  • the bridge type of network uses network namespaces behind the scenes. so, on running ip addr, we see docker0, which is basically the virtual switch in network namespaces. each container is encapsulated inside its own network namespace. an e.g. is shown below -
    1
    2
    3
    
    docker container run nginx
    docker container inspect <<container_name>> | grep SandboxKey
    # the output is /var/run/docker/netns/<<namespace-name>>
    

Docker Compose

  • docker compose helps in preventing having to run docker commands from cli repeatedly
  • it has syntax in yml which is easier to read and can be shipped with our code
  • services in docker compose are containers, for which we can define environment variables, network, image, etc
  • version of docker compose I had to use was 3.8 based on my docker engine version
    • note - i think it comes with docker in newer version, i just have to use docker compose now
  • all container names are one level nested under the services key
  • can specify networks, volumes key for each container
  • for named volumes, we should also mention them under the volumes key in the root of the file
  • all the containers are a part of the default network created by docker-compose
  • docker-compose up starts all the containers and builds the images as well
    • flag -d can be used to start in detached mode
    • add flag --build to force the images to be rebuilt
  • docker-compose down - deletes all the containers and the default network that docker-compose creates
    • flag -v also removes the volumes which were created
  • use depends_on key to ensure the order in which containers start e.g. server depends_on mongodb container
  • docker-compose build to build the images
  • docker-compose run service_name to run a specific container in the compose file under the services key

Issues while Containerizing Frontend Apps

  • docker doesn’t work in the web browser for e.g. when we make xhr requests
    • so referring the backend application just by container name won’t work as it utilizes docker networks
    • so, we publish the backend on a host port and simply use localhost:that_port in frontend
  • reactJS needs the terminal to be in interactive mode to ensure it continues to run
    • it is like adding -it flag while using docker container run..., or setting stdin_open: true and tty: true inside of the docker compose

CMD and ENTRYPOINT

  • when we specify docker container run image_name xyz, xyz replaces what there is in CMD
  • however xyz appends what is there in ENTRYPOINT
  • we can replace what is there in ENTRYPOINT using --entrypoint
  • useful tip - since a space separated command needs to be a part of different items in an array, use sh -c. i.e. CMD ["a", "b", "c"] can become CMD ["sh", "-c", "a b c"]
1
2
3
FROM ubuntu
ENTRYPOINT [ "sleep" ]
CMD [ "10" ]
  • docker image build -t ubuntu-sleeper .
  • run docker container run ubuntu-sleeper, sleep is of 10 seconds
  • run docker container run ubuntu-sleeper 20, sleep is of 20 seconds
  • run docker container run -it --entrypoint=bash ubuntu-sleeper, run bash in interactive mode

Setup Containers

how do we set up initial project e.g. how to run npm init when we don’t have node installed locally? below is an example for setup using node

Dockerfile.setup -

1
2
FROM node:14-alpine
WORKDIR /app

docker-compose-setup.yml -

1
2
3
4
5
6
7
8
9
10
11
version: "3.8"
services:
  npm:
    build:
      context: ./
      dockerfile: Dockerfile.setup
    stdin_open: true
    tty: true
    volumes:
      - ./:/app
    entrypoint: npm

now, we can use commands to help during development like -

  • docker-compose -f docker-compose-setup.yml run npm init
  • docker-compose -f docker-compose-setup.yml run npm i express

the npm in the command is the service name inside docker compose, and entrypoint was given as npm in docker-compose, otherwise we would have to run docker-compose -f docker-compose-setup.yml run npm npm init

About Kubernetes

  • kubernetes is the most widely used container scheduler
  • modern infrastructure is created using immutable images, and an upgrade is performed by replacing the older images with newer ones using rolling updates
  • we specify how many resources to run and kubernetes maintains that number
  • it ensures that the resources run within the specified memory and cpu constraints
  • kubernetes is cloud-agnostic and can also be run on-prem
  • it has features like service discovery, load balancing, secret and configuration management, etc

Minikube and Kubectl

  • minikube allows us set up a single node cluster on our local workstation
  • minikube is useful for development purpose
  • kubectl is the kubernetes command line tool which allows to manage a kubernetes cluster
  • add alias to .bashrc - alias kubectl="minikube kubectl --"
  • configuring autocomplete for kubectl (restart terminal after running the command) -
    1
    
    echo 'source <(kubectl completion bash)' >> ~/.bashrc
    
  • minikube can be deployed as a vm or as a container (i am trying as a container for now)
  • configuring minikube -
    • minikube config set driver docker
    • minikube config set memory 8192
    • minikube config set cpus 4
  • view config using minikube config view or cat ~/.minikube/config/config.json
  • start minikube -
    1
    2
    
    minikibe start
    minikube status
    
  • pointing docker client installed locally to minikube’s docker daemon -
    1
    2
    3
    4
    
    docker container ls
    minikube docker-env
    eval $(minikube -p minikube docker-env)
    docker container ls
    
  • to ssh into minikube - minikube ssh. now also we can run commands like docker container ls etc
  • to get all running components, we can use kubectl get all --all-namespaces
  • to shut down minikube, use minikube stop. it preserves the state
  • to start minikube again, minikube start
  • to delete the cluster, minikube delete
  • can format output e.g. kubectl version --output=yaml. output format can be json as well
  • minikube ip to get the ip address of minikube cluster
  • an issue on my laptop - minikube cannot pull docker images at times. temporary fix is to pull manually using docker image pull after pointing docker client to minikube’s docker daemon

Cluster Architecture

  • the cluster has master nodes and worker nodes. note: there can be multiple masters in the cluster
  • the master nodes schedule and monitor the containers assigned to it on the worker nodes
  • different methods of viewing information related to the different components e.g. etcd
    • ps aux | grep etcd
    • sudo cat /etc/kubernetes/manifests/etcd.yaml
    • docker container ls | grep etcd
  • it has 7 major components as described below

Etcd

  • a distributed key-value store that allows for fast storage and retrieval
  • it runs on the port 2379
  • etcdctl is the etcd control client which helps communicate with etcd
  • it is used for storing and retrieving information about all kubernetes resources
  • the etcd clusters can either be present on the master nodes or be entirely decoupled from them
  • kubeadm runs etcd as a static pod on the master nodes
  • we specify its ip address and port on the api server
  • an example of using etcdctl api version 3 -
    1
    2
    3
    4
    5
    6
    
    kubectl exec etcd-minikube --namespace=kube-system -- sh -c \
      "ETCDCTL_API=3 etcdctl get / \
      --prefix --keys-only --limit=100 \
      --cacert /var/lib/minikube/certs/etcd/ca.crt \
      --cert /var/lib/minikube/certs/etcd/server.crt \
      --key /var/lib/minikube/certs/etcd/server.key"
    

    to get the location of the certs, use kubectl get pod etcd-minikube --namespace=kube-system --output=yaml

  • peer to peer communication in etcd clusters when there are multiple master nodes happens through 2380
  • etcd is distributed i.e. we can read from any of the instances, while all writes go to the master in the etcd cluster which syncs the data on the other replicas
  • in case of inconsistencies, the quorum determines if the update is valid. it is the minimum number of nodes in the etcd cluster which should have processed the update, which is floor(n / 2) + 1. the value of fault tolerance is total instances - quorum. so, it is recommended to have an odd number of etcd instances / master nodes depending on the configuration, since fault tolerance is the same for n and n - 1 nodes where n is even

Api Server

  • it runs on the master node
  • external clients like kubectl communicate changes to the cluster via the api server
  • schedulers, controllers, kubelets, etc. monitor the api server for new resources
  • they also send updates to the api server which then updates it on the etcd cluster
  • so, api server is the only component that directly interacts with the etcd cluster
  • the api server on the multiple master nodes can run concurrently i.e. all api servers on all the master nodes can be active at once. however, in case of controller manager and scheduler, to avoid duplication and inconsistencies, they are in the active state on the master node which is elected as the leader while they are in standby mode on the other master nodes
  • in case of multiple masters, clients like kubectl interact with a load balancer, where the load balancer routes requests to the multiple api servers

Controllers

  • also called controller manager
  • different kinds of controllers run on the master node
  • for instance, the master node expects heartbeats from the worker nodes. the node controller monitors them and if the heartbeats do not reach the master nodes for a certain time period, the pods on it are evicted
  • similarly, we have replication controller to maintain the number of pods of the same type
  • the controller manager package installs all the different controllers. to view the different controllers, use -
    1
    2
    
    kubectl get pod kube-controller-manager-minikube \
      --namespace=kube-system --output=yaml | grep controllers
    

Scheduler

  • runs on the master node
  • it assigns pods to a specific node
  • it does this based on available resources like cpu and memory and filters out nodes which cannot run the pod
  • it then based on a priority function ranks the remaining nodes
  • the pod then gets scheduled on one of the remaining nodes

Kubelet

  • it runs on all worker nodes (and optionally on the master node). see how this is different from the components seen above - etcd, api server, controller manager, scheduler
  • unlike the rest, kubelet does not run as a static pod, daemon set etc. it runs via a binary installed on the vms - i think this flow deviates since is doesn’t use static pods / daemon set
  • it registers the nodes with the cluster
  • picks up the pods from the api server to run on the node and then runs it
  • it then sends updates of the status of the pod to the api server
  • so, to view information, use ps aux | grep kubelet
  • this will show the files locations, so, for e.g., use - cat /var/lib/kubelet/config.yaml

Kube Proxy

  • it runs on all nodes, since it runs as a daemon set
  • pods in a node can reach pods on other nodes as well because of this
  • the kube proxy assigns an ip to the service
  • to view the ip range from which services are assigned ip addresses, we can use kubectl get pod kube-apiserver-minikube --namespace=kube-system --output=yaml | grep service-cluster-ip-range
  • it configures ip tables, which maps the ip address of services to endpoints
  • an endpoint = the ip address of the pod + port of the pod. this port can belong to any one of the containers, set by target port field in the service definition
  • if we have multiple pods sitting behind a service, an algorithm similar to round robbin is used

Kube Dns

  • it maps the service name to the service ip address
  • so, it configures the dns server
  • kubernetes uses coredns for achieving this functionality
  • on running kubectl get deployment coredns --namespace=kube-system --output=yaml, we can see that a config map is mounted as a volume on it
  • we get the contents of it using kubectl get configmap coredns --namespace=kube-system --output=yaml. it shows the plugins being used by coredns
  • there is also a service associated with kube dns, which we can get using kubectl get service kube-dns --namespace=kube-system --output=yaml | grep clusterIP. the pods point to this ip, which can be confirmed by inspecting the pod using kubectl exec any_pod_name -- cat /etc/resolv.conf

Pods

  • the smallest unit in kubernetes
  • represents a single running process
  • a pod encapsulates one or more containers, but usually we run only one container in a pod
  • sidecar pattern - helper containers can be spun alongside the application container in the same pod
  • to create a pod in an imperative way using commands, use kubectl run db --image=mongo
  • to get all running pods, use kubectl get pods
    • to get more information, we can use --output=wide, --output=yaml or --output=json
  • we can do a dry run and get the yaml, e.g. kubectl run db --image=mongo --dry-run=client --output=yaml
  • to see the list of events that occurred, use kubectl describe pod db
  • yaml file to create a pod declaratively
  • now, we can run kubectl apply -f file_name.yml
  • we can specify the file name in commands instead of the resource name - kubectl describe -f file_name.yml, kubectl delete -f file_name.yml, kubectl get -f file_name.yml
  • suppose we want to execute a command against a container. one way would be to issue commands using docker, e.g. docker container exec container_name command, this is just like spawning off another process in an already running container. however, we have to issue this command on a particular node of the cluster. this may not matter for minikube since everything is on our local but would matter for a production cluster. another way is to run kubectl exec pod_name -- command. this would by default execute the command on the first container of the pod. we can also specify the container using --container flag
  • to view logs, use kubectl logs pod_name. like in exec, the container can be specified explicitly
  • containers of the same pod run on the same node. they can talk via localhost i.e. if a container is running in a pod on port 8080, the other container can make requests to localhost:8080. they also share the same volumes
  • if we stop the container using docker container stop container_id, the pod will restart the container
  • to stop a pod, use kubectl delete pod db
  • when a pod is deleted
    • it first sends TERM (terminate) signal to all processes of all containers of the pod
    • if it does not stop within the gracePeriod, KILL signal is sent for a forceful shutdown
  • all containers part of the same pod coexist in the same node i.e. they cannot be distributed across the nodes
  • all of them can also access the same volume

Process of Creation

pod creation

ReplicaSets

  • it is a type of controller i.e. it tries to maintain a specified number of pods
  • this provides high fault tolerance, high availability, self-healing mechanism, etc.
  • replica sets are the newer version of replication controllers, since replication controllers are deprecated
  • note: when setting image using kubectl set image... in replica sets, i had to delete the older pods
  • yaml example
  • look how in yml syntax, spec.template is the exact same as that of the contents of a pod
  • labels defined for a pod should match the labels defined for its replica set i.e. value of spec.selector.matchLabels should match spec.template.metadata.labels else kubernetes gives an error
  • however, the replicaset can manage pods not defined in the spec.template section as well. in this case, the labels of pods should match the selector of the replica set
  • spec.replicas defines the number of pods to run
  • use kubectl get replicasets and kubectl get pods to verify
  • verifying the self-healing feature - if we try to delete a pod using kubectl delete pod pod_name, we will see that the replica set will automatically spin up a new pod
  • deleting the replica set will delete the pods it spun up as well
  • kubectl delete -f replica-set.yml --cascade=orphan. this will delete the replica set but not the pods. so, in general, to prevent removal of downstream objects, use the cascade flag

Process of Creation

only the first part has been described here, the remaining parts are similar to that of a pod

replica set creation

Services

  • pods are short-lived, so using addresses of pods for inter-pod communication is not reliable
  • services can be used to expose pods, replication controllers, replication sets, etc
  • the controller used here is called endpoint controller
  • service can be of different types
    • NodePort - target port on every node is exposed to the outside world. if we have multiple worker nodes, to hit a particular set of pods, we would have to use worker_ip:node_port. this also indicates that the services span multiple nodes without us having to configure anything
    • ClusterIP - this is the default. exposes the port only inside and not from outside the cluster
    • LoadBalancer - useful when deploying to cloud
    • ExternalName - map a service to an external address like a url
  • imperative command - kubectl expose replicaset rs --name=svc --target-port=28017 --type=NodePort
  • note: the node port cannot be specified when using kubectl expose
  • if we run kubectl describe service svc, we see that it has inherited all the labels of the replica set. recall how replica set is associated to the pods using labels, services are associated to the pods in the same way
  • when describing a service, it also shows all endpoints aka pods it directs traffic to
  • the three ports involved in node port are -
    • node port - how to access from outside the cluster. hit http://minikube_ip:NodePort. if not specified, a free port is chosen at random for its value
    • port - incoming traffic i.e. traffic from other pods or outside the cluster hit this port of the service
    • target port - port of the pod to which the service should forward traffic. if not specified, it takes the same value as port. so, in yml, usually only the port is specified
  • we can run kubectl get endpoints to get a list of all the endpoint objects. we can also get more information about a specific endpoint using kubectl get endpoints endpoint_name --output=yaml
  • we can run kubectl exec pod_name env - here, we will get environment variables like <<SERVICE_NAME>>_SERVICE_HOST, <<SERVICE_NAME>>_SERVICE_PORT. this will have the ip address and port of the different services respectively
  • communication - till now, we were using <<service-name>> for communication. it can be expanded to <<service-name>>.<<namespace-name>>. if we don’t specify the namespace-name, it defaults to the namespace in which the resource initiating the request is
  • communication to services can be further expanded to <<service-name>>.<<namespace-name>>.svc or <<service-name>>.<<namespace-name>>.svc.cluster.local. this bit can be confirmed using kubectl exec any_pod_name -- cat /etc/resolv.conf under the search field
  • by default, direct communication to pods is not enabled. if we enable it, we can use <<modified-pod-ip>>.<<namespace-name>>.pod.cluster.local. here, the modified-pod-ip is constructed by replacing . with - i.e. 10.244.2.5 becomes 10-244-2-5. again, because of the search field in /etc/resolv.conf, we can skip cluster.local (but not pod since svc is the default). my doubt - at this point, i might as well use the pod’s ip address directly, i mean pod’s ip is directly used using modified ip anyway

Process of Creation

service creation

Liveliness Probe

  • used for configuring health checks, done at a container level
  • if the health check fails, it applies the restart policy which defaults to always
  • the restart policy is specified at the pod level and applies to all containers
  • initialDelaySeconds - when should the probe start
  • timeoutSeconds - after waiting for how many seconds should the probe fail
  • periodSeconds - after how many seconds should the probe be repeated
  • failureThreshold - how many consecutive health checks are allowed to fail
  • code example
    1
    2
    3
    4
    5
    6
    7
    8
    9
    10
    
    name: api
    image: user-service
    livenessProbe:
      httpGet:
        path: /actuator/health
        port: 8080
      initialDelaySeconds: 20
      timeoutSeconds: 5
      periodSeconds: 5
      failureThreshold: 3
    

Readiness Probe

  • it is used to determine whether a pod is ready to serve requests
  • it has the same configuration as liveliness probe
  • ip addresses of unhealthy pods are removed from ip tables, so that the future requests do not make it to them

An Example

  • a complete example of
    • a database and exposing it using cluster ip
    • backend service which talks to db, exposing it using node port, configuring health checks

Deployments

  • helps us achieve zero downtime when we deploy services
  • we should not create pods or even replica sets directly
  • deployments create replica sets behind the scenes
  • when we make an update to for e.g. the image version, the deployment will first create a new replica set with the desired number of pods, and once that replica set has successfully scaled the pods, the deployment would mark the desired replicas of the older replica set as 0. a part of kubectl describe deployment db -
    1
    2
    3
    4
    5
    
    Type    Reason             Age    Message
    ----    ------             ----   -------
    Normal  ScalingReplicaSet  12m    Scaled up replica set db-5cc56bf6fb to 1
    Normal  ScalingReplicaSet  4m22s  Scaled up replica set db-76774bbdf to 1
    Normal  ScalingReplicaSet  92s    Scaled down replica set db-5cc56bf6fb to 0
    
  • a side note - the random characters that we see are actually the hash value of the pod template
  • to create a deployment imperatively, use kubectl create deployment nginx --image=nginx --replicas=2
    • we can also add flags --dry-run=client --output=yaml to generate the yaml
  • deployment strategy can be rolling update (default) or recreate
  • in recreate, the old pods are stopped and new ones are created in its place. this leads to some downtime. use recreate when the coexistence of two versions of the applications can cause inconsistencies e.g. db migrations
  • in rolling deployments, the new replica set is scaled up and the old replica set is scaled down simultaneously gradually. they can be tweaked using maxSurge and maxUnavailable fields. at any given time, we can have a maximum of desired + maxSurge or a minimum of desired - maxUnavailable pods running. both can be absolute numbers or % and both default to 25%. since both versions of applications run in parallel, the response can be returned from either of the versions at random during deployment
  • e.g. of rolling deployment - by using the following code, the deployment order is 3 old ➝ 3 old, 1 new ➝ 2 old, 1 new ➝ 2 old, 2 new ➝ 1 old, 2 new ➝ 1 old, 3 new ➝ 3 new
    1
    2
    3
    4
    5
    6
    
    replicas: 3
    
    strategy:
      rollingUpdate:
        maxSurge: 1
        maxUnavailable: 1
    
  • everytime we deploy in kubernetes, a rollout takes place and a revision is created
  • we can monitor the status of the update to deployment using kubectl rollout status -f deployment.yml
  • we can view the history of updates using kubectl rollout history -f deployment.yml
  • we can also create a rollback using kubectl rollout undo -f deployment.yml
    • if we want to go back to a much older version and not just the previous one, we can use kubectl rollout undo -f deployment.yml --to-revision=2
  • side note: rollbacks might not always be possible e.g. if we had database migrations. so, we may need to roll forward in some cases i.e. implement a hot fix and redeploy the new changes
  • using labels -
    • kubectl get all --show-labels - show the resources with their labels
    • kubectl get all --selector=name=db,app=demo - filter the resources using their labels
    • e.g. to count the total number of resources in dev environment, use kubectl get all --selector=env=dev --no-headers | wc -l
  • we can set image of a deployment using kubectl set image deployment db db=mongo:3.3, where the first db is the deployment name and the second db is the container name, since we can have multi container pod
  • to add the default change cause to the kubectl rollout history ... output, append commands with --record, e.g. kubectl apply -f infra --record. this flag is deprecated but i cannot find its replacement
  • to scale deployments imperatively, use kubectl scale deployment api --replicas=2
  • both in deployments and in services, any one of the labels on pod need to be present in spec.selector

Process of Creation

  • a deployment controller will watch for new deployment creation requests
  • it will then create replica set definitions on api server
  • after this, the process of replica set creation is continued

Imperative vs Declarative

  • in declarative, we just tell the desired state which kubernetes tries to achieve
  • e.g. apply follows the declarative approach
  • however, in the imperative approach, we have to give clear instructions
  • all commands like create, edit, replace, expose, run etc. are imperative
  • using declarative approach we can track configuration using version control as well for iac
  • imperative approach can be used for hot fixes / experimental purpose
  • when using apply, we can see the last yaml configuration converted to json which we had sent under metadata.annotations in kubectl.kubernetes.io/last-applied-configuration. this is used by kubernetes to keep track of changes and is only available when we use apply
  • if for e.g. we use edit to edit a resource, and that resource is not allowed to be edited, we just use wq to exit out of vim, and then that file gets saved to /tmp. we can then use kubectl replace --force -f <<file>> to replace the existing resource with our newly configured one

Ingress

  • it is like a layer 7 load balancer built inside the kubernetes cluster
  • makes the services inside cluster accessible from outside
  • we also want features like ssl termination, route requests based on domain, etc
  • my understanding - recall how a service based on labels can only expose a set of pods. instead of multiple node ports / load balancers i.e. one for each set of pods, we have one node port / load balancer which directs traffic to the ingress service. the ingress service can then direct traffic to the different cluster ips in the cluster
  • kubernetes provides the ingress resource but not the ingress controller i.e. it provides the api which can be utilized by other third party implementations
  • minikube has an addon that can be enabled
    1
    2
    
    minikube addons enable ingress
    minikube addons list | grep ingress
    
  • to verify, kubectl get all --all-namespaces should show the ingress-nginx-controller-* pod running
  • ingress is spun up using a deployment and a node port to expose it outside the cluster
  • it also deploys configmaps to manage configuration and cluster roles to monitor kubernetes resources
  • all resources are deployed in the ingress-nginx namespace
  • we can also hit the endpoint http://minikube_ip/healthz to verify the working of ingress
  • we can also provide a domain so that the requests are routed based on domain names
  • we can also provide a catch-all entry
  • in this example, if requests come from custom-api.com, and start with request path /api they are routed to the api service, but all other requests are routed to the devops service
  • note: to simulate that requests are coming from a specific domain on our local, we can use curl -H "Host: custom-api.com" http://192.168.49.2/api/
  • the ingress resource provided by kubernetes has limited functionality, so to configure the ingress controller provided by third party, we use annotations
  • e.g. we want traffic from ingress-service/calendar to our calendar-cluster-ip:port. so, the calendar prefix should be removed. we can do this by using the annotation below -
    1
    2
    
    annotations:
      nginx.ingress.kubernetes.io/rewrite-target: /
    
  • so, /calendar in rules[x].http.paths.path gets replaced by the value in rewrite-target which is / here
  • use kubectl get ingress to view the ingress resources
  • my understanding - view the port of the node port service ingress-nginx-controller inside the nginx-ingress namespace. this is the port we hit when making requests to worker_node_ip

Volumes

  • references to files and directories made available to containers
  • the file system can be anywhere, e.g. outside the host as well i.e. this could be used for aws ebs as well
  • e.g. it helps us preserve data across pod restarts
  • there can be several types of volumes like host path, git repo (like host path but the path is a git repository) and even cloud specific like aws elastic block store
  • empty dir volume type - if a container crashes, a new container is spun up in the same pod. however, if we don’t specify any volume, the container crash results in a loss of data. this can be prevented using empty dir volume type, which can survive container restarts but not pod restarts. it is usually chosen as the default by third party manifests and is expected to be replaced by a better solution like nfs
  • an issue with using host volumes - it needs to be available on each node so that pods on different nodes can have access to it, and this of course is not an issue with minikube
  • so, in cloud, we should ideally mount an nfs on each node, else we would have to copy this file on all nodes. for e.g., we should use aws efs. the syntax should be similar

Example 1

  • for docker client to be able to communicate to the correct docker daemon, use the file /var/run/docker.sock
  • e.g. we want to run docker commands on the host from pods. so, the container running inside the pod should have docker client installed to issue docker commands, and it should point to the docker daemon of the host
  • so, we can use hostPath volume type
  • full yaml here
  • now, we run the pod using kubectl apply -f docker.yml
  • then, we can issue commands like kubectl exec docker -- docker image ls to list the images on minikube

Example 2

  • recall how for bind volumes in docker, we needed to specify a path in the host. the host now is minikube, so the host path needs to be that of minikube. before running minikube start, if i copy files to the path in ~/.minikube/files directory on my workstation, i can see those files in the root on minikube host. we can verify this using minikube ssh and then by running ls /
  • so, suppose we want to specify a configuration file for prometheus
  • we can copy this configuration file to minikube and then use host path volumes to reference it
  • in this example, a better solution would have been to create a custom image and use COPY in the docker file
1
2
3
4
5
6
7
8
9
10
11
12
13
# ...
spec:
  containers:
    - # ...
      volumeMounts:
        - mountPath: /etc/prometheus/prometheus.yml
          name: prom-conf

  volumes:
    - name: prom-conf
      hostPath:
        path: /prometheus-conf.yml
        type: File

Config Maps

  • we can have different sources of configuration like environment variables, files, env files, literals, etc

Default Config Map

  • it is used to make calls to the kubernetes api from containers
  • kubectl get configmaps - kube-root-ca.crt is the config map created by default
  • kubectl describe pods pod_name will give the mount location of this config map. note that this config map may not be mounted to the pods in kube-system though
  • kubectl exec pod_name -- ls /var/run/secrets/kubernetes.io/serviceaccount shows that there are three files - namespace, ca.crt and token
  • on reading online, i see that this can also be a secret instead of a configmap

Mount Volumes

  • config maps can mount the configuration as volumes to running containers
  • imperative command - kubectl create configmap prometheus-config --from-file=prometheus-conf.yml
  • kubectl describe configmap prometheus-config
  • using in the yml file -
    1
    2
    3
    4
    5
    6
    7
    8
    9
    10
    11
    
    spec:
      containers:
        - # ...
          volumeMounts:
            - mountPath: /etc/prometheus
              name: prometheus-config
      # ...
      volumes:
        - name: prometheus-config
          configMap:
            name: prometheus-config
    

    verify using kubectl exec prometheus -- cat /etc/prometheus/prometheus-conf.yml

  • instead of providing a file, we can use literals, e.g. kubectl create configmap --from-literal=foo=bar
  • in this case, if we use volume mounts, a file called foo would be created with its contents as bar

Environment Variables

  • e.g. create a file called .env
    1
    2
    
    client-id=qwerty
    client-secret=12345
    
  • kubectl create configmap api-credentials --from-env-file=.env
  • my understanding - the difference between --from-file vs --from-env-file is in from file, kubernetes does not care about the file’s content, while in from env file, it knows how to treat as different key value pairs, so that it can inject them all at once / individually as discussed below
  • usage -
    1
    2
    3
    4
    5
    
    containers:
        #...
        envFrom:
          - configMapRef:
              name: api-credentials
    
  • verify using kubectl exec alpine -- env
  • we can also inject the variables of the config map individually -
    1
    2
    3
    4
    5
    6
    7
    8
    
    containers:
        #...
        env:
          name: CLIENT_ID
          valueFrom:
            configMapKeyRef:
              name: api-credentials
              key: client-id
    

Secrets

  • secrets are similar to config maps
  • secrets can be of three types -
    • docker-registry - for pulling images from private registry
    • tls - for storing certificates
    • generic - works like config maps, so can have sources like --from-env-file, --from-file, --from-literal
  • creating a secret imperatively -
    1
    2
    3
    
    kubectl create secret generic jenkins-credential \
      --from-literal=username=johndoe \
      --from-literal=password=incognito
    
  • to retrieve the original value -
    1
    2
    
    kubectl get secret jenkins-credential --output=json
    kubectl get secret jenkins-credential --output=jsonpath="{.data.password}" | base64 --decode
    
  • to use the secrets, we put them into files /etc/secret/jenkins-user and /etc/secret/jenkins-pass -
    1
    2
    3
    4
    5
    6
    7
    8
    9
    10
    11
    12
    13
    14
    15
    16
    17
    
    spec:
      containers:
        - # ...
          volumeMounts:
            - mountPath: /etc/secrets
              name: jenkins-credentials
    
      volumes:
        - name: jenkins-credentials
          secret:
            secretName: jenkins-credential
            defaultMode: 0444
            items:
              - key: username
                path: jenkins-user
              - key: password
                path: jenkins-pass
    
  • we made it read only for all users using 0444 as the mode
  • verify using kubectl exec pod_name -- cat /etc/secrets/jenkins-pass
  • if creating secrets declaratively, the values should be base64 encoded first
    1
    2
    3
    
    # ...
    data:
      username: am9obmRvZQ==
    

    using sh base64 utility -

    1
    2
    3
    4
    5
    
    # to encode
    echo -n johndoe | base64
      
    # to decode
    echo -n am9obmRvZQ== | base64 --decode
    
  • the only difference between config maps and secrets is that secrets are stored in tmpfs (temporary file storage) thus leaving no trace on the nodes
  • secrets should be combined with rbac for limited access
  • cons of using kubernetes secrets: secrets are stored in plain text in etcd, so anyone with access to etcd can read the secrets. so, we should use solutions like hashicorp vault, integrating it with kubernetes is smooth

Docker Registry

  • by default, we use public docker registry
  • sometimes we might need private registry
  • we use docker login when using vanilla docker
  • when using kubernetes, we can create the secret of type docker-registry
    1
    2
    3
    4
    5
    
    kubectl create secret docker-registry registry-credential \
      --docker-server=...\
      --docker-username=...\
      --docker-password=...\
      --docker-email=...
    
  • we can then specify the name of the secret in pod
    1
    2
    3
    
    spec:
      imagePullSecrets:
        - name: registry-credential
    

Namespaces

  • we can spin up multiple clusters to isolate the different environments. this can help prevent accidental changes to the production cluster
  • however this has operational and resource overhead
  • namespaces help us create different segments on a cluster
  • namespaces are like virtual clusters
  • we can scope resource limits and permissions to namespaces
  • we use the “default namespace” by default
  • we can run kubectl get namespaces to view all the available namespaces
  • kube-public - the resources in this namespace are accessible to all (including unauthenticated) users
  • kube-system - kubectl get all --namespace=kube-system shows the resources managed by kubernetes itself
  • to create a namespace, use kubectl create namespace staging
  • if we set the namespace in context, we do not have to repeatedly suffix commands by --namespace=staging. e.g. we can use kubectl config set-context $(kubectl config current-context) --namespace=staging
  • when we delete a namespace, the cascading effect deletes all the resources within it as well. the command is kubectl delete namespace staging
  • in the resource files, we can also specify the namespace key under metadata

RBAC

  • it allows us to control access to resources
  • each request goes through three stages - authentication, authorization, and then through admission control
  • authentication is done using static password files, static token files, certificates or identity services like ldap
  • we can combine rbac with namespaces as well

Static Files

  • we create a csv file with 4 columns - password, username, user id and optionally a group
  • when starting the api server, we pass the file using a flag --basic-auth-file=user-details.csv
  • then, if sending requests using curl, we can pass credentials using -u username:password
  • we can also have tokens instead of passwords
  • the flag we pass to api server in this case is --token-auth-file
  • we can pass the token via curl using --header "Authorization: Bearer <token>"
  • this method is deprecated / unavailable in newer versions

Certificates

  • we need to be able to generate certificates for creating users
  • we need two files -
    • the certificate which is a signed public key (files suffixed with .crt)
    • a private key (files suffixed with .key)
  • certificates are needed by both servers and clients. there can be multiple combinations of server and client as well. the certificates required are listed below -
    • 7 client certificates
      • 3 - scheduler, controller manager, kube proxy (to api server)
      • api server (to kubelet)
      • kubelet (to api server)
      • api server (to etcd)
      • user using kubectl (to api server)
    • 3 server certificates - etcd, kubelet, api server
    • the ca’s certificate
  • etcd cluster can have multiple nodes running on different servers for high availability
  • so, we need to generate peer certificates so that the communication between the nodes is encrypted as well
  • we can specify a config file while creating certificates using --config
  • config can be needed for e.g. to provide alternative names
  • docs for the certificates required by kubernetes components and their respective cas
  • verify if openssl version works to be able to generate certificates
  • generating keys and certificates -
    1
    2
    3
    4
    5
    6
    7
    8
    9
    10
    11
    12
    13
    14
    15
    16
    17
    
    openssl genrsa -out johndoe.key 2048
    
    openssl req -new \
      -key johndoe.key \
      -out johndoe.csr \
      -subj "/CN=johndoe/O=developers"
    
    openssl x509 -req \
      -in johndoe.csr \
      -CA ~/.minikube/ca.crt \
      -CAkey ~/.minikube/ca.key \
      -CAcreateserial \
      -out johndoe.crt \
      -days 365
    
    # view the certificate details
    openssl x509 -noout -text -in johndoe.crt
    
  • first a private key (.key) is generated
  • then a certificate signing request (.csr) is generated. here CN or common name is analogous to the username and O or organization to the group
  • then, using the csr and the ca certificate, the signed public key or the signed certificate (.crt) is generated
  • instead of doing the steps after generating the csr manually, we can utilize the kubernetes api
  • basically a resource type of CertificateSigningRequest is created, with the contents of the csr mentioned in a base64 encoded format in the yaml file
  • the admin can then run kubectl get certificatesigningrequests to view pending requests
  • requests can be approved via kubectl certificate approve <name_of_csr>
  • similarly, requests can be denied via kubectl certificate deny <name_of_csr>
  • to view the generated signed certificate, we can use kubectl get certificatesigningrequest <name_of_csr> --output=yaml. it is again in the base64 encoded format, so we decode and store it in a file

Config

  • we can specify flags like --key, --cert and --cacert when making requests via curl to api server
  • we can also specify flags in kubectl everytime, e.g. kubectl get pods --client-key=... --client-certificate=... --certificate-authority=...
  • by default for all our requests using kubectl, the configuration is specified in ~/.kube/config. it also is like a kubernetes resource with kind: Config. so, instead of defaulting to ~/.kube/config in every command, we can specify the file using the --kubeconfig flag
  • it has three parts - clusters, users and contexts
  • clusters refer to the different kubernetes clusters that we would like to access
  • the cluster requires the path to the ca server certificate and the api server address
  • to get the server address, use kubectl config view --output=jsonpath="{.clusters[0].cluster.server}"
  • the user requires the path to private key and signed certificate
  • we can also provide the base64 encoded data directly instead of the path for the user / cluster
  • contexts pair the clusters to users. so, they have the cluster, user and even the namespace to use by default
  • the one used by default by kubectl is defined via current-context
  • create a new cluster -
    1
    2
    3
    4
    5
    6
    7
    
    cp ~/.minikube/ca.crt .
      
    kubectl config set-cluster johndoe \
      --certificate-authority ca.crt \
      --server https://192.168.49.2:8443 # cluster server address
      
    kubectl config get-clusters # verify that the cluster is created
    
  • create a new user -
    1
    2
    3
    
    kubectl config set-credentials johndoe \
      --client-certificate johndoe.crt \
      --client-key johndoe.key
    
  • create and set the context -
    1
    2
    3
    4
    5
    6
    7
    
    kubectl config set-context johndoe \
      --user johndoe \
      --cluster johndoe # create / edit the context
      
    kubectl config use-context johndoe # change the context
      
    kubectl config get-contexts # verify that the context is set
    
  • we can view the entire config using kubectl config view or cat ~/.kube/config
  • note: the context section can also take the namespace as an argument

Authorization Modes

  • node - e.g. used by worker nodes, for kubelet to interact with api server
  • rbac - attaching policies to role
  • abac - attaching policies to users / groups directly
  • webhooks - a third party agent like open policy agent
  • to view authorization modes configured on the api server, we can use - kubectl describe pod kube-apiserver-minikube --namespace=kube-system | grep authorization-mode

Roles and Bindings

  • rules, comprise of -
    • verbs, e.g. get, list, create
    • resources, e.g. pods
    • resource names
    • api groups of the resources
  • roles - they are a collection of rules. a role is applied to a namespace
  • cluster role - same as roles but scoped to clusters
  • resources like pods are namespaced while resources like nodes are cluster scoped. to get an exhaustive list, we can use kubectl api-resources --namespaced=true or set the flag to false
  • subjects - can be user, service accounts (used by pods to interact with kubernetes api) or groups, which are a collection of users and service accounts
  • we also have role bindings and cluster role bindings
  • kubectl config get-users - by default we have only one user minikube
  • a few commands we can use include - kubectl get roles, kubectl get clusterroles, kubectl get rolebindings, kubectl get clusterrolebindings
  • we can also use kubectl describe clusterrole view
  • we already have some cluster roles and cluster role bindings created by default
  • ones prefixed with system: should be generally avoided, so we can run kubectl get clusterroles | grep -v system. we get four roles, each of them has been described below
  • view - can perform get, list and watch operations (verbs) on almost everything
  • edit - everything that admin can do except modify roles and role bindings
  • admin - everything that cluster-admin can do except modification to namespaces and resource quotas
  • cluster-admin - can perform all operations. e.g. the default user minikube has this role. this can be verified by running kubectl auth can-i "*" "*"
  • to verify if an operation can be performed, we can use for instance kubectl auth can-i get pods
  • we can impersonate as someone else using kubectl auth can-i get pods --as=johndoe
  • creating a role binding -
    1
    2
    3
    4
    
    kubectl create rolebinding johndoe \
      --clusterrole=view \
      --user=johndoe \
      --namespace=default
    
  • verify using kubectl describe rolebinding johndoe. note: sometimes in kubectl outputs, the namespace field is empty when it is referring to the default namespace
  • my understanding - role bindings can reference cluster roles, it just means that the permissions would be granted on the specified namespace only. this allows for role reuse. the view role allows to view in any namespace, and by creating a role binding we can limit the user’s usage to a namespace
  • delete a role binding using kubectl delete rolebinding johndoe
  • using role bindings, we can attach one role to multiple subjects
  • declaratively creating a cluster role binding using yaml -
    1
    2
    3
    4
    5
    6
    7
    8
    9
    10
    11
    12
    13
    14
    15
    16
    
    apiVersion: rbac.authorization.k8s.io/v1
    
    kind: ClusterRoleBinding
    
    metadata:
      name: johndoe-view
    
    roleRef:
      apiGroup: rbac.authorization.k8s.io
      kind: ClusterRole
      name: view
    
    subjects:
      - apiGroup: rbac.authorization.k8s.io
        kind: User
        name: johndoe
    
  • note how role bindings have a single role but can have multiple subjects
  • verify using kubectl auth can-i get pods --as=johndoe --all-namespaces
  • describing the admin cluster role created by default - kubectl describe clusterrole admin
  • e.g. yml to create a custom role -
    1
    2
    3
    4
    5
    6
    7
    8
    9
    10
    11
    12
    13
    14
    
    apiVersion: rbac.authorization.k8s.io/v1
      
    kind: ClusterRole
      
    metadata:
      name: release-manager
      
    rules:
      - resources: ["pods", "pods/attach", "pods/exec", "pods/log", "pods/status"]
        verbs: ["*"]
        apiGroups: [""]
      - resources: ["deployments", "replicasets"]
        verbs: ["create", "get", "list", "update", "watch"]
        apiGroups: ["", "apps", "extensions"]
    
  • note: to grant permissions for different operations on pods, specifying the resources as pods is not enough, as there can be other sub resources like pods/logs etc
  • instead of users, we can also use groups in role bindings and cluster role bindings, by changing the kind to groups. the group a user belongs to was specified while generating the certificate
  • my understanding - when relying on groups for role bindings, using the correct context is important when running kubectl auth can-i... command, since the certificates associated with the context are used to validate the request. by just using --as, we specify the user, but for e.g. the group comes from the value of O in the subject when generating signed certificates. optionally, we can specify the flags like --client-certificate, --client-key etc

Service Accounts

  • it is used by applications to interact with the kubernetes cluster
  • if we run kubectl get serviceaccounts, we see the default service account is already present
  • we can create a service account using kubectl create serviceaccount app
  • recall the default config map
  • creating a service account creates a secret, which has a token behind the scenes
  • so, we can use these tokens to make requests, by adding the header - Authorization: Bearer <token>
  • the method above is useful when the functionality is outside the kubernetes cluster
  • however, if the application is a part of our kubernetes cluster, we can simply mount the service account. i.e. use the serviceAccountName field in pods
  • we can also manually set the automountServiceAccountToken to false in order to disable the automatic mounting of the default service account
  • in newer versions of kubernetes, some changes have been made - service accounts now no longer automatically have a token associated with them. instead, the token request api is used to generate a token which then gets mounted on the pods
  • we can generate a token for a service account using kubectl create token <<service-account-name>>, and provide configuration parameters like expiration time

NodeName

  • the scheduler schedules a pod on any node
  • if we run kubectl get pod pod_name --output=yaml, we can see the node it was scheduled on under nodeName
  • behind the scenes, a binding object is created which binds the pod to a node
  • we can manually specify the node a pod should be scheduled on using the nodeName property
  • we can use this if we didn’t have a scheduler, and this would schedule the pod on the specified node

Taint and Toleration

  • taint is set on nodes which prevent any random pod from being scheduled on it
  • toleration is set on pods which allows them to be scheduled on a node with taint
  • by default, the pods have no toleration
  • use case - a worker node has resources to enable running of a certain type of pod
  • it means that only pods with toleration can be scheduled on the node with that taint
  • however, the pods with this toleration can be scheduled on other nodes as well
  • this feature is used by kubernetes as well to help ensure that normal pods are not scheduled on the master and only the management pods scheduled by kubernetes itself are
  • to taint nodes, use kubectl taint node node_name key=value:taint-effect
  • similarly, to remove the taint, use kubectl taint node node_name key=value:taint-effect-, i.e. suffix the prior command with a - symbol
  • taint effects can be -
    • NoSchedule - do not schedule any new pods without the right toleration
    • PreferNoSchedule - prefer not scheduling
    • NoExecute - like NoSchedule but also evicts the existing pods on the node without the correct toleration
  • to apply toleration on pods, use -
    1
    2
    3
    4
    5
    6
    
    spec:
      tolerations:
        - key: key
          operator: Equal
          value: value
          effect: NoSchedule
    

Node Selectors and Node Affinity

  • we add labels to nodes and then add selectors for them to pod definitions
  • this way, the pods with the node affinity can only be run only on specific nodes
  • however, pods without the node affinity can still be spun up on the nodes with labels
  • to label a node, we use - kubectl label node node_name key=value
  • we can use kubectl get nodes --show-labels to verify
  • to apply selectors for labels on nodes, use -
    1
    2
    3
    
    spec:
      nodeSelector:
        size: large
    
  • but, using node selectors we cannot specify complex conditions
  • so, we use node affinity
    1
    2
    3
    4
    5
    6
    7
    8
    9
    10
    11
    
    spec:
      affinity:
        nodeAffinity:
          requiredDuringSchedulingIgnoredDuringExecution:
            nodeSelectorTerms:
              - matchExpressions:
                  - key: size
                    operator: In
                    values:
                      - large
                      - medium
    
  • what if the node labels are changed after the pod was already scheduled? what if there are no nodes found with the conditions matching the node affinity value? for these, the value can be requiredDuringSchedulingIgnoredDuringExecution or preferredDuringSchedulingIgnoredDuringExecution
  • my understanding - requiredDuringSchedulingRequiredDuringExecution is not available by default
  • some operators to use - Equal, In, Exists
  • so, overall, to ensure pods of a particular type and only this type end up on a particular node, we need to use node selectors / node affinity and taints and tolerations in conjunction

Resource Management

  • we can give an indication and set limits for the resources that can be used by kubernetes components
  • specified at the container level
  • this helps kubernetes in scheduling
  • to enable metrics server, use - minikube addons enable metrics-server
  • can be written as for e.g. 0.5 or 500m (500 milli cpu). 1 milli cpu is equivalent to 1 hyperthread / 1 vcpu
  • memory can be written as K or Ki for kilobyte, M or Mi for megabyte and so on. we can only specify the numerical value as well, its value is in bytes e.g. 256Mi or 268435456
  • syntax -
    1
    2
    3
    4
    5
    6
    7
    8
    9
    
    containers:
        #... 
        resources:
          limits:
            memory: 100Mi
            cpu: 200m
          requests:
            memory: 50Mi
            cpu: 100m
    
  • limits - amount of resources that containers should not cross
  • if the container crosses the memory limit, it will be terminated / restarted. the pod has status OOMKilled (out of memory killed). the pod remains the same, the container changes
  • containers are not allowed to use more than the cpu limit for an extended period, so there are no restarts / termination of the containers for crossing the cpu limits as cpu usage gets throttled automatically
  • requests - amount of resources that containers are expected to use
  • only when the node runs out of memory, the pod that the container exceeding the requests is a part of is evicted from the node, and it gets rescheduled
  • if a container’s memory request exceeds the available memory on any node (technically sum of the memory requests of all the containers of a pod), the pod stays in Pending state indefinitely
  • if the memory usage exceeds only the requested amount (and not the limit), the pod can be evicted if another pod enters with a higher qos and needs that memory
  • so, memory limit cannot be exceeded while memory request can be exceeded if the node has enough memory
  • kubectl describe nodes gives details of the available and in use resources
  • kubectl top pods gives cpu and memory usage details because of the metrics-server addon. to get information related to the containers in the pod as well, we can use kubectl top pods --containers
  • similarly, we can use kubectl top nodes
  • prometheus is a better solution than metrics-server for real world use cases
  • qos - quality of service determines the priority - guaranteed > burstable > best effort
  • guaranteed - resource limit = resource request. note: remember that if only limits are defined, request = limit
  • burstable - at least one container has limit / request defined, unequal limits and requests, etc
  • best effort - no resources are defined at all
  • we can view the qos assigned by kubernetes using kubectl describe pod pod_name | grep QoS
  • additional concept - priority classes are useful for e.g. when two pods have the same qosClass. we can run k get priorityClasses and then assign one of the values using priorityClassName under spec of the pod

Limit Ranges

  • limit ranges help us specify the following at a namespace level (a yaml example) -
  • default - default resources limit
  • defaultRequest - the default resources request
  • max and min the maximum and minimum permitted values for the requests and limits
  • maxLimitRequestRatio - the maximum limit to request ratio (limit should ideally be higher than request?)

Resource Quotas

  • limits the resources that can be consumed by a namespace. so, if we have multiple namespaces to support environments like dev and prod in our clusters, we can distribute resources equally so that there is no starvation for any of the environments. a yaml example
  • using resource quotas, we can limit compute (e.g. cpu and memory requests and limits), storage (e.g. persistent volume claims) and object count (e.g. number of pods, number of node ports, etc.)

Daemon Sets

  • ensures one pod runs on each node
  • e.g. logging and monitoring pods which need to be run on every node can be created using daemon sets
  • even kube-proxy is run this way. to verify, use - kubectl get daemonsets --all-namespaces
  • it used node name but in newer versions node affinity underneath

Init Containers

  • defined inside a pod
  • before the long-running containers start, we might want to install binaries, etc
  • init containers are run to completion one at a time sequentially before the normal containers start running
  • their syntax in yaml is the same as normal containers

Static Pods

  • a pod created by kubelet itself on the node without involving the api server / etcd is called a static pod
  • the kubelet continuously monitors a directory for changes
  • so, when we can create a file in it, it gets picked up by the kubelet
  • if we edit the file / remove the file, the kubelet automatically changes / terminates the pod accordingly
  • this does not work for deployments etc. since they require controllers
  • if the node is a part of a cluster, it will notify the api server about the static pod. so, kubectl get pods will show the pod, since the etcd cluster etc. know about these pods. however, unlike a normal pod, the only way to modify this pod is to modify the file
  • use case - since static pods do not depend on control plane components like controllers, scheduler, etc., they are used to deploy the control plane components themselves
  • unless configured otherwise, the directory is /etc/kubernetes/manifests/
  • we can verify this in minikube after running minikube ssh by running sudo ls /etc/kubernetes/manifests that it has files for etcd, scheduler, api server and controller manager
  • static pods will be suffixed by -nodename - kubectl get pods --all-namespaces | grep minikube
  • if we run kubectl get pod pod_name --output=yaml, we can confirm that the owner is a node by going to ownerReferences.kind which should have the value Node
  • to get the static pod path, use - cat /var/lib/kubelet/config.yaml | grep staticPodPath

Persistent Volumes

  • persistence of state should be decoupled from pods since they can be added / removed easily
  • nfs is the way to go for disk storage in cloud. here, aws ebs has been shown
  • note: ebs volumes should only be spun up in azs where worker nodes exist, since ebs is scoped to an az
  • the spec.capacity.storage in the persistent volume defn. should be <= the capacity of ebs
  • access modes can be ReadWriteOnce, ReadOnlyMany, ReadWriteMany
  • we can run kubectl get storageclasses to get the available storage classes
  • e.g. if we were using kops with aws, it would automatically add the storage class of gp2 for us
  • default storage class admission controller observe requests for persistent volume claims and when a claim does not specify the storage class, it gets assigned the default storage class. when we run kubectl get storageclasses, we see that gp2 is marked as default
  • yaml example of persistent volume
  • persistent volumes are used through persistent volume claims. the idea is that admins create a set of persistent volumes, and developers use them via persistent volume claims
  • there is a one to one mapping i.e. one persistent volume can only be used by one persistent volume claim
  • spec.storageClassName and spec.accessModes should have the same value as that of persistent volume while the value of spec.resources.requests.storage should be <= the value of spec.capacity.storage so that the persistent volume claim can get a segment of the persistent volume
  • because of this, if the persistent volume has more storage than what the persistent volume claim asks for, the claim gets the extra storage as well
  • if no matching persistent volume is found, the persistent volume claim remains unbound indefinitely
  • yaml example of persistent volume claim
  • usage -
    1
    2
    3
    4
    5
    6
    7
    8
    9
    10
    11
    
    spec:
      containers:
        - # ...
          volumeMounts:
          - name: jenkins-home
            mountPath: /var/jenkins_home
      
        volumes:
        - name: jenkins-home
          persistentVolumeClaim:
            claimName: jenkins-storage
    
  • the status of a persistent volume can be -
    • Available when no persistent volume claim is bound to it
    • Bound when a persistent volume claim is bound to it
    • Released when the persistent volume claim is deleted
  • the default reclaim policy of a persistent volume is Retain. first, the pod / deployment is deleted, and then the persistent volume claim is deleted. now, the persistent volume has status of released. but it is not available to be bound because it already has existing data from previous pods which need to be deleted first
  • so, we delete the persistent volume manually, try to clean up / delete the aws ebs manually and then can create new persistent volumes for the persistent volume claims
  • till now, we used the manual method of provisioning volumes, i.e. static persistent volumes
  • the dynamic method requires lesser intervention
  • however, in case of a conflict, kubernetes will choose the static one
  • the persistent volume is created automatically in case of dynamic persistent volumes
  • when we delete the deployment and then the persistent volume claim now, the persistent volume as well as the actual nfs ebs volume is deleted automatically. this is because when using dynamic persistent volumes, the reclaim policy of the persistent volume is Delete
  • yaml example for persistent volume claim for dynamic persistent volume
  • the storage classes have a field called volume binding mode. this can be set to WaitForFirstConsumer i.e. persistent volume will not be bound to the persistent volume claim till there is a pod for the persistent volume claim. the other value that the binding mode can take is Immediate

Commands and Arguments

  • difference between command and entrypoint in docker is described here
  • for e.g. in the pod definition -
    • spec.containers[*].command is used for replacing ENTRYPOINT of docker file
    • spec.containers[*].args is used for replacing CMD of docker file

Security Context

  • when we run a container, we can specify the id of the user used to run the container, capabilities, etc
  • we can specify this at the pod level so that all the containers inherit it or at the container level as well
  • note: capabilities can only be defined at the container level
  • syntax -
    1
    2
    3
    4
    
    securityContext:
      runAsUser: 1000
      capabilities:
        add: ["MAC_ADMIN"]
    
  • this is more of a docker functionality which can be specified via kubernetes. basically, since containers run using namespaces, and we do not want root users inside namespaces to perform any critical operations on the host itself. so, docker adds only some capabilities, which we extend using the add capabilities mechanism
  • to verify, we can use kubectl exec -- whoami. it should display the username or the user id

Patching Nodes

  • if we remove a node suddenly, the pods scheduled on it are lost
  • if it was a part of a replica set, it would be rescheduled, but not if it was a normal pod
  • to stop any further scheduling on the current node, run kubectl cordon node_name
  • to stop any further scheduling and also evict existing pods, use kubectl drain node_name
  • if the node included a pod not spun as a part of a controller, we have to add --force. this is because that pod would be lost forever
  • so, i think drain already does what cordon does
  • pods part of replica sets etc. will be rescheduled on other nodes, because that is the job of controllers
  • after running the drain, we can start the upgrade
  • to enable scheduling pods on the node again, run kubectl uncordon node_name
  • my understanding - suppose a node goes down. so do the pods running on it. the time a controller waits to reconsider rescheduling the pod on another node is defined via podEvictionTimeout. this is why draining nodes is important, we don’t rely on the timeout, and instead, rescheduling of pods happens gracefully
  • to verify, use kubectl describe node <<node-name>> | grep Unschedulable

Network Policy

  • ingress - incoming traffic, egress - outgoing traffic
  • the response is not considered when classifying traffic as ingress and egress (like security groups in aws)
  • kubernetes has an all allow rule by default
  • we use labels to link a pod to a network policy
  • whether network policies are supported also depends on the networking solution our cluster depends on
  • if we don’t specify the podSelector, traffic from all pods are allowed, same for namespaceSelector
  • for traffic from outside the cluster, we can also use the ipBlock section
  • a yaml example for a database
  • notice how there are multiple rules. so, it works like ipBlock or (namespaceSelector and podSelector)
  • if you want to apply it to all pods, use podSelector: {}
  • my understanding - if we want to restrict traffic between namespaces, e.g. ns-1 can only make requests to ns-2, we can use network policies. so, i created one with egress rule using key kubernetes.io/metadata.name under namespaceSelector.matchLabels. however, when i tried making requests to the cluster ip service in ns-2 from ns-1, the request was failing. i guess this was happening because kube dns is needed for the dns name resolution. so, i added this policy as well to get it to work

Interface

  • for implementations to work with kubernetes, they should be compatible with the interface
  • this allows for kubernetes to be extended with multiple implementations seamlessly
  • e.g. cri for container runtime interface, cni for network, csi for storage, etc
  • e.g. csi lays a set of rpc calls that the cri will make. the csi implementations should implement these rpcs

Json Path

  • using jsonpath, we get a subset of json data
  • jsonpath always returns the results in the form of an array
  • the root element can be represented as $
  • we can use [] for indexing arrays
  • we can have conditions using ?(), e.g. to get elements greater than 40 in an array, use $[?(@ > 40)]
  • we can use in and nin to represent in and not in, e.g. $[?(@ nin [1, 2])]
  • use * to retrieve all elements in an object ($.*) or in an array ($[*])
  • a convoluted e.g. - we want to extract the laureate whose first and last name we know -
    1
    2
    3
    4
    5
    6
    7
    8
    9
    10
    11
    12
    13
    14
    15
    16
    17
    
    {
      "prizes": [
        {
          "category": "physics",
          "laureates": [
            { "firstname": "Arthur", "surname": "Ashkin" }
          ]
        },
        {
          "category": "chemistry",
          "laureates": [
            { "firstname": "Frances H.", "surname": "Arnold" },
            { "firstname": "George P.", "surname": "Smith" }
          ]
        }
      ]
    }
    

    $.prizes[*].laureates[?(@.firstname == 'Malala' && @.surname == 'Yousafzai')]

  • to retrieve multiple elements in a list, we can use $[0,3]
  • we can also use the syntax $[start:end:step] e.g. $[1:5:2]
  • we can use negative indices to retrieve the last elements, e.g. $[-3:] for the last three elements
  • element present at index end is not included
  • step defaults to 1, start and end default to 0
  • while $[positive] works, $[negative] fails in some cases, so specify an end when using negative indices

Usage in Kubectl

  • we mention the json path using --output=jsonpath="{.items}" i.e. enclose within {} and remove $
  • we can get merge multiple queries - jsonpath="{.name}{.cpu}"
  • we can print the results in a tabular format. e.g. to get the node name and its cpu usage in a table, use - kubectl get nodes --output=custom-columns="NODE:.metadata.name,CPU:.status.capacity.cpu
  • we can also specify a sort order using --sort-by, e.g. - kubectl get pods --all-namespaces --output=custom-columns="NAME:.metadata.name" --sort-by=".metadata.name"
  • to reverse the order of sort by, k get ... --sort-by=... | tac

ArgoCD

  • advertises gitops approach
    • git repository is the source of truth
    • an agent ensures and keeps the cluster and repository in sync
  • first part - testing and building of application code, creating a docker image out of it and pushing this image to docker hub
  • second part - somehow updating the image version in kubernetes manifests to point to this new version e.g. jenkins can do this
  • third part - running kubectl apply on the new manifests
  • some challenges with the flow above -
    • kubectl client, which lives outside the cluster needs to have credentials for kubernetes cluster - this is a security challenge
    • kubectl cannot monitor things like health of applications
  • now, first and second part stay the same, the third part is swapped with argocd
  • argocd is a part of the cluster itself
  • pull workflow - the argocd agent pulls the manifests and applies them to the cluster
  • argocd keeps both git repository and cluster in sync - thus avoiding drifts. this way, the git repository stays as the single source of truth
  • since this is all git, argocd allows easy rollback i.e. have argocd point to the earlier git revision
  • one common configuration - do the sync automatically in non production environments, have someone apply the changes in production environments
  • both challenges are solved -
    • permissions around changes is now managed via git - approvals for changes can be done by selective people only, and the main workhorse argocd stays inside the cluster
    • does monitoring of resources
  • using crd (custom resource definitions) it extends kubernetes functionality, so that we can use kubernetes like manifests to provide argocd configuration
  • we basically create application like below - line 8-11 is for the repository configuration, line 13 is for kubernetes cluster. we just need to run kubectl apply on the file below -
    1
    2
    3
    4
    5
    6
    7
    8
    9
    10
    11
    12
    13
    14
    15
    16
    17
    
    apiVersion: argoproj.io/v1alpha1
    kind: Application
    metadata:
      name: demo
      namespace: argocd
    spec:
      project: default
      source:
        repoURL: https://gitlab.com/shameekagarwal/argocd-demo.git
        targetRevision: HEAD
        path: k8s
      destination:
        server: https://kubernetes.default.svc
      syncPolicy:
        automated:
          selfHeal: true
          prune: true
    
  • note - the namespace defines where this argocd application should go, not where the resources should go
  • kubernetes.default.svc is the endpoint to access the api server from inside the cluster
  • head - default branch latest commit
  • for private git repositories, we need to provide the private ssh key on the application manifest above
  • self heal - false by default. if someone makes manual changes to the kubernetes cluster, using for e.g. kubectl commands, argocd will not try to abandon those changes and sync them with the git repository by default. use true to keep the git repository as the single source of truth
  • prune - delete corresponding resource if manifest is deleted. false by default to prevent accidental deletions
  • alternatives of argocd - flux, jenkins x
  • deploying argocd to our minikube cluster -
    1
    2
    
    kubectl create namespace argocd
    kubectl apply -n argocd -f https://raw.githubusercontent.com/argoproj/argo-cd/stable/manifests/install.yaml 
    
  • to access the argocd api server on our local, we first need to expose it outside the cluster. then, methods to obtain password, ip and port are described below (username is admin)
    1
    2
    3
    4
    5
    
    kubectl patch svc argocd-server -n argocd -p '{"spec": {"type": "LoadBalancer"}}'
    minikube ip # ip
    kubectl --namespace argocd get services | grep argocd-server # port
    kubectl --namespace argocd get secret argocd-initial-admin-secret \
      --output=jsonpath="{.data.password}" | base64 --decode # password
    
  • by default, argocd will poll the git repository at a fixed interval like 5 minutes. to get rid of this delay, we can use git webhooks i.e. git will trigger a notification to argocd on change
  • in the ui, we can click on refresh so that instead of the default interval, argocd polls the repository immediately
  • we can have argocd deployed to one cluster and have it monitor multiple kubernetes clusters
    • imagine due to a disaster, one of the kubernetes clusters came down (not the one where argocd is running 🤣). lets assume no backups were needed (stateless workloads). if we create a new cluster and have argocd point to this new cluster instead of the old closed one, argocd will automatically deploy all outstanding workloads from the git repositories and bring it up to speed
  • app project - group multiple such applications and manage them easily
  • argocd health status - aggregates status from different resources that have been deployed
  • application sets - factory for applications i.e. unlike application where kubernetes cluster and git repository is one to one, in application sets, kubernetes cluster and git repository is many to many
  • in application sets, templates represent an application and generators are used to populate these applications using parameters. e.g. of list generator -
    1
    2
    3
    4
    5
    6
    7
    8
    9
    10
    11
    12
    13
    14
    15
    16
    17
    18
    19
    20
    21
    22
    23
    24
    25
    26
    27
    28
    
    apiVersion: argoproj.io/v1alpha1
    kind: ApplicationSet
    metadata:
      name: demo
      namespace: argocd
    spec:
      generators:
        - list:
            elements:
              - cluster: asia
                url: https://kubernetes.default.svc
              - cluster: americas
                url: https://kubernetes.default.svc
      template:
        metadata:
          name: demo
        spec:
          project: default
          source:
            repoURL: https://gitlab.com/shameekagarwal/argocd-demo.git
            targetRevision: HEAD
            path: k8s
          destination:
            server: \{\{ url \}\}
          syncPolicy:
            automated:
              selfHeal: true
              prune: true
    
  • there are a lot of more generators that we can use in an application set. some common generators have been described below
  • git directory generator - useful for monorepo pattern i.e. one single git project has multiple applications, e.g. one folder say app has a kustomize based manifests, another folder say monitoring has helm based manifests and so on. in the application set, we only have to prodvide the git repository link. argocd will scan the git repository for different such kubernetes manifests and generate 2 applications automatically
  • cluster generator - we register clusters inside argocd. our application set can then automatically generate applications for all the clusters that argocd knows off. we can use some methods to match certain clusters only
  • matrix generator - combine different generators, e.g. cluster generator and git directory generator. this way, if git directory generator finds 4 kustomize applications and cluster generator finds 3 clusters, we end up deploying 4 * 3 = 12 argocd applications

Kustomize

  • helps use kubernetes templating using yaml easily
  • inbuilt inside newer versions of kubectl
  • base folder - manifests file + kustomization.yaml
  • overlays folder - environment specific configuration
  • each environment inside overlay gets a different directory
  • to view the effective generated manifest, use kubectl kustomize overlays/env
  • to apply, use kubectl apply --kustomize overlays/env
  • kustomization.yaml can have things like common labels, annotations, common prefixes and suffixes, etc
  • resources - what files to process
  • commonLabels - add this to all labels and selectors - recall how this can easily remove human errors while for e.g. linking services to pods etc
  • namePrefix and nameSuffix - add for e.g. application specific prefixes and suffixes to all manifests
  • therefore, base/kustomization.yaml -
    1
    2
    3
    4
    5
    6
    7
    8
    9
    
    resources:
    \- deployment.yaml
    \- service.yaml
      
    commonLabels:
      name: demo
    
    namePrefix: kustomize-
    nameSuffix: -v1
    
  • patchesStrategicMerge - which files to use as patches. some properties in the patch files like name, kind, etc can be used to match its corresponding raw file in base, and its properties like replicas can override / add propeties to the raw file
  • therefore, overlays/dev/replicas.yaml -
    1
    2
    3
    4
    5
    6
    
    apiVersion: apps/v1
    kind: Deployment
    metadata:
      name: webapp
    spec:
      replicas: 2
    
  • configMapGenerator - we just specify the name and the file to use here. the name that we specify here should match what we use in deployment.yaml when populating volumes / environment variables using config maps
  • therefore, overlays/dev/kustomization.yaml -
    1
    2
    3
    4
    5
    6
    7
    8
    9
    10
    11
    
    resources:
      \- ../../base
    
    patchesStrategicMerge:
      \- "replicas.yaml"
    
    namespace: dev
    
    configMapGenerator:
      \- name: some-config
        env: config.properties
    
  • overlays/dev/config.properties -
    1
    
    CUSTOM_HEADER=Inside Prod
    
This post is licensed under CC BY 4.0 by the author.
Contents

Hadoop

Spring Reactive