Learn how to deploy a sample application on a DKP cluster
After you have a basic DKP cluster installed and ready to use, you might want to test operations by deploying a simple, sample application. This task is optional and is only intended to demonstrate the basic steps for deploying applications in a production environment. If you are configuring the DKP cluster for a production deployment, you can use this section to learn the deployment process. However, deploying applications on a production cluster typically involves more planning and custom configuration than covered in this example.
This tutorial shows how to deploy a simple application that connects to the
redis service. The sample application used in this tutorial is a condensed form of the Kubernetes sample guestbook application.
Before You Begin
You must have a DKP cluster running.
Before running the commands below, ensure that your
kubectl configuration references the DKP cluster on which you want to install the application. You can do this by setting the
KUBECONFIG environment variable to the appropriate kubeconfig file’s location.
Deploy a Sample Application
Deploy the Redis pods and service by running the following commands:
kubectl apply -f https://k8s.io/examples/application/guestbook/redis-leader-deployment.yaml kubectl apply -f https://k8s.io/examples/application/guestbook/redis-leader-service.yamlCODE
Deploy Redis followers. The leader deployment created above is a single pod. Adding followers (or replicas) makes it highly available to meet greater traffic demands. You must then setup the guestbook application to communicate with the Redis followers to read the data. To do this, set up another service (the
redis-follower-service.yamlbelow). Do this by running the following commands:
kubectl apply -f https://k8s.io/examples/application/guestbook/redis-follower-deployment.yaml kubectl apply -f https://k8s.io/examples/application/guestbook/redis-follower-service.yamlCODE
Deploy the web app frontend by running the following command:
kubectl apply -f https://k8s.io/examples/application/guestbook/frontend-deployment.yamlCODE
Confirm that there are three frontend replicas running:
kubectl get pods -l app=guestbook -l tier=frontendCODE
Apply the frontend Service:
kubectl apply -f https://k8s.io/examples/application/guestbook/frontend-service.yamlCODE
Configure the frontend Service to use a cloud load balancer:
cat << EOF | kubectl apply -f - apiVersion: v1 kind: Service metadata: name: frontend labels: app: guestbook tier: frontend spec: type: LoadBalancer ports: - port: 80 selector: app: guestbook tier: frontend EOFCODE
View the frontend service via the LoadBalancer by running the following command to get the IP address for the frontend Service:
kubectl get service frontendCODE
the external IP address, and load the page in your browser to view your guestbook.
The service properties provide the name of the load balancer. You can connect to the application by accessing that load balancer address in your web browser. Because this sample deployment creates a cloud load balancer, you should keep in mind that creating the load balancer can take up to a few minutes. You also might experience a slight delay before it is running properly due to DNS propagation and synchronization.
Remove the sample application by running the following commands:
kubectl delete deployment -l app=redis kubectl delete service -l app=redis kubectl delete deployment frontend kubectl delete service frontendCODE
WARNING: This step is required because the sample deployment attaches a cloud provider load balancer to the DKP cluster. Therefore, you must delete the sample application before tearing down the cluster.