AWS Air-gapped FIPS: Create a Managed Cluster Using the DKP CLI
In previous steps, you created a new Management cluster, which is self-managed. When you use these steps to create new Managed clusters, they will become Attached clusters under your Management Cluster.
When creating Managed clusters, you do not need to create and move CAPI objects, or install the Kommander component. Those tasks are only done on Management clusters!
Create New Managed Air-gapped FIPS Cluster with the CLI
The below instructions tell you how to create a cluster and have it automatically attach to the workspace you set above.
If you do not set a workspace, it will be created in the default
workspace, and you need to take additional steps to attach to a workspace later. For instructions on how to do this, see Attach a Kubernetes Cluster.
Your new, managed cluster needs to be part of either a new Workspace or an existing Workspace under a management cluster. Create or locate a workspace name to get started with your managed cluster.
Make New Cluster Part of a Workspace
If you have an existing Workspace name, run this command to find the name:
⚠️ NOTE: If you need to create a new Workspace, follow the instructions to Create a New Workspace.CODEkubectl get workspace -A
When you have the Workspace name, set the
WORKSPACE_NAMESPACE
environment variable:CODEexport WORKSPACE_NAMESPACE=<workspace_namespace>
Name Your Cluster
Follow these steps:
Give your cluster a unique name suitable for your environment.
In AWS it is critical that the name is unique, as no two clusters in the same AWS account can have the same name.
Set the environment variable:
CODEexport MANAGED_CLUSTER_NAME=<aws-additional>
The cluster name may only contain the following characters: a-z
, 0-9
, .
, and -
. Cluster creation will fail if the name has capital letters. See Kubernetes for more naming information.
3. Execute this command to create your Kubernetes cluster using any relevant flags but note that the --self-managed
is absent so that this cluster can be attached to your previously created Management Cluster:
dkp create cluster aws
--cluster-name=${MANAGED_CLUSTER_NAME} \
--additional-tags=owner=$(whoami) \
--namespace ${WORKSPACE_NAMESPACE}
--with-aws-bootstrap-credentials=true \
--vpc-id=${AWS_VPC_ID} \
--ami=${AWS_AMI_ID} \
--subnet-ids=${AWS_SUBNET_IDS} \
--internal-load-balancer=true \
--additional-security-group-ids=${AWS_ADDITIONAL_SECURITY_GROUPS} \
--registry-mirror-url=${REGISTRY_URL} \
--registry-mirror-cacert=${REGISTRY_CA} \
--registry-mirror-username=${REGISTRY_USERNAME} \
--registry-mirror-password=${REGISTRY_PASSWORD} \
--kubeconfig=<management-cluster-kubeconfig-path> \
If your environment uses HTTP/HTTPS proxies, you must include the flags --http-proxy
, --https-proxy
, and --no-proxy
and their related values in this command for it to be successful. More information is available in Configuring an HTTP/HTTPS Proxy.
Retrieve the kubeconfig and Explore New AWS Cluster
Follow these steps:
Fetch the kubeconfig file with the command:
CODEdkp get kubeconfig --cluster-name ${MANAGED_CLUSTER_NAME} --kubeconfig <management-cluster-kubeconfig-path> -n ${WORKSPACE_NAMESPACE} > ${MANAGED_CLUSTER_NAME}.conf
List the nodes with the following command:
CODEkubectl --kubeconfig=${MANAGED_CLUSTER_NAME}.conf get nodes
List the pods with the following command:
NOTE: Wait for the Status to move to Ready while calico-node
pods are being deployed.
kubectl --kubeconfig=${MANAGED_CLUSTER_NAME}.conf get pods -A
Manually Attach a DKP CLI Cluster to the Management Cluster
These steps are only applicable if you do not set a WORKSPACE_NAMESPACE when creating a cluster. If you already set a WORKSPACE_NAMESPACE, then you do not need to perform these steps since the cluster is already attached to the workspace.
Starting with DKP 2.6, when you create a Managed Cluster with the DKP CLI, it attaches automatically to the Management Cluster after a few moments.
However, if you do not set a workspace, the attached cluster will be created in the default
workspace. To ensure that the attached cluster is created in your desired workspace namespace, follow these instructions:
Confirm you have your MANAGED_CLUSTER_NAME variable set with the following command:
CODEecho ${MANAGED_CLUSTER_NAME}
Retrieve your kubeconfig from the cluster you have created without setting a workspace:
CODEdkp get kubeconfig --cluster-name ${MANAGED_CLUSTER_NAME} > ${MANAGED_CLUSTER_NAME}.conf
You can now either [attach it in the UI](link to attaching it to workspace via UI that was earlier), or attach your cluster to the workspace you want in the CLI.
NOTE: This is only necessary if you never set the workspace of your cluster upon creation.Retrieve the workspace where you want to attach the cluster:
CODEkubectl get workspaces -A
Set the WORKSPACE_NAMESPACE environment variable:
CODEexport WORKSPACE_NAMESPACE=<workspace-namespace>
You need to create a secret in the desired workspace before attaching the cluster to that workspace. Retrieve the kubeconfig secret value of your cluster:
CODEkubectl -n default get secret ${MANAGED_CLUSTER_NAME}-kubeconfig -o go-template='{{.data.value}}{{ "\n"}}'
This will return a lengthy value. Copy this entire string for a secret using the template below as a reference. Create a new attached-cluster-kubeconfig.yaml file:
CODEapiVersion: v1 kind: Secret metadata: name: <your-managed-cluster-name>-kubeconfig labels: cluster.x-k8s.io/cluster-name: <your-managed-cluster-name> type: cluster.x-k8s.io/secret data: value: <value-you-copied-from-secret-above>
Create this secret in the desired workspace:
CODEkubectl apply -f attached-cluster-kubeconfig.yaml --namespace ${WORKSPACE_NAMESPACE}
Create this
kommandercluster
object to attach the cluster to the workspace:CODEcat << EOF | kubectl apply -f - apiVersion: kommander.mesosphere.io/v1beta1 kind: KommanderCluster metadata: name: ${MANAGED_CLUSTER_NAME} namespace: ${WORKSPACE_NAMESPACE} spec: kubeconfigRef: name: ${MANAGED_CLUSTER_NAME}-kubeconfig clusterRef: capiCluster: name: ${MANAGED_CLUSTER_NAME} EOF
You can now view this cluster in your Workspace in the UI and you can confirm its status by running the below command. It may take a few minutes to reach "Joined" status:
CODEkubectl get kommanderclusters -A
If you have several Essential Clusters and want to turn one of them to a Managed Cluster to be centrally administrated by a Management Cluster, refer to Platform Expansion: Convert a DKP Essential Cluster to a DKP Enterprise Managed Cluster.
When you have existing clusters or want to create other new clusters to attach, there are many ways to attach a cluster with various requirements and restrictions. To see all the options, visit the section in the Day 2- Cluster Operations Management section for documentation on how to Attach an Existing Kubernetes Cluster.