This section will illustrate how to create an AWS Elastic Kubernetes Service. Assuming you have already installed the aws CLI and the eksctl CLI you can proceed by creating the kubernetes cluster with following characteristics (that you may change):

  • Cluster name: stackgres
  • Kubernetes version: 1.13
  • Zone: us-west-2
  • Machine type: m5.large
  • Number of nodes: 3
  • Disk size 20GB
eksctl create cluster --name stackgres \
  --region us-west-2 \
  --node-type m5.large \
  --node-volume-size 20 \
  --nodes 3 \
  --version 1.13
[]  eksctl version 0.13.0
[]  using region us-west-2
[]  setting availability zones to [us-west-2a us-west-2c us-west-2b]
[]  subnets for us-west-2a - public: private:
[]  subnets for us-west-2c - public: private:
[]  subnets for us-west-2b - public: private:
[]  nodegroup "ng-308f6134" will use "ami-09bcf0b1f5b446c5d" [AmazonLinux2/1.13]
[]  using Kubernetes version 1.13
[]  creating EKS cluster "stackgres" in "us-west-2" region with un-managed nodes
[]  will create 2 separate CloudFormation stacks for cluster itself and the initial nodegroup
[]  if you encounter any issues, check CloudFormation console or try 'eksctl utils describe-stacks --region=us-west-2 --cluster=stackgres'
[]  CloudWatch logging will not be enabled for cluster "stackgres" in "us-west-2"
[]  you can enable it with 'eksctl utils update-cluster-logging --region=us-west-2 --cluster=stackgres'
[]  Kubernetes API endpoint access will use default of {publicAccess=true, privateAccess=false} for cluster "stackgres" in "us-west-2"
[]  2 sequential tasks: { create cluster control plane "stackgres", create nodegroup "ng-308f6134" }
[]  building cluster stack "eksctl-stackgres-cluster"
[]  deploying stack "eksctl-stackgres-cluster"
[]  building nodegroup stack "eksctl-stackgres-nodegroup-ng-308f6134"
[]  --nodes-min=3 was set automatically for nodegroup ng-308f6134
[]  --nodes-max=3 was set automatically for nodegroup ng-308f6134
[]  deploying stack "eksctl-stackgres-nodegroup-ng-308f6134"
[]  all EKS cluster resources for "stackgres" have been created
[]  saved kubeconfig as "/home/matteom/.kube/config-aws"
[]  adding identity "arn:aws:iam::661392101474:role/eksctl-stackgres-nodegroup-ng-NodeInstanceRole-C8R84QGP5UYX" to auth ConfigMap
[]  nodegroup "ng-308f6134" has 1 node(s)
[]  node "ip-192-168-66-45.us-west-2.compute.internal" is not ready
[]  waiting for at least 3 node(s) to become ready in "ng-308f6134"
[]  nodegroup "ng-308f6134" has 3 node(s)
[]  node "ip-192-168-2-185.us-west-2.compute.internal" is ready
[]  node "ip-192-168-58-166.us-west-2.compute.internal" is ready
[]  node "ip-192-168-66-45.us-west-2.compute.internal" is ready
[]  kubectl command should work with "/home/matteom/.kube/config-aws", try 'kubectl --kubeconfig=/home/matteom/.kube/config-aws get nodes'
[]  EKS cluster "stackgres" in "us-west-2" region is ready

To cleanup the kubernetes cluster you may issue following command:

eksctl delete cluster --name stackgres \
  --region us-west-2 \

You may also want to cleanup EBS used by persistence volumes that may have been created:

aws ec2 describe-volumes --region us-west-2 --filters Name=tag-key,Values=kubernetes.io/cluster/stackgres \
  | jq -r '.Volumes[].VolumeId' | xargs -r -n 1 -I % aws ec2 delete-volume --region us-west-2 --volume-id %