Setting up EKS with Terraform for Altinity.Cloud

The easiest way to configure an EKS cluster

The Altinity Terraform module for EKS makes it easy to set up an EKS Kubernetes cluster for a Bring Your Own Kuberentes (BYOK) environment.


  • Terraform version >= 1.5
  • kubectl
  • Authentication with your AWS account

To authenticate with your AWS account, set the environment variables AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY, and AWS_SESSION_TOKEN. You must also ensure your AWS account has sufficient permissions for EKS and related services.

Terraform module for BYOK on EKS

We created the Terraform EKS module to make it easy to spin up an AWS EKS cluster optimized for working with Altinity.Cloud Anywhere. This configuration is tailored for the best performance of ClickHouse, following Altinity’s best practices and recommended specs for AWS:

  • Instance Types
  • Node Labels
  • EBS Controller with custom Storage Class (gp3-encrypted)
  • Cluster Autoscaler with multi-zones High Availability

See the module’s repo for detailed information about the module's architecture.

EKS cluster and node groups setup

  1. Create a new directory for your terraform project and create a file named .
  2. Copy and paste the following code into the file:
locals {
  region = "us-east-1"

module "eks_clickhouse" {
  source  = ""

  # We don't need to install the operator and the cluster, ACM will take care of this task.
  install_clickhouse_operator = false
  install_clickhouse_cluster  = false

  eks_cluster_name = "clickhouse-cluster"
  eks_region       = local.region
  eks_cidr         = ""

  eks_availability_zones = [
  eks_private_cidr = [
  eks_public_cidr = [
  eks_node_pools_config = {
    scaling_config = {
      desired_size = 2
      max_size     = 10
      min_size     = 0

    disk_size      = 20
    instance_types = ["m6i.large", "t3.large"]

  eks_tags = {
    CreatedBy = "mr-robot"

This configuration will create 6 different node groups using the combination of eks_availability_zones and instance_types.

Be sure to tweak configuration values like eks_cluster_name, region (for the availability zones), instance_types, CIDR information and whatever else you need for your requirements. If you need further customization, you can always fork the Terraform module and create your personalized version of it.

Again, remember to authenticate with your AWS account before going forward.

Applying the configuration

Open the terminal, navigate into the created directory and run these commands to initialize the Terraform project and apply it:

# initialize the terraform project
terraform init

# apply module changes
# btw, did you remember to authenticate with your AWS account? 
terraform apply

This operation will take several minutes to complete. When it completes, you’ll have a running AWS EKS cluster with high availability and other features.

Verify your EKS cluster

  1. Update your kubeconfig with the new AWS EKS cluster data using the following command:
aws eks update-kubeconfig --region us-east-1 --name clickhouse-cluster
  1. List your AWS EKS cluster nodes:
kubectl get nodes

You should see six nodes:

NAME                         STATUS   ROLES    AGE     VERSION
ip-10-0-1-174.ec2.internal   Ready    <none>   6m17s   v1.28.5-eks-5e0fdde
ip-10-0-1-35.ec2.internal    Ready    <none>   6m55s   v1.28.5-eks-5e0fdde
ip-10-0-2-181.ec2.internal   Ready    <none>   6m35s   v1.28.5-eks-5e0fdde
ip-10-0-2-63.ec2.internal    Ready    <none>   6m38s   v1.28.5-eks-5e0fdde
ip-10-0-3-128.ec2.internal   Ready    <none>   6m37s   v1.28.5-eks-5e0fdde
ip-10-0-3-164.ec2.internal   Ready    <none>   6m36s   v1.28.5-eks-5e0fdde
  1. List the kube-system pods:
kubectl get pods -n kube-system

All these pods are created by default under the kube-system namespace. You’ll see something like this:

NAME                                                        READY   STATUS    RESTARTS   AGE
aws-node-gw24w                                              2/2     Running   0          6m12s
aws-node-j9sr6                                              2/2     Running   0          6m14s
aws-node-kwdgq                                              2/2     Running   0          6m13s
aws-node-spdtd                                              2/2     Running   0          6m32s
aws-node-stnjm                                              2/2     Running   0          6m15s
aws-node-wq528                                              2/2     Running   0          5m54s
cluster-autoscaler-aws-cluster-autoscaler-686c7f4f5-bhw9n   1/1     Running   0          4m34s
coredns-86969bccb4-dpfb4                                    1/1     Running   0          10m
coredns-86969bccb4-mlpt6                                    1/1     Running   0          10m
ebs-csi-controller-68ff8856fc-ffc4s                         6/6     Running   0          4m44s
ebs-csi-controller-68ff8856fc-jqw2k                         6/6     Running   0          4m44s
ebs-csi-node-62h5k                                          3/3     Running   0          4m44s
ebs-csi-node-69924                                          3/3     Running   0          4m44s
ebs-csi-node-9snbj                                          3/3     Running   0          4m44s
ebs-csi-node-qgdtk                                          3/3     Running   0          4m44s
ebs-csi-node-r5lq9                                          3/3     Running   0          4m44s
ebs-csi-node-sjqdw                                          3/3     Running   0          4m44s
kube-proxy-2ntw7                                            1/1     Running   0          6m15s
kube-proxy-9q4rs                                            1/1     Running   0          5m54s
kube-proxy-kv76v                                            1/1     Running   0          6m32s
kube-proxy-lkmp4                                            1/1     Running   0          6m12s
kube-proxy-pzwht                                            1/1     Running   0          6m13s
kube-proxy-sfz7z                                            1/1     Running   0          6m14s

Your AWS EKS cluster is now ready, and you are able to connect it to the Altinity Cloud Manager (ACM). Remember that the given configuration is just a starting point. Before using this in production, you should review the module documentation and ensure it fits your security needs.

To connect your new Kubernetes cluster to the ACM, see the section Connecting to Altinity.Cloud Anywhere.

Deleting the configuration

When you no longer need your EKS cluster, use kubectl to delete all the ClickHouse clusters in the environment you created, then delete the environment itself. Start by deleting the ClickHouse installation (chi):

kubectl delete chi --all --all-namespaces

Once that’s done, use the following command to clean up the EKS cluster and all of its resources:

terraform destroy

When this command finishes, all of the resources associated with your EKS environment are gone.