Helm is an open source packaging tool that helps install applications and services on Kubernetes. It uses a packaging format called charts. A Helm chart is a package containing all resource definitions necessary to create an instance of a Kubernetes application, tool, or service in a Kubernetes cluster.

Prerequisites

You must have a Kubernetes cluster that has Helm configured. If you have not installed the Helm client (helm), see Install Helm.

The YugabyteDB Helm chart has been tested with the following software versions:

  • Kubernetes 1.20 or later with nodes such that a total of 12 CPU cores and 18 GB RAM can be allocated to YugabyteDB. This can be three nodes with 4 CPU core and 6 GB RAM allocated to YugabyteDB.
  • Helm 3.4 or later.
  • YugabyteDB Docker image (yugabytedb/yugabyte) 2.1.0 or later
  • For optimal performance, ensure you have set the appropriate system limits using ulimit on each node in your Kubernetes cluster.

Confirm that helm and kubectl are configured correctly, as follows:

helm version
version.BuildInfo{Version:"v3.2.1", GitCommit:"fe51cd1e31e6a202cba7dead9552a6d418ded79a", GitTreeState:"clean", GoVersion:"go1.13.10"}
kubectl version

Create cluster

Creating a cluster includes adding a repository for charts and updating this repository, checking the version, and installing YugabyteDB.

Add charts repository

To add the YugabyteDB charts repository, run the following command:

helm repo add yugabytedb https://charts.yugabyte.com

Fetch updates from the repository

Make sure that you have the latest updates to the repository by running the following command:

helm repo update

Validate the chart version

To check the chart version, run the following command:

helm search repo yugabytedb/yugabyte --version 2.23.1

Expect output similar to the following:

NAME                 CHART VERSION  APP VERSION   DESCRIPTION
yugabytedb/yugabyte  2.23.1          2.23.1.0-b220  YugabyteDB is the high-performance distributed ...

Install YugabyteDB

Install YugabyteDB in the Kubernetes cluster using the commands described in the following sections.

On multi-node Kubernetes

Create a namespace and then install YugabyteDB, as follows:

kubectl create namespace yb-demo

helm install yb-demo yugabytedb/yugabyte --version 2.23.1 --namespace yb-demo --wait

On Minikube

If you are running in a resource-constrained environment or a local environment, such as Minikube, you have to change the default resource requirements.

Create a yb-demo namespace, as follows:

kubectl create namespace yb-demo

helm install yb-demo yugabytedb/yugabyte \
--version 2.23.1 \
--set resource.master.requests.cpu=0.5,resource.master.requests.memory=0.5Gi,\
resource.tserver.requests.cpu=0.5,resource.tserver.requests.memory=0.5Gi --namespace yb-demo

Note that in Minikube, the LoadBalancers for yb-master-ui and yb-tserver-service will remain in pending state as load balancers are not available in a Minikube environment. If you would like to disable these services, pass the enableLoadBalancer=False flag, as follows:

helm install yb-demo yugabytedb/yugabyte \
--version 2.23.1 \
--set resource.master.requests.cpu=0.5,resource.master.requests.memory=0.5Gi,\
resource.tserver.requests.cpu=0.5,resource.tserver.requests.memory=0.5Gi,\
enableLoadBalancer=False --namespace yb-demo

In some environments, such as macOS, Minikube may run inside a virtual machine. Make sure to configure the VM with at least 4 CPUs and 5 GB memory so the cluster has room to start up. The following is an example command:

minikube start --cpus 4 --memory 5120

Check the cluster status

You can check the status of the cluster using the following commands:

helm status yb-demo -n yb-demo

Expect output similar to the following:

NAME: yb-demo
LAST DEPLOYED: Thu Feb 13 13:29:13 2020
NAMESPACE: yb-demo
STATUS: deployed
REVISION: 1
TEST SUITE: None
NOTES:
1. Get YugabyteDB Pods by running this command:
  kubectl --namespace yb-demo get pods

2. Get list of YugabyteDB services that are running:
  kubectl --namespace yb-demo get services

3. Get information about the load balancer services:
  kubectl get svc --namespace yb-demo

4. Connect to one of the tablet server:
  kubectl exec --namespace yb-demo -it yb-tserver-0 -- bash

5. Run YSQL shell from inside of a tablet server:
  kubectl exec --namespace yb-demo -it yb-tserver-0 -- ysqlsh  -h yb-tserver-0.yb-tservers.yb-demo

6. Cleanup YugabyteDB Pods
  helm delete yb-demo --purge
  NOTE: You need to manually delete the persistent volume
  kubectl delete pvc --namespace yb-demo -l app=yb-master
  kubectl delete pvc --namespace yb-demo -l app=yb-tserver

Check the pods, as follows:

kubectl get pods --namespace yb-demo

Expect output similar to the following:

NAME           READY     STATUS    RESTARTS   AGE
yb-master-0    2/2       Running   0          4m
yb-master-1    2/2       Running   0          4m
yb-master-2    2/2       Running   0          4m
yb-tserver-0   2/2       Running   0          4m
yb-tserver-1   2/2       Running   0          4m
yb-tserver-2   2/2       Running   0          4m

Check the services, as follows:

kubectl get services --namespace yb-demo

Expect output similar to the following:

NAME                 TYPE           CLUSTER-IP      EXTERNAL-IP    PORT(S)                                        AGE
yb-master-ui         LoadBalancer   10.109.39.242   35.225.153.213 7000:31920/TCP                                 10s
yb-masters           ClusterIP      None            <none>         7100/TCP,7000/TCP                              10s
yb-tserver-service   LoadBalancer   10.98.36.163    35.225.153.214 6379:30929/TCP,9042:30975/TCP,5433:30048/TCP   10s
yb-tservers          ClusterIP      None            <none>         7100/TCP,9000/TCP,6379/TCP,9042/TCP,5433/TCP   10s

You can also check the history of the yb-demo deployment, as follows:

helm history yb-demo -n yb-demo

Expect output similar to the following:

REVISION  UPDATED                   STATUS    CHART           APP VERSION   DESCRIPTION
1         Thu Apr 13 13:29:13 2020  deployed  yugabyte-2.13.0  2.13.0.1-b2  Install complete

Connect using YugabyteDB shells

To connect and use the YSQL Shell (ysqlsh), run the following command:

kubectl exec -n yb-demo -it yb-tserver-0 -- ysqlsh -h yb-tserver-0.yb-tservers.yb-demo

To connect and use the YCQL Shell (ycqlsh), run the following command:

kubectl exec -n yb-demo -it yb-tserver-0 -- ycqlsh yb-tserver-0.yb-tservers.yb-demo

Connect using external clients

To connect an external program, get the load balancer EXTERNAL-IP address of the yb-tserver-service service and connect using port 5433 for YSQL or port 9042 for YCQL, as follows:

kubectl get services --namespace yb-demo

Expect output similar to the following:

NAME                 TYPE           CLUSTER-IP      EXTERNAL-IP        PORT(S)                                        AGE
...
yb-tserver-service   LoadBalancer   10.98.36.163    35.225.153.214     6379:30929/TCP,9042:30975/TCP,5433:30048/TCP   10s
...

Configure cluster

Instead of using the default values in the Helm chart, you can modify the configuration of the YugabyteDB cluster according to your requirements.

CPU, memory, and replica count

The default values for the Helm chart are in the helm/yugabyte/values.yaml file. The following is a listing of the most important values. As noted in Prerequisites, the defaults are set for a 3-node Kubernetes cluster, each node with 4 CPU cores and 6 GB RAM allocated to YugabyteDB.

storage:
  master:
    count: 2
    size: 10Gi
    storageClass: standard
  tserver:
    count: 2
    size: 10Gi
    storageClass: standard

resource:
  master:
    requests:
      cpu: 2
      memory: 2Gi
    limits:
      cpu: 2
      memory: 2Gi
  tserver:
    requests:
      cpu: 2
      memory: 4Gi
    limits:
      cpu: 2
      memory: 4Gi

replicas:
  master: 3
  tserver: 3

partition:
  master: 3
  tserver: 3

If you want to change the defaults, you can use the following command. You can even do helm install instead of helm upgrade when you are installing on a Kubernetes cluster with configuration different than the defaults:

helm upgrade --set resource.tserver.requests.cpu=8,resource.tserver.requests.memory=15Gi yb-demo ./yugabyte

Replica count can be changed using the following command. Note that only the YB-TServers need to be scaled in a replication factor 3 cluster which keeps the masters count at 3:

helm upgrade --set replicas.tserver=5 yb-demo ./yugabyte

Readiness probes

Readiness probes provide readiness checks for your Kubernetes deployment. Probes are compatible with both direct Helm deployments and YugabyteDB Anywhere-managed deployments, and work with TLS enabled or restricted authorization environments. Use the probes to ensure pods are ready before being marked as available. The probes verify connectivity using ysqlsh for YSQL and ycqlsh for YCQL.

The following probes are available:

  • YSQL Readiness. Uses ysqlsh to verify connectivity via local socket for credentialed setups.

  • YCQL Readiness. Uses ycqlsh to validate connectivity.

  • Master Readiness. Uses httpGet to probe master.

  • Custom Readiness. Supports custom readiness probe parameters, such as delays, timeouts, and thresholds.

  • Startup probes to delay enforcing of readiness probes.

Readiness probes are disabled by default for compatibility.

Enable probes via values.yaml or using Kubernetes overrides.

master:
  readinessProbe:
    enabled: true

tserver:
  readinessProbe:
    enabled: true

The following is example of custom readiness parameters:

tserver:
  customReadinessProbe:
    initialDelaySeconds: 30
    periodSeconds: 20
    timeoutSeconds: 10

Independent LoadBalancers

By default, the YugabyteDB Helm chart exposes the client API endpoints and master UI endpoint using two load balancers. If you want to expose the client APIs using independent LoadBalancers, you can execute the following command:

helm install yb-demo yugabytedb/yugabyte -f https://raw.githubusercontent.com/yugabyte/charts/master/stable/yugabyte/expose-all.yaml --version 2.23.1 --namespace yb-demo --wait

You can also bring up an internal load balancer (for either YB-Master or YB-TServer services), if required. To do so, you specify the annotation required for your cloud provider. See Amazon EKS and Google Kubernetes Engine for examples.

Reserved LoadBalancer IP Addresses

If you intend to use a preallocated (reserved) IP for the exposed YB-Master and YB-TServer services, you need to specify the load balancer IP. If you do not set this IP, a so-called ephemeral, semi-random IP will be allocated.

The following is an example of the values-overrides.yaml file that allows you to override IP values in the Helm charts:

serviceEndpoints:
  - name: "yb-master-ui"
    app: "yb-master"
    loadBalancerIP: "11.11.11.11"
    type: "LoadBalancer"
    ports:
      http-ui: "7000"
  - name: "yb-tserver-service"
    app: "yb-tserver"
    loadBalancerIP: "22.22.22.22"
    type: "LoadBalancer"
    ports:
      tcp-yql-port: "9042"
      tcp-ysql-port: "5433"

You apply the override by executing the following Helm command:

helm install yb-demo ./yugabyte -f values-overrides.yaml

Assuming that you already reserved the IP addresses (11.11.11.11 and 22.22.22.22), yb-master-ui and yb-tserver-service will use the predetermined addresses.

Note that setting the load balancer IP can result in behavior that might not be entirely consistent across cloud providers.

Storage class

If you want to use a storage class other than the standard class for your deployment, provision the storage class and then pass in the name of the class while running the helm install command, as follows:

helm install yugabyte --version 2.23.1 --namespace yb-demo --name yb-demo --set storage.master.storageClass=<desired storage class>,storage.tserver.storageClass=<desired storage class> --wait

Configure YB-Master and YB-TServer pods

Flags on the YB-Master and YB-TServer pods can be specified via the command line or by overriding the values.yaml file in the charts repository. The following example shows how to set the three geo-distribution-related flags placement_cloud, placement_region, and placement_zone on a Minikube cluster:

helm install yb-demo yugabytedb/yugabyte \
--version 2.23.1 \
--set resource.master.requests.cpu=0.5,resource.master.requests.memory=0.5Gi,\
resource.tserver.requests.cpu=0.5,resource.tserver.requests.memory=0.5Gi,\
gflags.master.placement_cloud=myk8s-cloud,gflags.master.placement_region=myk8s-region,gflags.master.placement_zone=myk8s-zone,\
gflags.tserver.placement_cloud=myk8s-cloud,gflags.tserver.placement_region=myk8s-region,gflags.tserver.placement_zone=myk8s-zone\
 --namespace yb-demo

Upgrade the software version of YugabyteDB

You can upgrade the software on the YugabyteDB cluster with the following command. By default, this performs a rolling update of the pods.

helm repo update
helm upgrade yb-demo yugabytedb/yugabyte --version 2.23.1 --wait -n yb-demo

Update the configuration of YugabyteDB pods

You can update most settings in the helm chart by running a helm upgrade with the new values. By default, this performs a rolling update of the pods.

helm upgrade yb-demo yugabytedb/yugabyte --set resource.tserver.requests.cpu=4 --wait -n yb-demo

Delete cluster

To delete the cluster, you need to purge the Helm chart, and then delete the PVCs, as follows:

helm uninstall yb-demo -n yb-demo
kubectl delete pvc --namespace yb-demo --all