Open source Kubernetes
This page documents the preview version (v2.23). Preview includes features under active development and is for development and testing only. For production, use the stable version (v2024.1). To learn more, see Versioning.
Helm is an open source packaging tool that helps install applications and services on Kubernetes. It uses a packaging format called charts
. A Helm chart is a package containing all resource definitions necessary to create an instance of a Kubernetes application, tool, or service in a Kubernetes cluster.
Prerequisites
You must have a Kubernetes cluster that has Helm configured. If you have not installed the Helm client (helm
), see Install Helm.
The YugabyteDB Helm chart has been tested with the following software versions:
- Kubernetes 1.20 or later with nodes such that a total of 12 CPU cores and 18 GB RAM can be allocated to YugabyteDB. This can be three nodes with 4 CPU core and 6 GB RAM allocated to YugabyteDB.
- Helm 3.4 or later.
- YugabyteDB Docker image (yugabytedb/yugabyte) 2.1.0 or later
- For optimal performance, ensure you have set the appropriate system limits using
ulimit
on each node in your Kubernetes cluster.
Confirm that helm
and kubectl
are configured correctly, as follows:
helm version
version.BuildInfo{Version:"v3.2.1", GitCommit:"fe51cd1e31e6a202cba7dead9552a6d418ded79a", GitTreeState:"clean", GoVersion:"go1.13.10"}
kubectl version
Create cluster
Creating a cluster includes adding a repository for charts and updating this repository, checking the version, and installing YugabyteDB.
Add charts repository
To add the YugabyteDB charts repository, run the following command:
helm repo add yugabytedb https://charts.yugabyte.com
Fetch updates from the repository
Make sure that you have the latest updates to the repository by running the following command:
helm repo update
Validate the chart version
To check the chart version, run the following command:
helm search repo yugabytedb/yugabyte --version 2.23.0
Expect output similar to the following:
NAME CHART VERSION APP VERSION DESCRIPTION
yugabytedb/yugabyte 2.23.0 2.23.0.0-b710 YugabyteDB is the high-performance distributed ...
Install YugabyteDB
Install YugabyteDB in the Kubernetes cluster using the commands described in the following sections.
On multi-node Kubernetes
Create a namespace and then install YugabyteDB, as follows:
kubectl create namespace yb-demo
helm install yb-demo yugabytedb/yugabyte --version 2.23.0 --namespace yb-demo --wait
On Minikube
If you are running in a resource-constrained environment or a local environment, such as Minikube, you have to change the default resource requirements.
Create a yb-demo
namespace, as follows:
kubectl create namespace yb-demo
helm install yb-demo yugabytedb/yugabyte \
--version 2.23.0 \
--set resource.master.requests.cpu=0.5,resource.master.requests.memory=0.5Gi,\
resource.tserver.requests.cpu=0.5,resource.tserver.requests.memory=0.5Gi --namespace yb-demo
Note that in Minikube, the LoadBalancers for yb-master-ui
and yb-tserver-service
will remain in pending state as load balancers are not available in a Minikube environment. If you would like to disable these services, pass the enableLoadBalancer=False
flag, as follows:
helm install yb-demo yugabytedb/yugabyte \
--version 2.23.0 \
--set resource.master.requests.cpu=0.5,resource.master.requests.memory=0.5Gi,\
resource.tserver.requests.cpu=0.5,resource.tserver.requests.memory=0.5Gi,\
enableLoadBalancer=False --namespace yb-demo
In some environments, such as macOS, Minikube may run inside a virtual machine. Make sure to configure the VM with at least 4 CPUs and 5 GB memory so the cluster has room to start up. The following is an example command:
minikube start --cpus 4 --memory 5120
Check the cluster status
You can check the status of the cluster using the following commands:
helm status yb-demo -n yb-demo
Expect output similar to the following:
NAME: yb-demo
LAST DEPLOYED: Thu Feb 13 13:29:13 2020
NAMESPACE: yb-demo
STATUS: deployed
REVISION: 1
TEST SUITE: None
NOTES:
1. Get YugabyteDB Pods by running this command:
kubectl --namespace yb-demo get pods
2. Get list of YugabyteDB services that are running:
kubectl --namespace yb-demo get services
3. Get information about the load balancer services:
kubectl get svc --namespace yb-demo
4. Connect to one of the tablet server:
kubectl exec --namespace yb-demo -it yb-tserver-0 -- bash
5. Run YSQL shell from inside of a tablet server:
kubectl exec --namespace yb-demo -it yb-tserver-0 -- ysqlsh -h yb-tserver-0.yb-tservers.yb-demo
6. Cleanup YugabyteDB Pods
helm delete yb-demo --purge
NOTE: You need to manually delete the persistent volume
kubectl delete pvc --namespace yb-demo -l app=yb-master
kubectl delete pvc --namespace yb-demo -l app=yb-tserver
Check the pods, as follows:
kubectl get pods --namespace yb-demo
Expect output similar to the following:
NAME READY STATUS RESTARTS AGE
yb-master-0 2/2 Running 0 4m
yb-master-1 2/2 Running 0 4m
yb-master-2 2/2 Running 0 4m
yb-tserver-0 2/2 Running 0 4m
yb-tserver-1 2/2 Running 0 4m
yb-tserver-2 2/2 Running 0 4m
Check the services, as follows:
kubectl get services --namespace yb-demo
Expect output similar to the following:
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
yb-master-ui LoadBalancer 10.109.39.242 35.225.153.213 7000:31920/TCP 10s
yb-masters ClusterIP None <none> 7100/TCP,7000/TCP 10s
yb-tserver-service LoadBalancer 10.98.36.163 35.225.153.214 6379:30929/TCP,9042:30975/TCP,5433:30048/TCP 10s
yb-tservers ClusterIP None <none> 7100/TCP,9000/TCP,6379/TCP,9042/TCP,5433/TCP 10s
You can also check the history of the yb-demo
deployment, as follows:
helm history yb-demo -n yb-demo
Expect output similar to the following:
REVISION UPDATED STATUS CHART APP VERSION DESCRIPTION
1 Thu Apr 13 13:29:13 2020 deployed yugabyte-2.13.0 2.13.0.1-b2 Install complete
Connect using YugabyteDB shells
To connect and use the YSQL Shell (ysqlsh
), run the following command:
kubectl exec -n yb-demo -it yb-tserver-0 -- ysqlsh -h yb-tserver-0.yb-tservers.yb-demo
To connect and use the YCQL Shell (ycqlsh
), run the following command:
kubectl exec -n yb-demo -it yb-tserver-0 -- ycqlsh yb-tserver-0.yb-tservers.yb-demo
Connect using external clients
To connect an external program, get the load balancer EXTERNAL-IP
address of the yb-tserver-service
service and connect using port 5433 for YSQL or port 9042 for YCQL, as follows:
kubectl get services --namespace yb-demo
Expect output similar to the following:
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
...
yb-tserver-service LoadBalancer 10.98.36.163 35.225.153.214 6379:30929/TCP,9042:30975/TCP,5433:30048/TCP 10s
...
Configure cluster
Instead of using the default values in the Helm chart, you can modify the configuration of the YugabyteDB cluster according to your requirements.
CPU, memory, and replica count
The default values for the Helm chart are in the helm/yugabyte/values.yaml
file. The following is a listing of the most important values. As noted in Prerequisites, the defaults are set for a 3-node Kubernetes cluster, each node with 4 CPU cores and 6 GB RAM allocated to YugabyteDB.
storage:
master:
count: 2
size: 10Gi
storageClass: standard
tserver:
count: 2
size: 10Gi
storageClass: standard
resource:
master:
requests:
cpu: 2
memory: 2Gi
limits:
cpu: 2
memory: 2Gi
tserver:
requests:
cpu: 2
memory: 4Gi
limits:
cpu: 2
memory: 4Gi
replicas:
master: 3
tserver: 3
partition:
master: 3
tserver: 3
If you want to change the defaults, you can use the following command. You can even do helm install
instead of helm upgrade
when you are installing on a Kubernetes cluster with configuration different than the defaults:
helm upgrade --set resource.tserver.requests.cpu=8,resource.tserver.requests.memory=15Gi yb-demo ./yugabyte
Replica count can be changed using the following command. Note that only the YB-TServers need to be scaled in a replication factor 3 cluster which keeps the masters count at 3
:
helm upgrade --set replicas.tserver=5 yb-demo ./yugabyte
Independent LoadBalancers
By default, the YugabyteDB Helm chart exposes the client API endpoints and master UI endpoint using two load balancers. If you want to expose the client APIs using independent LoadBalancers, you can execute the following command:
helm install yb-demo yugabytedb/yugabyte -f https://raw.githubusercontent.com/yugabyte/charts/master/stable/yugabyte/expose-all.yaml --version 2.23.0 --namespace yb-demo --wait
You can also bring up an internal load balancer (for either YB-Master or YB-TServer services), if required. To do so, you specify the annotation required for your cloud provider. See Amazon EKS and Google Kubernetes Engine for examples.
Reserved LoadBalancer IP Addresses
If you intend to use a preallocated (reserved) IP for the exposed YB-Master and YB-TServer services, you need to specify the load balancer IP. If you do not set this IP, a so-called ephemeral, semi-random IP will be allocated.
The following is an example of the values-overrides.yaml
file that allows you to override IP values in the Helm charts:
serviceEndpoints:
- name: "yb-master-ui"
app: "yb-master"
loadBalancerIP: "11.11.11.11"
type: "LoadBalancer"
ports:
http-ui: "7000"
- name: "yb-tserver-service"
app: "yb-tserver"
loadBalancerIP: "22.22.22.22"
type: "LoadBalancer"
ports:
tcp-yql-port: "9042"
tcp-ysql-port: "5433"
You apply the override by executing the following Helm command:
helm install yb-demo ./yugabyte -f values-overrides.yaml
Assuming that you already reserved the IP addresses (11.11.11.11 and 22.22.22.22), yb-master-ui
and yb-tserver-service
will use the predetermined addresses.
Note that setting the load balancer IP can result in behavior that might not be entirely consistent across cloud providers.
Storage class
If you want to use a storage class other than the standard class for your deployment, provision the storage class and then pass in the name of the class while running the helm install command, as follows:
helm install yugabyte --version 2.23.0 --namespace yb-demo --name yb-demo --set storage.master.storageClass=<desired storage class>,storage.tserver.storageClass=<desired storage class> --wait
Configure YB-Master and YB-TServer pods
Flags on the YB-Master and YB-TServer pods can be specified via the command line or by overriding the values.yaml
file in the charts repository. The following example shows how to set the three geo-distribution-related flags placement_cloud
, placement_region
, and placement_zone
on a Minikube cluster:
helm install yb-demo yugabytedb/yugabyte \
--version 2.23.0 \
--set resource.master.requests.cpu=0.5,resource.master.requests.memory=0.5Gi,\
resource.tserver.requests.cpu=0.5,resource.tserver.requests.memory=0.5Gi,\
gflags.master.placement_cloud=myk8s-cloud,gflags.master.placement_region=myk8s-region,gflags.master.placement_zone=myk8s-zone,\
gflags.tserver.placement_cloud=myk8s-cloud,gflags.tserver.placement_region=myk8s-region,gflags.tserver.placement_zone=myk8s-zone\
--namespace yb-demo
Upgrade cluster
You can perform rolling upgrades on the YugabyteDB cluster with the following command. Change the Image.tag
value to any valid tag from YugabyteDB's listing on the Docker Hub registry. By default, the installation uses the latest
Docker image. In the following example, the Docker image specified is 2.1.6.0-b17
:
helm upgrade yb-demo yugabytedb/yugabyte --set Image.tag=2.1.6.0-b17 --wait -n yb-demo
Delete cluster
To delete the cluster, you need to purge the Helm chart, and then delete the PVCs, as follows:
helm uninstall yb-demo -n yb-demo
kubectl delete pvc --namespace yb-demo --all