DigitalOcean Kubernetes (DOKS) is a managed Kubernetes service. Deploy Kubernetes clusters with a fully managed control plane, high availability, autoscaling, and native integration with DigitalOcean Load Balancers and volumes. You can add node pools using shared and dedicated CPUs, and NVIDIA H100 GPUs in a single GPU or 8 GPU configuration. DOKS clusters are compatible with standard Kubernetes toolchains and the DigitalOcean API and CLI.
The API and CLI often require region, node size, and version values. Clusters run on Droplets and determine the node sizes available for the size
field when creating a cluster. You can retrieve a list of available option values using the /v2/kubernetes/options
endpoint or the doctl kubernetes options
command. You can also reference our Droplet pricing page for a full list of Droplet sizes and their slugs.
See the region availability table for a list of region slugs.
You can also create a new cluster with GPU node pools.
To run GPU workloads after you create a cluster, use the GPU nodes-specific labels and taint in your workload specifications to schedule pods that match. You can use a configuration spec, similar to the pod spec shown below, for your actual workloads:
apiVersion: v1
kind: Pod
metadata:
name: gpu-workload
spec:
restartPolicy: Never
nodeSelector:
doks.digitalocean.com/gpu-brand: nvidia
containers:
- name: cuda-container
image: nvcr.io/nvidia/k8s/cuda-sample:vectoradd-cuda10.2
tolerations:
- key: nvidia.com/gpu
operator: Exists
effect: NoSchedule
The above spec shows how to create a pod that runs NVIDIA’s CUDA image and uses the labels and taint for GPU worker nodes.
You can use the cluster autoscaler to scale the GPU node pool down to 1 or use the DigitalOcean CLI or API to manually scale the node pool down to 0. For example:
doctl kubernetes cluster node-pool update <your-cluster-id> <your-nodepool-id> --count 0
You can also create a new cluster with VPC-native networking. This feature is only available on new clusters using Kubernetes 1.31+.
To create a cluster with VPC-native networking, you need to provide two network subnets in CIDR notation (for example, 192.168.0.0/16
), one for the cluster and another for services running on the cluster. The subnets must follow RFC 1918 guidelines and must not overlap with each other or with any VPCs or VPC-native clusters within the same team.
The chosen subnet size determines how many nodes can run in a cluster, and how many services can be created, and it cannot be extended later. Every node is assigned a /25
network out of the cluster subnet, and you can create one service for each IP address in the services subnet.
We recommend using /16
for the cluster and /19
for services, which results in a cluster that can have up to 512 nodes and 8192 services. Avoid choosing smaller than /20
for the cluster (64 nodes) and /22
for services (1024 services).
When you have determined your subnet ranges, use the doctl
CLI or the API to create a cluster.
You can create a DigitalOcean Kubernetes Cluster at any time from the DigitalOcean Control Panel by opening the Create menu in the top right.
In the Create menu, click Kubernetes to go to the Create a cluster page. On this page, you choose a Kubernetes version, datacenter region, and cluster capacity for your cluster and then create it.
The latest patch version for the three most recent minor versions of Kubernetes are available for new cluster creation. The latest stable release is selected by default.
Choose the region for your cluster. Your cluster’s control plane and worker nodes are located in the same region.
If you plan to use GPU worker nodes or volumes for persistent data storage, choose a region with GPU worker nodes support or volumes support. If you add a DigitalOcean Load Balancer to your deployment, it is automatically placed in the same region as the cluster.
In the VPC Network section, choose a VPC network for the cluster. You can choose one you’ve created or use your default network for the datacenter region. VPC enables an additional networking interface that can only be accessed by other resources within the same VPC network. This keeps traffic between Droplets and other applicable resource from being routed outside the datacenter over the public internet.
As of 12 December 2024 new clusters created in the control panel using Kubernetes 1.31 or higher are created with VPC-native cluster networking. VPC-native networking enables routing traffic directly between pods and services and other resources on VPC networks.
Choose Size network subnets for me to use the default /16
(512 nodes) pod network and /19
(8192 services) service network.
To customize the pod and service network sizes, choose Configure my own network subnet sizes. The network prefixes are automatically populated from available network ranges in your team account.
A DOKS cluster has one or more node pools. Each node pool consists of a group of identical worker nodes. Worker nodes are built on Droplets.
To create a cluster, specify the following for the node pool:
Node pool name: Choose a name for the node pool when it’s created. Nodes inside this pool inherit this naming scheme when they are created. If you rename the node pool later, the nodes only inherit the new naming scheme when they are recreated (when you recycle the nodes or resize the node pool).
Choose the machine type:
Shared CPUs are built on Basic Droplet plans.
Dedicated CPUs are built on General Purpose, CPU-Optimized (Regular Intel CPU or Premium Intel CPU), or Memory-Optimized Droplet plans.
GPUs are built on GPU Droplets which are powered by NVIDIA’s H100 GPUs and can be in a single GPU or 8 GPU configuration.
All machine types, except the Basic nodes, are dedicated CPU Droplets.
Select the node type from the Machine type (Droplet) dropdown list. Choosing the right Kubernetes plan highly depends on your workload. See Choosing the Right Kubernetes Plan for more information.
Node plan: Choose the specific plan you want for your worker nodes. Each of the workers in a node pool has identical resources.
Select the Set node pool to autoscale option to enable autoscaling.
Nodes: For a fixed-size cluster, choose how many nodes to include in the node pool. By default, three worker nodes are selected.
Minimum nodes and Maximum nodes: For an autoscaling-enabled cluster, choose the minimum and maximum number of nodes for when the load decreases or increases.
To take advantage of different resource capacities, you can add additional node pools with the Add Additional Node Pool button and assign pods to the node pools with the appropriate scheduling constraints.
minikube
generally gives you better performance than a one-node cluster for the same cost.Then, you can optionally select the following settings.
DigitalOcean Kubernetes provides a high availability (HA) option that increases uptime and provides 99.95% SLA uptime for control planes. In the Get extra reliability for critical workloads section, select the Add high availability checkbox. Using this option creates multiple backup replicas of each control plane component and provides extra reliability for critical workloads.
To integrate your Kubernetes cluster with a DigitalOcean database, we recommend adding a free database operator during creation, now in beta. The database operator allows you to create new databases, automatically link them to your Kubernetes cluster, and manage them. For detailed information on using the operator, see our documentation on GitHub and how-to guide for diagrams.
While the operator is free, any databases you create with the operator are billed normally.
Finalize the cluster settings. You can specify a name, assign a project, and optionally add tags to the cluster.
By default, cluster names begin with k8s
, followed by the version of Kubernetes, the datacenter region, and the cluster ID. You can customize the cluster name, which is also used in the tag.
The new cluster belongs to your default project. You can assign the cluster to a different project.
You can also change the project after you create the cluster. Go to the Kubernetes page in the control panel. From the cluster’s More menu, select Move to and select the project you want to move the cluster to.
Associated resources, such as load balancers and volumes, also move when you move the cluster to a different project.
Due to a known issue, resources associated with a DOKS cluster – such as load balancers, volumes, and volume snapshots – belong to the default project upon creation, regardless of which project the cluster belongs to. You can associate resources to the correct project by reconciling the cluster or moving the cluster to a different project.
Clusters automatically have three tags:
k8s
k8s:EXAMPLEc-3515-4a0c-91a3-2452eEXAMPLE
k8s:worker
You can also add custom tags to a cluster and its node pools in the cluster’s Overview and Resources pages. Any custom tags you add to worker nodes in a node pool (for example, from the Droplets page), are deleted to maintain consistency between the node pool and its worker nodes.
At the bottom of this section, you see the Total monthly cost for your cluster based on the resources you’ve chosen. When you create the cluster, billing begins for each resource (for example, worker nodes, volumes, load balancers) as it is created and ends as it is destroyed.
When you have entered your settings, create the cluster by clicking the Create Cluster button. It can take several minutes for cluster creation to finish.
Once your cluster is provisioned, you can view cluster information such as the cluster configuration, networking details, cost, cluster ID, and control plane high availability indicator in the Overview tab.
To manage the cluster, use kubectl
, the official Kubernetes command-line client. See How to Connect to a DigitalOcean Kubernetes Cluster to get set up.