# Using managed AKS clusters¶

## Initial setup¶

If you already have an Azure Container Registry (ACR) up and ready, you can skip this section and go to Install the AKS plugin.

Otherwise, follow the Azure documentation on how to create your ACR registry.

Warning

We recommend that you pay extra attention to the Azure container registry pricing plan, as it is directly related to the registry storage capacity.

### Install the AKS plugin¶

To use Microsoft Azure Kubernetes Service (AKS), begin by installing the “AKS clusters” plugin from the Plugins store in Dataiku DSS. For more details, see the instructions for installing plugins.

### Prepare your local az, docker, and kubectl commands¶

Follow the Azure documentation to ensure the following on your local machine (where DSS is installed):

• The az command is properly logged in. As of October 2019, this implies running the az login --service-principal --username client_d --password client_secret --tenant tenant_id command.

• The docker command can successfully push images to the ACR repository. As of October 2019, this implies running the az acr login --name your-registry-name command.

• The kubectl command can interact with the cluster. As of October 2019, this implies running the az aks get-credentials --resource-group your-rg --name your-cluster-name command.

### Create base images¶

Build the base image by following these instructions.

### Create a new containerized execution configuration¶

Go to Administration > Settings > Containerized execution, and add a new execution configuration of type “Kubernetes”.

• In particular, to set up the image registry, the URL must be of the form your-registry-name.azurecr.io.

• Finish by clicking Push base images.

## Cluster configuration¶

• The connection is where you define how to connect to Azure. This can be done either inline in each cluster (not recommended) or via a preset in the “AKS connection” plugin settings (recommended).

• By default, the service principal associated to the cluster will be the same as the one used on the DSS machine. You can change this by enabling the Use distinct credentials option and defining a specific connection, either inline or via a preset.

• Cluster nodes is where you define the number and type of nodes that you want in your cluster. You can define the properties of a node pool either inline (not recommended) or as a preset in the “Node pools” plugin settings (recommended). You have the possibility to define multiple node pools, each with its own properties.

## Using GPUs¶

Azure provides GPU-enabled instances with NVidia GPUs. Using GPUs for containerized execution requires the following steps.

### Building an image with CUDA support¶

The base image that is built by default (see setup) does not have CUDA support and cannot use NVidia GPUs. You need to build a CUDA-enabled base image. To enable CUDA add the --with-cuda option to the command line:

./bin/dssadmin build-base-image --type container-exec --with-cuda

We recommend that you give this image a specific tag using the --tag option and keep the default base image “pristine”. We also recommend that you add the DSS version number in the image tag.

./bin/dssadmin build-base-image --type container-exec --with-cuda --tag dataiku-container-exec-base-cuda:X.Y.Z

where X.Y.Z is your DSS version number

Note

• This image contains CUDA 10.0 and CuDNN 7.6. You can use --cuda-version X.Y to specify another DSS-provided version (9.0, 10.0, 10.1, 10.2, 11.0 and 11.2 are available). If you require other CUDA versions, you would have to create a custom image.

• Remember that depending on which CUDA version you build the base image (by default 10.0) you will need to use the corresponding tensorflow version.

Warning

After each upgrade of DSS, you must rebuild all base images and update code envs.

Thereafter, create a new container configuration dedicated to running GPU workloads. If you specified a tag for the base image, report it in the “Base image tag” field.

### Enable GPU support on the cluster¶

When you create your cluster using the AKS plugin, be sure to select a VM with a GPU. See Azure documentation for a full list. You’ll also need to enable the “With GPU” option in the node pool settings.

At cluster creation, the plugin will run the NVidia driver “DaemonSet” installation procedure, which needs several minutes to complete.

For your containerized execution task to run on nodes with GPUs, and for AKS to configure the CUDA driver on your containers, the corresponding pods must be created with a custom limit (in Kubernetes parlance). This indicates that you need a specific type of resource (standard resource types are CPU and memory).

You must configure this limit in the containerized execution configuration. To do this:

• In the “Custom limits” section, add a new entry with key nvidia.com/gpu and value 1 (to request 1 GPU).

To disable this, set a DISABLE_AZURE_USAGE_ATTRIBUTION environment variable to 1 in DATADIR/bin/env-site.sh.