# Deploy and monitor DataRobot models in Azure Kubernetes Service

> Deploy and monitor DataRobot models in Azure Kubernetes Service - Deploy and monitor DataRobot
> models in Azure Kubernetes Service

This Markdown file sits beside the HTML page at the same path (with a `.md` suffix). It summarizes the topic and lists links for tools and LLM context.

Companion generated at `2026-04-24T16:03:56.549175+00:00` (UTC).

## Primary page

- [Deploy and monitor DataRobot models in Azure Kubernetes Service](https://docs.datarobot.com/en/docs/classic-ui/integrations/azure/aks-deploy-and-monitor.html): Full documentation for this topic (HTML).

## Sections on this page

- [Create a model](https://docs.datarobot.com/en/docs/classic-ui/integrations/azure/aks-deploy-and-monitor.html#create-a-model): In-page section heading.
- [Create an image with the model package](https://docs.datarobot.com/en/docs/classic-ui/integrations/azure/aks-deploy-and-monitor.html#create-an-image-with-the-model-package): In-page section heading.
- [Create an Azure Container Registry](https://docs.datarobot.com/en/docs/classic-ui/integrations/azure/aks-deploy-and-monitor.html#create-an-azure-container-registry): In-page section heading.
- [Push the model image to ACR](https://docs.datarobot.com/en/docs/classic-ui/integrations/azure/aks-deploy-and-monitor.html#push-the-model-image-to-acr): In-page section heading.
- [Create an AKS cluster](https://docs.datarobot.com/en/docs/classic-ui/integrations/azure/aks-deploy-and-monitor.html#create-an-aks-cluster): In-page section heading.
- [Deploy a model to AKS](https://docs.datarobot.com/en/docs/classic-ui/integrations/azure/aks-deploy-and-monitor.html#deploy-a-model-to-aks): In-page section heading.
- [Score predictions with Postman](https://docs.datarobot.com/en/docs/classic-ui/integrations/azure/aks-deploy-and-monitor.html#score-predictions-with-postman): In-page section heading.
- [Configure autoscaling and high availability](https://docs.datarobot.com/en/docs/classic-ui/integrations/azure/aks-deploy-and-monitor.html#configure-autoscaling-and-high-availability): In-page section heading.
- [Report usage to DataRobot MLOps via monitoring agents](https://docs.datarobot.com/en/docs/classic-ui/integrations/azure/aks-deploy-and-monitor.html#report-usage-to-datarobot-mlops-via-monitoring-agents): In-page section heading.
- [Report prediction details](https://docs.datarobot.com/en/docs/classic-ui/integrations/azure/aks-deploy-and-monitor.html#report-prediction-details): In-page section heading.

## Related documentation

- [Classic UI documentation](https://docs.datarobot.com/en/docs/classic-ui/index.html): Linked from this page.
- [Integrations](https://docs.datarobot.com/en/docs/classic-ui/integrations/index.html): Linked from this page.
- [Azure](https://docs.datarobot.com/en/docs/classic-ui/integrations/azure/index.html): Linked from this page.
- [Portable Prediction Server](https://docs.datarobot.com/en/docs/classic-ui/predictions/port-pred/pps/portable-pps.html): Linked from this page.
- [documentation](https://docs.datarobot.com/en/docs/classic-ui/mlops/deployment/registry/reg-create.html#register-external-model-packages): Linked from this page.

## Documentation content

# Deploy and monitor DataRobot models in Azure Kubernetes Service

> [!NOTE] Availability information
> The MLOps model package export feature is off by default. Contact your DataRobot representative or administrator for information on enabling this feature for DataRobot MLOps.
> 
> Feature flag: Enable MMM model package export

This page shows how to deploy machine learning models on Azure Kubernetes Services (AKS) to create production scoring pipelines with DataRobot's MLOps Portable Prediction Server (PPS).

DataRobot Automated Machine Learning provides a dedicated prediction server as a low-latency, synchronous REST API suitable for real-time predictions. The DataRobot MLOps PPS extends this functionality to serve ML models in container images, giving you portability and control over your ML model deployment architecture.

A containerized PPS is well-suited to deployment in a Kubernetes cluster, allowing you to take advantage of this deployment architecture's auto-scaling and high availability. The combination of PPS and Kubernetes is ideal for volatile, irregular workloads such as those you can find in IoT use cases.

## Create a model

The examples on this page use the [public LendingClub dataset](https://datarobot-doc-assets.s3.us-east-1.amazonaws.com/10K_Lending_Club_Loans.csv) to predict the loan amount for each application.

1. To upload the training data to DataRobot, do either of the following:
2. Enterloan_amtas your target (what you want to predict)(1) and clickStart(2) to run Autopilot.
3. After Autopilot finishes, clickModelsand select a model at the top of the Leaderboard.
4. Under the model you selected, clickPredict > Deployto access theGet model packagedownload.
5. ClickDownload .mlpkgto start the.mlpkgfile download.

> [!NOTE] Note
> For more information, see the documentation on the [Portable Prediction Server](https://docs.datarobot.com/en/docs/classic-ui/predictions/port-pred/pps/portable-pps.html).

## Create an image with the model package

After you [obtain the PPS base image](https://docs.datarobot.com/en/docs/classic-ui/predictions/port-pred/pps/portable-pps.html#obtain-the-pps-docker-image), create a new version of it by creating a Dockerfile with the content below:

```
FROM datarobot/datarobot-portable-prediction-api:<TAG> AS build

COPY . /opt/ml/model
```

> [!NOTE] Note
> For more information on how to structure this Docker command, see the [Docker build](https://docs.docker.com/engine/reference/builder/) documentation.

For the `COPY` command to work, you must have the `.mlpkg` file in the same directory as the Dockerfile. After creating your Dockerfile, run the command below to create a new image that includes the model:

```
docker build . --tag regressionmodel:latest
```

## Create an Azure Container Registry

Before deploying your image to AKS, push it to a container registry such as the Azure Container Registry (ACR) for deployment:

1. In the Azure Portal, clickCreate a resource > Containers, then clickContainer Registry.
2. On theCreate container registry blade, enter the following: FieldDescriptionRegistry nameEnter a suitable nameSubscriptionSelect your subscriptionResource groupUse your existing resource groupLocationSelect your regionAdmin userEnableSKUStandard
3. ClickCreate.
4. Navigate to your newly-generated registry and selectAccess Keys.
5. Copy the admin password to authenticate with Docker and push the.mlpkgimage to this registry.

## Push the model image to ACR

To push your new image to Azure Container Registry (ACR), log in with the following command (replace `<DOCKER_USERNAME>` with your previously-selected repository name):

```
docker login <DOCKER_USERNAME>.azurecr.io
```

The password is the administrator password you created with the Azure Container Registry (ACR).

Once logged in, make sure your Docker image is correctly tagged, and then push it to the repo with the following command (replace `<DOCKER_USERNAME>` with your previously selected repository name):

```
docker tag regressionmodel <DOCKER_USERNAME>.azurecr.io/regressionmodel
docker push <DOCKER_USERNAME>.azurecr.io/regressionmodel
```

## Create an AKS cluster

This section assumes you are familiar with AKS and Azure's Command Line Interface (CLI).

> [!NOTE] Note
> For more information on AKS, see [Microsoft's Quickstart tutorial](https://docs.microsoft.com/en-us/azure/aks/kubernetes-walkthrough).

1. If you don't have a running AKS cluster, create one: RESOURCE_GROUP=ai_success_engCLUSTER_NAME=AIEngineeringDemo

azakscreate\--resource-group$RESOURCE_GROUP\--name$CLUSTER_NAME\-sStandard_B2s\--node-count1\--generate-ssh-keys\--service-principalXXXXXX\--client-secretXXXX\--enable-cluster-autoscaler\--min-count1\--max-count2
2. Create a secret Docker registry so that AKS can pull images from the private repository. In the command below, replace the following with your actual credentials: kubectlcreatesecretdocker-registry<SECRET_NAME>--docker-server=<YOUR_REPOSITORY_NAME>.azurecr.io--docker-username=<DOCKER_USERNAME>--docker-password=<YOUR_SECRET_ADMIN_PW>
3. Deploy your Portable Prediction Server image. There are many ways to deploy applications, but the easiest method is via the Kubernetes dashboard. Start the Kubernetes dashboard with the following command: azaksbrowse--resource-group$RESOURCE_GROUP--name$CLUSTER_NAME

## Deploy a model to AKS

To install and deploy the PPS containing your model:

1. Click CREATE > CREATE AN APP .
2. On theCREATE AN APPpage, specify the following: FieldValueApp namee.g.,portablepredictionserverContainer imagee.g.,aisuccesseng.azurecr.io/regressionmodel:latestNumber of podse.g.,1ServiceExternalPort,Target port, andProtocol8080,8080, andTCPImage pull secretpreviously createdCPU requirement (cores)1Memory requirement (MiB)250
3. ClickDeploy—it may take several minutes to deploy.

## Score predictions with Postman

To test the model, download the [DataRobot PPS Examples](https://github.com/datarobot-community/ai_engineering/blob/master/mlops/DRMLOps_PortablePredictionServer_examples/DR%20MLOps%20Portable%20Prediction%20Server%20Public.postman_collection.json) a [Postman Collection](https://www.postman.com/collection/), and update the hostname from `localhost` to the external IP address assigned to your service. You can find the IP address in the Services tab on your Kubernetes dashboard:

To make a prediction, execute the make predictions request:

## Configure autoscaling and high availability

Kubernetes supports horizontal pod autoscaling to adjust the number of pods in a deployment depending on CPU utilization or other selected metrics. The Metrics Server provides resource utilization to Kubernetes and is automatically deployed in AKS clusters.

In the previous sections, you deployed one pod for your service and defined only the minimum requirement for CPU and memory resources.

To use the autoscaler, you must define CPU requests and utilization limits.

By default, the Portable Prediction Server spins up one worker, which means it can handle only one HTTP request simultaneously. The number of workers you can run, and thus the number of HTTP requests that it can handle simultaneously, is tied to the number of CPU cores available for the container.

Because you set the minimum CPU requirement to `1`, you can now set the limit to `2` in the `patchSpec.yaml` file:

```
apiVersion: apps/v1
kind: Deployment
metadata:
  name: portablepredictionserver
spec:
  selector:
    matchLabels:
      app: portablepredictionserver
  replicas: 1
  template:
    metadata:
      labels:
        app: portablepredictionserver
    spec:
      containers:
      - name: portablepredictionserver
        image: aisuccesseng.azurecr.io/regressionmodel:latest
        ports:
        - containerPort: 8080
        resources:
          requests:
            cpu: 1000m
          limits:
            cpu: 2000m
      imagePullSecrets:
      - name: aisuccesseng
```

Then run the following command:

```
kubectl patch deployment portablepredictionserver --patch "$(cat patchSpec.yaml)"
```

Alternatively, you can update the deployment directly in the Kubernetes dashboard by editing the JSON as shown below and clicking UPDATE.

Now that the CPU limits are defined, you can configure autoscaling with the following command:

```
kubectl autoscale deployment portablepredictionserver --cpu-percent=50 --min=1 --max=10
```

This enables Kubernetes to autoscale the number of pods in the `portablepredictionserver` deployment. If the average CPU utilization across all pods exceeds 50% of their requested usage, the autoscaler increases the pods from a minimum of one instance up to ten instances.

To run a load test, download the sample JMeter test plan below and update the URLs/ authentication. Run it with the following command:

```
jmeter -n -t LoadTesting.jmx -l results.csv
```

The output will look similar to the following example:

## Report usage to DataRobot MLOps via monitoring agents

After deploying your model to AKS, you can monitor this model, along with all of your other models, in one central dashboard by reporting the telemetrics for these predictions to your DataRobot MLOps server and dashboards.

1. Navigate to theModel Registry > Model Packages > Add New Packageand follow the instructions in thedocumentation.
2. SelectAdd new external model packageand specify a package name and description (1 and 2), upload the corresponding training data for drift tracking (3), and identify the model location (4), target (5), environment (6), and prediction type (7), then clickCreate package(8).
3. After creating the external model package, note the model ID in the URL as shown below (blurred in the image for security purposes).
4. While still on theModel Registrypage and within the expanded new package, select theDeploymentstab and clickCreate new deployment. The deployment page loads prefilled with information from the model package you created.
5. Complete any missing information for the deployment and clickCreate deployment.
6. Navigate toDeployments > Overviewand copy the deployment ID (from the URL).

Now that you have your model ID and deployment ID, you can report predictions as described in the next section.

### Report prediction details

To report prediction details to DataRobot, you need to provide a few environment variables to your Portable Prediction Server container.

Update the deployment directly in the Kubernetes dashboard by editing the JSON and then clicking UPDATE:

```
"env": [
             {
                "name": "PORTABLE_PREDICTION_API_WORKERS_NUMBER",
                "value": "2"
              },
              {
                "name": "PORTABLE_PREDICTION_API_MONITORING_ACTIVE",
                "value": "True"
              },
              {
                "name": "PORTABLE_PREDICTION_API_MONITORING_SETTINGS",
                "value": "output_type=output_dir;path=/tmp;max_files=50;file_max_size=10240000;model_id=<modelId>;deployment_id=<deployment_id>"
              },
              {
                "name": "MONITORING_AGENT",
                "value": "False"
              },
              {
                "name": "MONITORING_AGENT_DATAROBOT_APP_URL",
                "value": "https://app.datarobot.com/"
              },
              {
                "name": "MONITORING_AGENT_DATAROBOT_APP_TOKEN",
                "value": "<YOUR TOKEN>"
              }
]
```

Even though you deployed a model outside of DataRobot on a Kubernetes cluster (AKS), you can monitor it like any other model and track service health and data drift in one central dashboard (see below).
