# Create monitoring jobs

> Create monitoring jobs - Use the job definition UI to create monitoring jobs, allowing DataRobot to
> monitor deployments running and storing feature data and predictions outside of DataRobot.

This Markdown file sits beside the HTML page at the same path (with a `.md` suffix). It summarizes the topic and lists links for tools and LLM context.

Companion generated at `2026-04-24T16:03:56.619551+00:00` (UTC).

## Primary page

- [Create monitoring jobs](https://docs.datarobot.com/en/docs/classic-ui/predictions/batch/pred-monitoring-jobs/ui-monitoring-jobs.html): Full documentation for this topic (HTML).

## Sections on this page

- [Set monitoring data source](https://docs.datarobot.com/en/docs/classic-ui/predictions/batch/pred-monitoring-jobs/ui-monitoring-jobs.html#set-monitoring-data-source): In-page section heading.
- [Set monitoring options](https://docs.datarobot.com/en/docs/classic-ui/predictions/batch/pred-monitoring-jobs/ui-monitoring-jobs.html#set-monitoring-options): In-page section heading.
- [Configure predictions and actuals options](https://docs.datarobot.com/en/docs/classic-ui/predictions/batch/pred-monitoring-jobs/ui-monitoring-jobs.html#configure-predictions-and-actuals-options): In-page section heading.
- [Set aggregation options](https://docs.datarobot.com/en/docs/classic-ui/predictions/batch/pred-monitoring-jobs/ui-monitoring-jobs.html#set-aggregation-options): In-page section heading.
- [Set output monitoring and data destination options](https://docs.datarobot.com/en/docs/classic-ui/predictions/batch/pred-monitoring-jobs/ui-monitoring-jobs.html#set-output-monitoring-and-data-destination-options): In-page section heading.
- [Configure custom metric options](https://docs.datarobot.com/en/docs/classic-ui/predictions/batch/pred-monitoring-jobs/ui-monitoring-jobs.html#configure-custom-metric-options): In-page section heading.
- [Schedule monitoring jobs](https://docs.datarobot.com/en/docs/classic-ui/predictions/batch/pred-monitoring-jobs/ui-monitoring-jobs.html#schedule-monitoring-jobs): In-page section heading.
- [Save monitoring job definition](https://docs.datarobot.com/en/docs/classic-ui/predictions/batch/pred-monitoring-jobs/ui-monitoring-jobs.html#save-monitoring-job-definition): In-page section heading.

## Related documentation

- [Classic UI documentation](https://docs.datarobot.com/en/docs/classic-ui/index.html): Linked from this page.
- [Predictions](https://docs.datarobot.com/en/docs/classic-ui/predictions/index.html): Linked from this page.
- [Batch prediction methods](https://docs.datarobot.com/en/docs/classic-ui/predictions/batch/index.html): Linked from this page.
- [Prediction monitoring jobs](https://docs.datarobot.com/en/docs/classic-ui/predictions/batch/pred-monitoring-jobs/index.html): Linked from this page.
- [view and manage](https://docs.datarobot.com/en/docs/classic-ui/predictions/batch/pred-monitoring-jobs/manage-monitoring-job-def.html): Linked from this page.
- [agent-monitored deployments](https://docs.datarobot.com/en/docs/classic-ui/mlops/deployment/mlops-agent/monitoring-agent/index.html): Linked from this page.
- [define the connection](https://docs.datarobot.com/en/docs/classic-ui/data/connect-data/data-conn.html): Linked from this page.
- [intake adapter](https://docs.datarobot.com/en/docs/api/reference/batch-prediction-api/intake-options.html): Linked from this page.
- [report accuracy for the modelandits challengers](https://docs.datarobot.com/en/docs/classic-ui/mlops/deployment/mlops-agent/monitoring-agent/agent-use.html#report-accuracy-for-challengers): Linked from this page.
- [output adapter](https://docs.datarobot.com/en/docs/api/reference/batch-prediction-api/output-options.html): Linked from this page.
- [Custom metrics](https://docs.datarobot.com/en/docs/api/reference/sdk/custom-metrics.html): Linked from this page.

## Documentation content

# Create monitoring jobs via the UI

To integrate more closely with external data sources, monitoring job definitions allow DataRobot to monitor deployments running and storing feature data, predictions, actuals, and custom metrics outside of DataRobot. For example, you can create a monitoring job to connect to Snowflake, fetch raw data from the relevant Snowflake tables, and send the data to DataRobot for monitoring purposes. You can then [view and manage](https://docs.datarobot.com/en/docs/classic-ui/predictions/batch/pred-monitoring-jobs/manage-monitoring-job-def.html) monitoring job definitions as you would any other job definition.

> [!NOTE] Service health information for external models and monitoring jobs
> Service health information is unavailable for external [agent-monitored deployments](https://docs.datarobot.com/en/docs/classic-ui/mlops/deployment/mlops-agent/monitoring-agent/index.html) and deployments with predictions uploaded through a [prediction monitoring job](https://docs.datarobot.com/en/docs/classic-ui/predictions/batch/pred-monitoring-jobs/index.html).

> [!NOTE] Time series model consideration
> Monitoring jobs don't support monitoring predictions made by time series models.

To create the monitoring jobs in DataRobot:

1. ClickDeploymentsand select a deployment from the inventory.
2. On the selected deployment'sOverview, clickJob Definitions.
3. On theJob Definitionspage, clickMonitoring Jobs, and then clickAdd Job Definition.
4. On theNew Monitoring Job Definitionpage, configure the following options: Field nameDescription1Monitoring job definition nameEnter the name of the monitoring job that you are creating for the deployment.2Monitoring data sourceSet thesource typeanddefine the connectionfor the data to be scored.3Monitoring optionsConfigurepredictions and actualsorcustom metrics (preview feature)monitoring options.4Data destination(Optional)Configure the data destination optionsif you enable output monitoring.5Jobs scheduleConfigure whether to run the job immediately and whether toschedule the job.

## Set monitoring data source

Select a monitoring source, called an [intake adapter](https://docs.datarobot.com/en/docs/api/reference/batch-prediction-api/intake-options.html), and complete the appropriate authentication workflow for the source type. Select a connection type below to view field descriptions:

> [!NOTE] Note
> When browsing for connections, invalid adapters are not shown.

Database connections

- JDBC

Cloud Storage Connections

- Azure
- GCP (Google Cloud Platform Storage)
- S3

Data Warehouse Connections

- BigQuery
- Snowflake
- Synapse

Other

- AI Catalog

After you set your monitoring source, DataRobot validates that the data is applicable to the deployed model.

> [!NOTE] Note
> DataRobot validates that a data source is compatible with the model when possible, but not in all cases. DataRobot validates for AI Catalog, most JDBC connections, Snowflake, and Synapse.

## Set monitoring options

In the Monitoring options section, you can configure a Predictions and actuals monitoring job or a Custom metrics monitoring job.

### Configure predictions and actuals options

Monitoring job definitions allow DataRobot to monitor deployments that are running and storing feature data, predictions, and actuals outside of DataRobot.

In the Monitoring Options section, on the Predictions and actuals tab, the options available depend on the model type: regression or classification.

> [!NOTE] Important: Association ID for monitoring agent and monitoring jobs
> You must set an association ID before making predictions to include those predictions in accuracy tracking. For [agent-monitored](https://docs.datarobot.com/en/docs/classic-ui/mlops/deployment/mlops-agent/monitoring-agent/index.html) external model deployments with challengers (and monitoring jobs for challengers), the association ID should be `__DataRobot_Internal_Association_ID__` to [report accuracy for the modelandits challengers](https://docs.datarobot.com/en/docs/classic-ui/mlops/deployment/mlops-agent/monitoring-agent/agent-use.html#report-accuracy-for-challengers).

**Regression models:**
[https://docs.datarobot.com/en/docs/images/monitoring-options-regression.png](https://docs.datarobot.com/en/docs/images/monitoring-options-regression.png)

Option
Description
Association ID column
Identifies the column in the data source containing the association ID for predictions.
Predictions column
Identifies the column in the data source containing prediction values. You must provide this field and/or
Actuals value column
.
Actuals value column
Identifies the column in the data source containing actual values. You must provide this field and/or
Predictions column
.
Actuals timestamp column
Identifies the column in the data source containing the timestamps for actual values.

**Classification models:**
[https://docs.datarobot.com/en/docs/images/monitoring-options-classification.png](https://docs.datarobot.com/en/docs/images/monitoring-options-classification.png)

Option
Description
Association ID column
Identifies the column in the data source containing the association ID for predictions.
Predictions column
Identifies the columns in the data source containing each prediction class. You must provide this field and/or
Actuals value column
.
Actuals value column
Identifies the column in the data source containing actual values. You must provide this field and/or
Predictions column
.
Actuals timestamp column
Identifies the column in the data source containing the timestamps for actual values.


#### Set aggregation options

To support challengers for external models with [large-scale monitoring](https://docs.datarobot.com/en/docs/classic-ui/mlops/deployment/mlops-agent/monitoring-agent/agent-use.html#enable-large-scale-monitoring) enabled (meaning that raw data isn't stored in the DataRobot platform), you can report a small sample of raw feature and predictions data; then, you can send the remaining data in aggregate format. Enable Use aggregation and configure the retention settings to indicate that raw data is aggregated by the MLOps library and define how much raw data should be retained for challengers.

> [!NOTE] Autosampling for large-scale monitoring
> To automatically report a small sample of raw data for challenger analysis and accuracy monitoring, you can define the `MLOPS_STATS_AGGREGATION_AUTO_SAMPLING_PERCENTAGE` when [enabling large-scale monitoring for an external model](https://docs.datarobot.com/en/docs/classic-ui/mlops/deployment/mlops-agent/monitoring-agent/agent-use.html#enable-large-scale-monitoring).

| Property | Description |
| --- | --- |
| Retention policy | The policy definition determines if the Retention value represents a number of Samples or a Percentage of the dataset. |
| Retention value | The amount of data to retain, either a percentage of data or the number of samples. |

If you define these properties, raw data is aggregated by the MLOps library. This means that the data isn't stored in the DataRobot platform. Stats aggregation only supports feature and prediction data, not actuals data for accuracy monitoring. If you've defined one or more of the Association ID column, Actuals value column, or Actuals timestamp column, DataRobot cannot aggregate data. If you enable the Use aggregation option, the association ID and actuals-related fields are disabled.

> [!NOTE] Preview: Accuracy monitoring with aggregation
> Now available for preview, monitoring jobs for external models with aggregation enabled can support accuracy tracking. With this feature enabled, when you enable Use aggregation and configure the retention settings, you can also define the Actuals value column for accuracy monitoring; however, you must also define the Predictions column and Association ID column.
> 
> Feature flag OFF by default: Enable Accuracy Aggregation

#### Set output monitoring and data destination options

After setting the prediction and actuals monitoring options, you can choose to enable Output monitoring status and configure the following options:

| Option | Description |
| --- | --- |
| Monitored status column | Identifies the column in the data destination containing the monitoring status for each row. |
| Unique row identifier columns | Identifies the columns from the data source to serve as unique identifiers for each row. These columns are copied to the data destination to associate each monitored status with its corresponding source row. |

With Output monitoring status enabled, you must also configure the Data destination options to specify where the monitored data results should be stored. Select a monitoring data destination, called an [output adapter](https://docs.datarobot.com/en/docs/api/reference/batch-prediction-api/output-options.html), and complete the appropriate authentication workflow for the destination type. Select a connection type below to view field descriptions:

> [!NOTE] Note
> When browsing for connections, invalid adapters are not shown.

Database connections

- JDBC

Cloud Storage Connections

- Azure
- GCP (Google Cloud Platform Storage)
- S3

Data Warehouse Connections

- BigQuery
- Snowflake
- Synapse

### Configure custom metric options

> [!NOTE] Preview
> Monitoring jobs for custom metrics are off by default. Contact your DataRobot representative or administrator for information on enabling this feature.
> 
> Feature flag: Enable Custom Metrics Job Definitions

Monitoring job definitions allow DataRobot to pull calculated custom metric values from outside of DataRobot into the custom metric defined on the [Custom metrics](https://docs.datarobot.com/en/docs/api/reference/sdk/custom-metrics.html) tab, supporting custom metrics with external data sources.

In the Monitoring options section, click Custom metrics and configure the following options:

| Field | Description |
| --- | --- |
| Custom metric | Select the custom metric you want to monitor from the current deployment. |
| Value column | Select the column in the dataset containing the calculated values of the custom metric. |
| Timestamp column | Select the column in the dataset containing a timestamp. |
| Date format | Select the date format used by the timestamp column. |

## Schedule monitoring jobs

You can schedule monitoring jobs to run automatically on a schedule. When outlining a monitoring job definition, enable Run this job automatically on a schedule, then specify the frequency (daily, hourly, monthly, etc.) and time of day to define the schedule on which the job runs.

For further granularity, select Use advanced scheduler. You can set the exact time (to the minute) you want to run the monitoring job.

## Save monitoring job definition

After setting all applicable options, click Save monitoring job definition. The button text changes to Save and run monitoring job definition if Run this job immediately is enabled.

> [!NOTE] Validation errors
> This button is disabled if there are any validation errors.
