Skip to content

On-premise users: click in-app to access the full platform documentation for your version of DataRobot.

Monitoring Jobs

This page outlines the operations, endpoints, parameters, and example requests and responses for the Monitoring Jobs.

POST /api/v2/batchMonitoring/

Submit the configuration for the job and it will be submitted to the queue

Code samples

curl -X POST https://app.datarobot.com/api/v2/batchMonitoring/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{BatchMonitoringJobCreate}'

Body parameter

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Parameters

Name In Type Required Description
body body BatchMonitoringJobCreate false none

Example responses

202 Response

{
  "batchMonitoringJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "batchPredictionJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "elapsedTimeSec": 0,
  "failedRows": 0,
  "hidden": "2019-08-24T14:15:22Z",
  "id": "string",
  "intakeDatasetDisplayName": "string",
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "jobSpec": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 1,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "links": {
    "csvUpload": "string",
    "download": "string",
    "self": "string"
  },
  "logs": [
    "string"
  ],
  "monitoringBatchId": "string",
  "percentageCompleted": 100,
  "queuePosition": 0,
  "queued": true,
  "resultsDeleted": true,
  "scoredRows": 0,
  "skippedRows": 0,
  "source": "string",
  "status": "INITIALIZING",
  "statusDetails": "string"
}

Responses

Status Meaning Description Schema
202 Accepted Job details for the created Batch Monitoring job BatchJobResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/batchMonitoringJobDefinitions/

List all Batch Monitoring jobs definitions available

Code samples

curl -X GET https://app.datarobot.com/api/v2/batchMonitoringJobDefinitions/?offset=0&limit=100 \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer true This many results will be skipped
limit query integer true At most this many results are returned
searchName query string false A human-readable name for the definition, must be unique across organisations.
deploymentId query string false Includes only definitions for this particular deployment

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "batchMonitoringJob": {
        "abortOnError": true,
        "batchJobType": "monitoring",
        "chunkSize": "auto",
        "columnNamesRemapping": {},
        "csvSettings": {
          "delimiter": ",",
          "encoding": "utf-8",
          "quotechar": "\""
        },
        "deploymentId": "string",
        "disableRowLevelErrorHandling": false,
        "explanationAlgorithm": "shap",
        "explanationClassNames": [
          "string"
        ],
        "explanationNumTopClasses": 1,
        "includePredictionStatus": false,
        "includeProbabilities": true,
        "includeProbabilitiesClasses": [],
        "intakeSettings": {
          "type": "localFile"
        },
        "maxExplanations": 0,
        "maxNgramExplanations": 0,
        "modelId": "string",
        "modelPackageId": "string",
        "monitoringAggregation": {
          "retentionPolicy": "samples",
          "retentionValue": 0
        },
        "monitoringBatchPrefix": "string",
        "monitoringColumns": {
          "actedUponColumn": "string",
          "actualsTimestampColumn": "string",
          "actualsValueColumn": "string",
          "associationIdColumn": "string",
          "customMetricId": "string",
          "customMetricTimestampColumn": "string",
          "customMetricTimestampFormat": "string",
          "customMetricValueColumn": "string",
          "monitoredStatusColumn": "string",
          "predictionsColumns": [
            {
              "className": "string",
              "columnName": "string"
            }
          ],
          "reportDrift": true,
          "reportPredictions": true,
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "monitoringOutputSettings": {
          "monitoredStatusColumn": "string",
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "numConcurrent": 0,
        "outputSettings": {
          "credentialId": "string",
          "format": "csv",
          "partitionColumns": [
            "string"
          ],
          "type": "azure",
          "url": "string"
        },
        "passthroughColumns": [
          "string"
        ],
        "passthroughColumnsSet": "all",
        "pinnedModelId": "string",
        "predictionInstance": {
          "apiKey": "string",
          "datarobotKey": "string",
          "hostName": "string",
          "sslEnabled": true
        },
        "predictionWarningEnabled": true,
        "redactedFields": [
          "string"
        ],
        "skipDriftTracking": false,
        "thresholdHigh": 0,
        "thresholdLow": 0,
        "timeseriesSettings": {
          "forecastPoint": "2019-08-24T14:15:22Z",
          "relaxKnownInAdvanceFeaturesCheck": false,
          "type": "forecast"
        }
      },
      "created": "2019-08-24T14:15:22Z",
      "createdBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      },
      "enabled": false,
      "id": "string",
      "lastFailedRunTime": "2019-08-24T14:15:22Z",
      "lastScheduledRunTime": "2019-08-24T14:15:22Z",
      "lastStartedJobStatus": "INITIALIZING",
      "lastStartedJobTime": "2019-08-24T14:15:22Z",
      "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
      "name": "string",
      "nextScheduledRunTime": "2019-08-24T14:15:22Z",
      "schedule": {
        "dayOfMonth": [
          "*"
        ],
        "dayOfWeek": [
          "*"
        ],
        "hour": [
          "*"
        ],
        "minute": [
          "*"
        ],
        "month": [
          "*"
        ]
      },
      "updated": "2019-08-24T14:15:22Z",
      "updatedBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      }
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Responses

Status Meaning Description Schema
200 OK List of all available jobs BatchMonitoringJobDefinitionsListResponse
422 Unprocessable Entity Your input data or query arguments did not work together None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/batchMonitoringJobDefinitions/

Create a Batch Monitoring Job definition. A configuration for a Batch Monitoring job which can either be executed manually upon request or on scheduled intervals, if enabled. The API payload is the same as for /batchJobs along with optional enabled and schedule items.

Code samples

curl -X POST https://app.datarobot.com/api/v2/batchMonitoringJobDefinitions/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{BatchMonitoringJobDefinitionsCreate}'

Body parameter

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "enabled": true,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "name": "string",
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Parameters

Name In Type Required Description
body body BatchMonitoringJobDefinitionsCreate false none

Example responses

202 Response

{
  "batchMonitoringJob": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 0,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "enabled": false,
  "id": "string",
  "lastFailedRunTime": "2019-08-24T14:15:22Z",
  "lastScheduledRunTime": "2019-08-24T14:15:22Z",
  "lastStartedJobStatus": "INITIALIZING",
  "lastStartedJobTime": "2019-08-24T14:15:22Z",
  "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
  "name": "string",
  "nextScheduledRunTime": "2019-08-24T14:15:22Z",
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "updated": "2019-08-24T14:15:22Z",
  "updatedBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  }
}

Responses

Status Meaning Description Schema
202 Accepted Job details for the created Batch Monitoring job definition BatchMonitoringJobDefinitionsResponse
403 Forbidden You are not authorized to create a job definition on this deployment due to your permissions role None
422 Unprocessable Entity You tried to create a job definition with incompatible or missing parameters to create a fully functioning job definition None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

DELETE /api/v2/batchMonitoringJobDefinitions/{jobDefinitionId}/

Delete a Batch Prediction job definition

Code samples

curl -X DELETE https://app.datarobot.com/api/v2/batchMonitoringJobDefinitions/{jobDefinitionId}/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
jobDefinitionId path string true ID of the Batch Prediction job definition

Responses

Status Meaning Description Schema
204 No Content none None
403 Forbidden You are not authorized to delete this job definition due to your permissions role None
404 Not Found Job was deleted, never existed or you do not have access to it None
409 Conflict Job could not be deleted, as there are currently running jobs in the queue. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/batchMonitoringJobDefinitions/{jobDefinitionId}/

Retrieve a Batch Monitoring job definition

Code samples

curl -X GET https://app.datarobot.com/api/v2/batchMonitoringJobDefinitions/{jobDefinitionId}/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
jobDefinitionId path string true ID of the Batch Prediction job definition

Example responses

200 Response

{
  "batchMonitoringJob": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 0,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "enabled": false,
  "id": "string",
  "lastFailedRunTime": "2019-08-24T14:15:22Z",
  "lastScheduledRunTime": "2019-08-24T14:15:22Z",
  "lastStartedJobStatus": "INITIALIZING",
  "lastStartedJobTime": "2019-08-24T14:15:22Z",
  "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
  "name": "string",
  "nextScheduledRunTime": "2019-08-24T14:15:22Z",
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "updated": "2019-08-24T14:15:22Z",
  "updatedBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  }
}

Responses

Status Meaning Description Schema
200 OK Job details for the requested Batch Monitoring job definition BatchMonitoringJobDefinitionsResponse
404 Not Found Job was deleted, never existed or you do not have access to it None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

PATCH /api/v2/batchMonitoringJobDefinitions/{jobDefinitionId}/

Update a Batch Monitoring job definition

Code samples

curl -X PATCH https://app.datarobot.com/api/v2/batchMonitoringJobDefinitions/{jobDefinitionId}/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "enabled": true,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "name": "string",
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Parameters

Name In Type Required Description
jobDefinitionId path string true ID of the Batch Prediction job definition
body body BatchMonitoringJobDefinitionsUpdate false none

Example responses

200 Response

{
  "batchMonitoringJob": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 0,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "enabled": false,
  "id": "string",
  "lastFailedRunTime": "2019-08-24T14:15:22Z",
  "lastScheduledRunTime": "2019-08-24T14:15:22Z",
  "lastStartedJobStatus": "INITIALIZING",
  "lastStartedJobTime": "2019-08-24T14:15:22Z",
  "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
  "name": "string",
  "nextScheduledRunTime": "2019-08-24T14:15:22Z",
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "updated": "2019-08-24T14:15:22Z",
  "updatedBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  }
}

Responses

Status Meaning Description Schema
200 OK Job details for the updated Batch Monitoring job definition BatchMonitoringJobDefinitionsResponse
403 Forbidden You are not authorized to alter the contents of this monitoring job due to your permissions role None
404 Not Found Job was deleted, never existed or you do not have access to it None
409 Conflict You chose a name of your monitoring job that was already existing within your organization None
422 Unprocessable Entity Could not update the monitoring job. Possible reasons: {} None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

Schemas

AzureDataStreamer

{
  "credentialId": "string",
  "format": "csv",
  "type": "azure",
  "url": "string"
}

Stream CSV data chunks from Azure

Properties

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false Use the specified credential to access the url

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
format string false Type of input file format
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
anonymous [redacted]
format [csv, parquet]
type azure

AzureIntake

{
  "credentialId": "string",
  "format": "csv",
  "type": "azure",
  "url": "string"
}

Stream CSV data chunks from Azure

Properties

Name Type Required Restrictions Description
credentialId string,null false Use the specified credential to access the url
format string false Type of input file format
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
format [csv, parquet]
type azure

AzureOutput

{
  "credentialId": "string",
  "format": "csv",
  "partitionColumns": [
    "string"
  ],
  "type": "azure",
  "url": "string"
}

Save CSV data chunks to Azure Blob Storage

Properties

Name Type Required Restrictions Description
credentialId string,null false Use the specified credential to access the url
format string false Type of output file format
partitionColumns [string] false maxItems: 100
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/").
type string true Type name for this output type
url string(url) true URL for the file or directory

Enumerated Values

Property Value
format [csv, parquet]
type azure

AzureOutputAdaptor

{
  "credentialId": "string",
  "format": "csv",
  "partitionColumns": [
    "string"
  ],
  "type": "azure",
  "url": "string"
}

Save CSV data chunks to Azure Blob Storage

Properties

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false Use the specified credential to access the url

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
format string false Type of output file format
partitionColumns [string] false maxItems: 100
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/").
type string true Type name for this output type
url string(url) true URL for the file or directory

Enumerated Values

Property Value
anonymous [redacted]
format [csv, parquet]
type azure

BatchJobCSVSettings

{
  "delimiter": ",",
  "encoding": "utf-8",
  "quotechar": "\""
}

The CSV settings used for this job

Properties

Name Type Required Restrictions Description
delimiter any true CSV fields are delimited by this character. Use the string "tab" to denote TSV (TAB separated values).

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous string false maxLength: 1
minLength: 1
minLength: 1
none

continued

Name Type Required Restrictions Description
encoding string true The encoding to be used for intake and output. For example (but not limited to): "shift_jis", "latin_1" or "mskanji".
quotechar string true maxLength: 1
minLength: 1
minLength: 1
Fields containing the delimiter or newlines must be quoted using this character.

Enumerated Values

Property Value
anonymous tab

BatchJobCreatedBy

{
  "fullName": "string",
  "userId": "string",
  "username": "string"
}

Who created this job

Properties

Name Type Required Restrictions Description
fullName string,null true The full name of the user who created this job (if defined by the user)
userId string true The User ID of the user who created this job
username string true The username (e-mail address) of the user who created this job

BatchJobDefinitionResponse

{
  "createdBy": "string",
  "id": "string",
  "name": "string"
}

The Batch Prediction Job Definition linking to this job, if any.

Properties

Name Type Required Restrictions Description
createdBy string true The ID of creator of this job definition
id string true The ID of the Batch Prediction job definition
name string true A human-readable name for the definition, must be unique across organisations

BatchJobDefinitionsSpecResponse

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "maxNgramExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "numConcurrent": 0,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionWarningEnabled": true,
  "redactedFields": [
    "string"
  ],
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

The Batch Monitoring Job specification to be put on the queue in intervals

Properties

Name Type Required Restrictions Description
abortOnError boolean true Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchJobCSVSettings true The CSV settings used for this job
deploymentId string false ID of deployment which is used in job for processing predictions dataset
disableRowLevelErrorHandling boolean true Skip row by row error handling
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean true Include prediction status column in the output
includeProbabilities boolean true Include probabilities for all classes
includeProbabilitiesClasses [string] true maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any true The response option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureDataStreamer false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous DataStageDataStreamer false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous CatalogDataStreamer false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous GCPDataStreamer false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryDataStreamer false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous S3DataStreamer false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeDataStreamer false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseDataStreamer false Stream CSV data chunks from Azure Synapse

xor

Name Type Required Restrictions Description
» anonymous DSSDataStreamer false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemDataStreamer false none

xor

Name Type Required Restrictions Description
» anonymous HTTPDataStreamer false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCDataStreamer false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileDataStreamer false Stream CSV data chunks from local file storage

continued

Name Type Required Restrictions Description
maxExplanations integer true maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
maxNgramExplanations any false The maximum number of text ngram explanations to supply per row of the dataset. The default recommended maxNgramExplanations is all (no limit)

oneOf

Name Type Required Restrictions Description
» anonymous integer false minimum: 0
none

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringAggregation MonitoringAggregation false Defines the aggregation policy for monitoring jobs.
monitoringBatchPrefix string,null false Name of the batch to create with this job
monitoringColumns MonitoringColumnsMapping false Column names mapping for monitoring
monitoringOutputSettings MonitoringOutputSettings false Output settings for monitoring jobs
numConcurrent integer false minimum: 0
Number of simultaneous requests to run against the prediction instance
outputSettings any false The response option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutputAdaptor false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous GCPOutputAdaptor false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutputAdaptor false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous S3OutputAdaptor false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutputAdaptor false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutputAdaptor false Save CSV data chunks to Azure Synapse in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutputAdaptor false none

xor

Name Type Required Restrictions Description
» anonymous HttpOutputAdaptor false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JdbcOutputAdaptor false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutputAdaptor false Save CSV data chunks to local file storage

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionWarningEnabled boolean,null false Enable prediction warnings.
redactedFields [string] true A list of qualified field names from intake- and/or outputSettings that was redacted due to permissions and sharing settings. For example: intakeSettings.dataStoreId
skipDriftTracking boolean true Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsHistorical false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
anonymous all
passthroughColumnsSet all

{
  "csvUpload": "string",
  "download": "string",
  "self": "string"
}

Links useful for this job

Properties

Name Type Required Restrictions Description
csvUpload string(url) false The URL used to upload the dataset for this job. Only available for localFile intake.
download string,null false The URL used to download the results from this job. Only available for localFile outputs. Will be null if the download is not yet available.
self string(url) true The URL used access this job.

BatchJobPredictionInstance

{
  "apiKey": "string",
  "datarobotKey": "string",
  "hostName": "string",
  "sslEnabled": true
}

Override the default prediction instance from the deployment when scoring this job.

Properties

Name Type Required Restrictions Description
apiKey string false By default, prediction requests will use the API key of the user that created the job. This allows you to make requests on behalf of other users.
datarobotKey string false If running a job against a prediction instance in the Managed AI Cloud, you must provide the organization level DataRobot-Key.
hostName string true Override the default host name of the deployment with this.
sslEnabled boolean true Use SSL (HTTPS) when communicating with the overriden prediction server.

BatchJobRemapping

{
  "inputName": "string",
  "outputName": "string"
}

Properties

Name Type Required Restrictions Description
inputName string true Rename column with this name
outputName string,null true Rename column to this name (leave as null to remove from the output)

BatchJobResponse

{
  "batchMonitoringJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "batchPredictionJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "elapsedTimeSec": 0,
  "failedRows": 0,
  "hidden": "2019-08-24T14:15:22Z",
  "id": "string",
  "intakeDatasetDisplayName": "string",
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "jobSpec": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 1,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "links": {
    "csvUpload": "string",
    "download": "string",
    "self": "string"
  },
  "logs": [
    "string"
  ],
  "monitoringBatchId": "string",
  "percentageCompleted": 100,
  "queuePosition": 0,
  "queued": true,
  "resultsDeleted": true,
  "scoredRows": 0,
  "skippedRows": 0,
  "source": "string",
  "status": "INITIALIZING",
  "statusDetails": "string"
}

Properties

Name Type Required Restrictions Description
batchMonitoringJobDefinition BatchJobDefinitionResponse false The Batch Prediction Job Definition linking to this job, if any.
batchPredictionJobDefinition BatchJobDefinitionResponse false The Batch Prediction Job Definition linking to this job, if any.
created string(date-time) true When was this job created
createdBy BatchJobCreatedBy true Who created this job
elapsedTimeSec integer true minimum: 0
Number of seconds the job has been processing for
failedRows integer true minimum: 0
Number of rows that have failed scoring
hidden string(date-time) false When was this job was hidden last, blank if visible
id string true The ID of the Batch job
intakeDatasetDisplayName string,null false If applicable (e.g. for AI catalog), will contain the dataset name used for the intake dataset.
jobIntakeSize integer,null true minimum: 0
Number of bytes in the intake dataset for this job
jobOutputSize integer,null true minimum: 0
Number of bytes in the output dataset for this job
jobSpec BatchJobSpecResponse true The job configuration used to create this job
links BatchJobLinks true Links useful for this job
logs [string] true The job log.
monitoringBatchId string,null true Id of the monitoring batch created by this job. Only present if the job runs on a deployment with batch monitoring enabled.
percentageCompleted number true maximum: 100
minimum: 0
Indicates job progress which is based on number of already processed rows in dataset
queuePosition integer,null false minimum: 0
To ensure a dedicated prediction instance is not overloaded, only one job will be run against it at a time. This is the number of jobs that are awaiting processing before this job start running. May not be available in all environments.
queued boolean true The job has been put on the queue for execution.
resultsDeleted boolean false Indicates if the job was subject to garbage collection and had its artifacts deleted (output files, if any, and scoring data on local storage)
scoredRows integer true minimum: 0
Number of rows that have been used in prediction computation
skippedRows integer true minimum: 0
Number of rows that have been skipped during scoring. May contain non-zero value only in time-series predictions case if provided dataset contains more than required historical rows.
source string false Source from which batch job was started
status string true The current job status
statusDetails string true Explanation for current status

Enumerated Values

Property Value
status [INITIALIZING, RUNNING, COMPLETED, ABORTED, FAILED]

BatchJobSpecResponse

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "maxNgramExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionWarningEnabled": true,
  "redactedFields": [
    "string"
  ],
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

The job configuration used to create this job

Properties

Name Type Required Restrictions Description
abortOnError boolean true Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchJobCSVSettings true The CSV settings used for this job
deploymentId string false ID of deployment which is used in job for processing predictions dataset
disableRowLevelErrorHandling boolean true Skip row by row error handling
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean true Include prediction status column in the output
includeProbabilities boolean true Include probabilities for all classes
includeProbabilitiesClasses [string] true maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any true The response option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureDataStreamer false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous DataStageDataStreamer false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous CatalogDataStreamer false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous GCPDataStreamer false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryDataStreamer false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous S3DataStreamer false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeDataStreamer false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseDataStreamer false Stream CSV data chunks from Azure Synapse

xor

Name Type Required Restrictions Description
» anonymous DSSDataStreamer false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemDataStreamer false none

xor

Name Type Required Restrictions Description
» anonymous HTTPDataStreamer false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCDataStreamer false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileDataStreamer false Stream CSV data chunks from local file storage

continued

Name Type Required Restrictions Description
maxExplanations integer true maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
maxNgramExplanations any false The maximum number of text ngram explanations to supply per row of the dataset. The default recommended maxNgramExplanations is all (no limit)

oneOf

Name Type Required Restrictions Description
» anonymous integer false minimum: 0
none

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringAggregation MonitoringAggregation false Defines the aggregation policy for monitoring jobs.
monitoringBatchPrefix string,null false Name of the batch to create with this job
monitoringColumns MonitoringColumnsMapping false Column names mapping for monitoring
monitoringOutputSettings MonitoringOutputSettings false Output settings for monitoring jobs
numConcurrent integer false minimum: 1
Number of simultaneous requests to run against the prediction instance
outputSettings any false The response option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutputAdaptor false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous GCPOutputAdaptor false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutputAdaptor false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous S3OutputAdaptor false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutputAdaptor false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutputAdaptor false Save CSV data chunks to Azure Synapse in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutputAdaptor false none

xor

Name Type Required Restrictions Description
» anonymous HttpOutputAdaptor false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JdbcOutputAdaptor false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutputAdaptor false Save CSV data chunks to local file storage

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionWarningEnabled boolean,null false Enable prediction warnings.
redactedFields [string] true A list of qualified field names from intake- and/or outputSettings that was redacted due to permissions and sharing settings. For example: intakeSettings.dataStoreId
skipDriftTracking boolean true Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsHistorical false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
anonymous all
passthroughColumnsSet all

BatchJobTimeSeriesSettingsForecast

{
  "forecastPoint": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "forecast"
}

Properties

Name Type Required Restrictions Description
forecastPoint string(date-time) false Used for forecast predictions in order to override the inferred forecast point from the dataset.
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Forecast mode makes predictions using forecastPoint or rows in the dataset without target.

Enumerated Values

Property Value
type forecast

BatchJobTimeSeriesSettingsHistorical

{
  "predictionsEndDate": "2019-08-24T14:15:22Z",
  "predictionsStartDate": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "historical"
}

Properties

Name Type Required Restrictions Description
predictionsEndDate string(date-time) false Used for historical predictions in order to override date to which predictions should be calculated. By default value will be inferred automatically from the dataset.
predictionsStartDate string(date-time) false Used for historical predictions in order to override date from which predictions should be calculated. By default value will be inferred automatically from the dataset.
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Historical mode enables bulk predictions which calculates predictions for all possible forecast points and forecast distances in the dataset within the predictionsStartDate/predictionsEndDate range.

Enumerated Values

Property Value
type historical

BatchMonitoringJobCreate

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Properties

Name Type Required Restrictions Description
abortOnError boolean true Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchPredictionJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchPredictionJobCSVSettings true The CSV settings used for this job
deploymentId string false ID of deployment which is used in job for processing predictions dataset
disableRowLevelErrorHandling boolean true Skip row by row error handling
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean true Include prediction status column in the output
includeProbabilities boolean true Include probabilities for all classes
includeProbabilitiesClasses [string] true maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any true The intake option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureIntake false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous BigQueryIntake false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous DataStageIntake false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous Catalog false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous DSS false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemIntake false none

xor

Name Type Required Restrictions Description
» anonymous GCPIntake false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPIntake false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCIntake false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileIntake false Stream CSV data chunks from local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Intake false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeIntake false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseIntake false Stream CSV data chunks from Azure Synapse

continued

Name Type Required Restrictions Description
maxExplanations integer true maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringAggregation MonitoringAggregation false Defines the aggregation policy for monitoring jobs.
monitoringBatchPrefix string,null false Name of the batch to create with this job
monitoringColumns MonitoringColumnsMapping false Column names mapping for monitoring
monitoringOutputSettings MonitoringOutputSettings false Output settings for monitoring jobs
numConcurrent integer false minimum: 1
Number of simultaneous requests to run against the prediction instance
outputSettings any false The output option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutput false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutput false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutput false none

xor

Name Type Required Restrictions Description
» anonymous GCPOutput false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPOutput false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JDBCOutput false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutput false Save CSV data chunks to local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Output false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutput false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutput false Save CSV data chunks to Azure Synapse in bulk

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchPredictionJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionThreshold number false maximum: 1
minimum: 0
Threshold is the point that sets the class boundary for a predicted value. The model classifies an observation below the threshold as FALSE, and an observation above the threshold as TRUE. In other words, DataRobot automatically assigns the positive class label to any prediction exceeding the threshold. This value can be set between 0.0 and 1.0.
predictionWarningEnabled boolean,null false Enable prediction warnings.
secondaryDatasetsConfigId string false Configuration id for secondary datasets to use when making a prediction.
skipDriftTracking boolean true Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsHistorical false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsTraining false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
passthroughColumnsSet all

BatchMonitoringJobDefinitionsCreate

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "enabled": true,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "name": "string",
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Properties

Name Type Required Restrictions Description
abortOnError boolean true Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchPredictionJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchPredictionJobCSVSettings true The CSV settings used for this job
deploymentId string true ID of deployment that the monitoring jobs is associated with.
disableRowLevelErrorHandling boolean true Skip row by row error handling
enabled boolean false If this job definition is enabled as a scheduled job. Optional if no schedule is supplied.
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean true Include prediction status column in the output
includeProbabilities boolean true Include probabilities for all classes
includeProbabilitiesClasses [string] true maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any true The intake option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureIntake false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous BigQueryIntake false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous DataStageIntake false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous Catalog false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous DSS false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemIntake false none

xor

Name Type Required Restrictions Description
» anonymous GCPIntake false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPIntake false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCIntake false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileIntake false Stream CSV data chunks from local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Intake false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeIntake false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseIntake false Stream CSV data chunks from Azure Synapse

continued

Name Type Required Restrictions Description
maxExplanations integer true maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringAggregation MonitoringAggregation false Defines the aggregation policy for monitoring jobs.
monitoringBatchPrefix string,null false Name of the batch to create with this job
monitoringColumns MonitoringColumnsMapping false Column names mapping for monitoring
monitoringOutputSettings MonitoringOutputSettings false Output settings for monitoring jobs
name string false maxLength: 100
minLength: 1
minLength: 1
A human-readable name for the definition, must be unique across organisations, if left out the backend will generate one for you.
numConcurrent integer false minimum: 1
Number of simultaneous requests to run against the prediction instance
outputSettings any false The output option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutput false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutput false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutput false none

xor

Name Type Required Restrictions Description
» anonymous GCPOutput false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPOutput false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JDBCOutput false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutput false Save CSV data chunks to local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Output false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutput false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutput false Save CSV data chunks to Azure Synapse in bulk

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchPredictionJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionThreshold number false maximum: 1
minimum: 0
Threshold is the point that sets the class boundary for a predicted value. The model classifies an observation below the threshold as FALSE, and an observation above the threshold as TRUE. In other words, DataRobot automatically assigns the positive class label to any prediction exceeding the threshold. This value can be set between 0.0 and 1.0.
predictionWarningEnabled boolean,null false Enable prediction warnings.
schedule Schedule false The scheduling information defining how often and when to execute this job to the Job Scheduling service. Optional if enabled = False.
secondaryDatasetsConfigId string false Configuration id for secondary datasets to use when making a prediction.
skipDriftTracking boolean true Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsHistorical false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsTraining false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
passthroughColumnsSet all

BatchMonitoringJobDefinitionsListResponse

{
  "count": 0,
  "data": [
    {
      "batchMonitoringJob": {
        "abortOnError": true,
        "batchJobType": "monitoring",
        "chunkSize": "auto",
        "columnNamesRemapping": {},
        "csvSettings": {
          "delimiter": ",",
          "encoding": "utf-8",
          "quotechar": "\""
        },
        "deploymentId": "string",
        "disableRowLevelErrorHandling": false,
        "explanationAlgorithm": "shap",
        "explanationClassNames": [
          "string"
        ],
        "explanationNumTopClasses": 1,
        "includePredictionStatus": false,
        "includeProbabilities": true,
        "includeProbabilitiesClasses": [],
        "intakeSettings": {
          "type": "localFile"
        },
        "maxExplanations": 0,
        "maxNgramExplanations": 0,
        "modelId": "string",
        "modelPackageId": "string",
        "monitoringAggregation": {
          "retentionPolicy": "samples",
          "retentionValue": 0
        },
        "monitoringBatchPrefix": "string",
        "monitoringColumns": {
          "actedUponColumn": "string",
          "actualsTimestampColumn": "string",
          "actualsValueColumn": "string",
          "associationIdColumn": "string",
          "customMetricId": "string",
          "customMetricTimestampColumn": "string",
          "customMetricTimestampFormat": "string",
          "customMetricValueColumn": "string",
          "monitoredStatusColumn": "string",
          "predictionsColumns": [
            {
              "className": "string",
              "columnName": "string"
            }
          ],
          "reportDrift": true,
          "reportPredictions": true,
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "monitoringOutputSettings": {
          "monitoredStatusColumn": "string",
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "numConcurrent": 0,
        "outputSettings": {
          "credentialId": "string",
          "format": "csv",
          "partitionColumns": [
            "string"
          ],
          "type": "azure",
          "url": "string"
        },
        "passthroughColumns": [
          "string"
        ],
        "passthroughColumnsSet": "all",
        "pinnedModelId": "string",
        "predictionInstance": {
          "apiKey": "string",
          "datarobotKey": "string",
          "hostName": "string",
          "sslEnabled": true
        },
        "predictionWarningEnabled": true,
        "redactedFields": [
          "string"
        ],
        "skipDriftTracking": false,
        "thresholdHigh": 0,
        "thresholdLow": 0,
        "timeseriesSettings": {
          "forecastPoint": "2019-08-24T14:15:22Z",
          "relaxKnownInAdvanceFeaturesCheck": false,
          "type": "forecast"
        }
      },
      "created": "2019-08-24T14:15:22Z",
      "createdBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      },
      "enabled": false,
      "id": "string",
      "lastFailedRunTime": "2019-08-24T14:15:22Z",
      "lastScheduledRunTime": "2019-08-24T14:15:22Z",
      "lastStartedJobStatus": "INITIALIZING",
      "lastStartedJobTime": "2019-08-24T14:15:22Z",
      "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
      "name": "string",
      "nextScheduledRunTime": "2019-08-24T14:15:22Z",
      "schedule": {
        "dayOfMonth": [
          "*"
        ],
        "dayOfWeek": [
          "*"
        ],
        "hour": [
          "*"
        ],
        "minute": [
          "*"
        ],
        "month": [
          "*"
        ]
      },
      "updated": "2019-08-24T14:15:22Z",
      "updatedBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      }
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Properties

Name Type Required Restrictions Description
count integer false Number of items returned on this page.
data [BatchMonitoringJobDefinitionsResponse] true maxItems: 10000
An array of scheduled jobs
next string,null(uri) true URL pointing to the next page (if null, there is no next page).
previous string,null(uri) true URL pointing to the previous page (if null, there is no previous page).
totalCount integer true The total number of items across all pages.

BatchMonitoringJobDefinitionsResponse

{
  "batchMonitoringJob": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 0,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "enabled": false,
  "id": "string",
  "lastFailedRunTime": "2019-08-24T14:15:22Z",
  "lastScheduledRunTime": "2019-08-24T14:15:22Z",
  "lastStartedJobStatus": "INITIALIZING",
  "lastStartedJobTime": "2019-08-24T14:15:22Z",
  "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
  "name": "string",
  "nextScheduledRunTime": "2019-08-24T14:15:22Z",
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "updated": "2019-08-24T14:15:22Z",
  "updatedBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  }
}

Properties

Name Type Required Restrictions Description
batchMonitoringJob BatchJobDefinitionsSpecResponse true The Batch Monitoring Job specification to be put on the queue in intervals
created string(date-time) true When was this job created
createdBy BatchJobCreatedBy true Who created this job
enabled boolean true If this job definition is enabled as a scheduled job.
id string true The ID of the Batch job definition
lastFailedRunTime string,null(date-time) false Last time this job had a failed run
lastScheduledRunTime string,null(date-time) false Last time this job was scheduled to run (though not guaranteed it actually ran at that time)
lastStartedJobStatus string,null true The status of the latest job launched to the queue (if any).
lastStartedJobTime string,null(date-time) true The last time (if any) a job was launched.
lastSuccessfulRunTime string,null(date-time) false Last time this job had a successful run
name string true A human-readable name for the definition, must be unique across organisations
nextScheduledRunTime string,null(date-time) false Next time this job is scheduled to run
schedule Schedule false The scheduling information defining how often and when to execute this job to the Job Scheduling service. Optional if enabled = False.
updated string(date-time) true When was this job last updated
updatedBy BatchJobCreatedBy true Who created this job

Enumerated Values

Property Value
lastStartedJobStatus [INITIALIZING, RUNNING, COMPLETED, ABORTED, FAILED]

BatchMonitoringJobDefinitionsUpdate

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "enabled": true,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "name": "string",
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Properties

Name Type Required Restrictions Description
abortOnError boolean false Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchPredictionJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchPredictionJobCSVSettings false The CSV settings used for this job
deploymentId string false ID of deployment which is used in job for processing predictions dataset
disableRowLevelErrorHandling boolean false Skip row by row error handling
enabled boolean false If this job definition is enabled as a scheduled job. Optional if no schedule is supplied.
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean false Include prediction status column in the output
includeProbabilities boolean false Include probabilities for all classes
includeProbabilitiesClasses [string] false maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any false The intake option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureIntake false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous BigQueryIntake false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous DataStageIntake false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous Catalog false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous DSS false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemIntake false none

xor

Name Type Required Restrictions Description
» anonymous GCPIntake false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPIntake false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCIntake false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileIntake false Stream CSV data chunks from local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Intake false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeIntake false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseIntake false Stream CSV data chunks from Azure Synapse

continued

Name Type Required Restrictions Description
maxExplanations integer false maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringAggregation MonitoringAggregation false Defines the aggregation policy for monitoring jobs.
monitoringBatchPrefix string,null false Name of the batch to create with this job
monitoringColumns MonitoringColumnsMapping false Column names mapping for monitoring
monitoringOutputSettings MonitoringOutputSettings false Output settings for monitoring jobs
name string false maxLength: 100
minLength: 1
minLength: 1
A human-readable name for the definition, must be unique across organisations, if left out the backend will generate one for you.
numConcurrent integer false minimum: 1
Number of simultaneous requests to run against the prediction instance
outputSettings any false The output option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutput false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutput false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutput false none

xor

Name Type Required Restrictions Description
» anonymous GCPOutput false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPOutput false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JDBCOutput false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutput false Save CSV data chunks to local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Output false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutput false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutput false Save CSV data chunks to Azure Synapse in bulk

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchPredictionJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionThreshold number false maximum: 1
minimum: 0
Threshold is the point that sets the class boundary for a predicted value. The model classifies an observation below the threshold as FALSE, and an observation above the threshold as TRUE. In other words, DataRobot automatically assigns the positive class label to any prediction exceeding the threshold. This value can be set between 0.0 and 1.0.
predictionWarningEnabled boolean,null false Enable prediction warnings.
schedule Schedule false The scheduling information defining how often and when to execute this job to the Job Scheduling service. Optional if enabled = False.
secondaryDatasetsConfigId string false Configuration id for secondary datasets to use when making a prediction.
skipDriftTracking boolean false Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsHistorical false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsTraining false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
passthroughColumnsSet all

BatchPredictionJobCSVSettings

{
  "delimiter": ",",
  "encoding": "utf-8",
  "quotechar": "\""
}

The CSV settings used for this job

Properties

Name Type Required Restrictions Description
delimiter any true CSV fields are delimited by this character. Use the string "tab" to denote TSV (TAB separated values).

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous string false maxLength: 1
minLength: 1
minLength: 1
none

continued

Name Type Required Restrictions Description
encoding string true The encoding to be used for intake and output. For example (but not limited to): "shift_jis", "latin_1" or "mskanji".
quotechar string true maxLength: 1
minLength: 1
minLength: 1
Fields containing the delimiter or newlines must be quoted using this character.

Enumerated Values

Property Value
anonymous tab

BatchPredictionJobPredictionInstance

{
  "apiKey": "string",
  "datarobotKey": "string",
  "hostName": "string",
  "sslEnabled": true
}

Override the default prediction instance from the deployment when scoring this job.

Properties

Name Type Required Restrictions Description
apiKey string false By default, prediction requests will use the API key of the user that created the job. This allows you to make requests on behalf of other users.
datarobotKey string false If running a job against a prediction instance in the Managed AI Cloud, you must provide the organization level DataRobot-Key.
hostName string true Override the default host name of the deployment with this.
sslEnabled boolean true Use SSL (HTTPS) when communicating with the overriden prediction server.

BatchPredictionJobRemapping

{
  "inputName": "string",
  "outputName": "string"
}

Properties

Name Type Required Restrictions Description
inputName string true Rename column with this name
outputName string,null true Rename column to this name (leave as null to remove from the output)

BatchPredictionJobTimeSeriesSettingsForecast

{
  "forecastPoint": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "forecast"
}

Properties

Name Type Required Restrictions Description
forecastPoint string(date-time) false Used for forecast predictions in order to override the inferred forecast point from the dataset.
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Forecast mode makes predictions using forecastPoint or rows in the dataset without target.

Enumerated Values

Property Value
type forecast

BatchPredictionJobTimeSeriesSettingsHistorical

{
  "predictionsEndDate": "2019-08-24T14:15:22Z",
  "predictionsStartDate": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "historical"
}

Properties

Name Type Required Restrictions Description
predictionsEndDate string(date-time) false Used for historical predictions in order to override date to which predictions should be calculated. By default value will be inferred automatically from the dataset.
predictionsStartDate string(date-time) false Used for historical predictions in order to override date from which predictions should be calculated. By default value will be inferred automatically from the dataset.
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Historical mode enables bulk predictions which calculates predictions for all possible forecast points and forecast distances in the dataset within the predictionsStartDate/predictionsEndDate range.

Enumerated Values

Property Value
type historical

BatchPredictionJobTimeSeriesSettingsTraining

{
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "training"
}

Properties

Name Type Required Restrictions Description
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Forecast mode used for making predictions on subsets of training data.

Enumerated Values

Property Value
type training

BigQueryDataStreamer

{
  "bucket": "string",
  "credentialId": "string",
  "dataset": "string",
  "table": "string",
  "type": "bigquery"
}

Stream CSV data chunks from Big Query using GCS

Properties

Name Type Required Restrictions Description
bucket string true The name of gcs bucket for data export
credentialId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false The ID of the GCP credentials

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataset string true The name of the specified big query dataset to read input data from
table string true The name of the specified big query table to read input data from
type string true Type name for this intake type

Enumerated Values

Property Value
anonymous [redacted]
type bigquery

BigQueryIntake

{
  "bucket": "string",
  "credentialId": "string",
  "dataset": "string",
  "table": "string",
  "type": "bigquery"
}

Stream CSV data chunks from Big Query using GCS

Properties

Name Type Required Restrictions Description
bucket string true The name of gcs bucket for data export
credentialId string true The ID of the GCP credentials
dataset string true The name of the specified big query dataset to read input data from
table string true The name of the specified big query table to read input data from
type string true Type name for this intake type

Enumerated Values

Property Value
type bigquery

BigQueryOutput

{
  "bucket": "string",
  "credentialId": "string",
  "dataset": "string",
  "table": "string",
  "type": "bigquery"
}

Save CSV data chunks to Google BigQuery in bulk

Properties

Name Type Required Restrictions Description
bucket string true The name of gcs bucket for data loading
credentialId string true The ID of the GCP credentials
dataset string true The name of the specified big query dataset to write data back
table string true The name of the specified big query table to write data back
type string true Type name for this output type

Enumerated Values

Property Value
type bigquery

BigQueryOutputAdaptor

{
  "bucket": "string",
  "credentialId": "string",
  "dataset": "string",
  "table": "string",
  "type": "bigquery"
}

Save CSV data chunks to Google BigQuery in bulk

Properties

Name Type Required Restrictions Description
bucket string true The name of gcs bucket for data loading
credentialId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false The ID of the GCP credentials

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataset string true The name of the specified big query dataset to write data back
table string true The name of the specified big query table to write data back
type string true Type name for this output type

Enumerated Values

Property Value
anonymous [redacted]
type bigquery

Catalog

{
  "datasetId": "string",
  "datasetVersionId": "string",
  "type": "dataset"
}

Stream CSV data chunks from AI catalog dataset

Properties

Name Type Required Restrictions Description
datasetId string true The ID of the AI catalog dataset
datasetVersionId string false The ID of the AI catalog dataset version
type string true Type name for this intake type

Enumerated Values

Property Value
type dataset

CatalogDataStreamer

{
  "datasetId": "string",
  "datasetVersionId": "string",
  "type": "dataset"
}

Stream CSV data chunks from AI catalog dataset

Properties

Name Type Required Restrictions Description
datasetId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false The ID of the AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
datasetVersionId string false The ID of the AI catalog dataset version
type string true Type name for this intake type

Enumerated Values

Property Value
anonymous [redacted]
type dataset

DSS

{
  "datasetId": "string",
  "partition": "holdout",
  "projectId": "string",
  "type": "dss"
}

Stream CSV data chunks from DSS dataset

Properties

Name Type Required Restrictions Description
datasetId string false The ID of the dataset
partition string,null false Partition used to predict
projectId string true The ID of the project
type string true Type name for this intake type

Enumerated Values

Property Value
partition [holdout, validation, allBacktests, null]
type dss

DSSDataStreamer

{
  "datasetId": "string",
  "partition": "holdout",
  "projectId": "string",
  "type": "dss"
}

Stream CSV data chunks from DSS dataset

Properties

Name Type Required Restrictions Description
datasetId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false The ID of the dataset

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
partition string,null false Partition used to predict
projectId string true The ID of the project
type string true Type name for this intake type

Enumerated Values

Property Value
anonymous [redacted]
partition [holdout, validation, allBacktests, null]
type dss

DataStageDataStreamer

{
  "dataStageId": "string",
  "type": "dataStage"
}

Stream CSV data chunks from data stage storage

Properties

Name Type Required Restrictions Description
dataStageId string true The ID of the data stage
type string true Type name for this intake type

Enumerated Values

Property Value
type dataStage

DataStageIntake

{
  "dataStageId": "string",
  "type": "dataStage"
}

Stream CSV data chunks from data stage storage

Properties

Name Type Required Restrictions Description
dataStageId string true The ID of the data stage
type string true Type name for this intake type

Enumerated Values

Property Value
type dataStage

FileSystemDataStreamer

{
  "path": "string",
  "type": "filesystem"
}

Properties

Name Type Required Restrictions Description
path string true Path to data on host filesystem
type string true Type name for this intake type

Enumerated Values

Property Value
type filesystem

FileSystemIntake

{
  "path": "string",
  "type": "filesystem"
}

Properties

Name Type Required Restrictions Description
path string true Path to data on host filesystem
type string true Type name for this intake type

Enumerated Values

Property Value
type filesystem

FileSystemOutput

{
  "path": "string",
  "type": "filesystem"
}

Properties

Name Type Required Restrictions Description
path string true Path to results on host filesystem
type string true Type name for this output type

Enumerated Values

Property Value
type filesystem

FileSystemOutputAdaptor

{
  "path": "string",
  "type": "filesystem"
}

Properties

Name Type Required Restrictions Description
path string true Path to results on host filesystem
type string true Type name for this output type

Enumerated Values

Property Value
type filesystem

GCPDataStreamer

{
  "credentialId": "string",
  "format": "csv",
  "type": "gcp",
  "url": "string"
}

Stream CSV data chunks from Google Storage

Properties

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false Use the specified credential to access the url

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
format string false Type of input file format
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
anonymous [redacted]
format [csv, parquet]
type gcp

GCPIntake

{
  "credentialId": "string",
  "format": "csv",
  "type": "gcp",
  "url": "string"
}

Stream CSV data chunks from Google Storage

Properties

Name Type Required Restrictions Description
credentialId string,null false Use the specified credential to access the url
format string false Type of input file format
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
format [csv, parquet]
type gcp

GCPOutput

{
  "credentialId": "string",
  "format": "csv",
  "partitionColumns": [
    "string"
  ],
  "type": "gcp",
  "url": "string"
}

Save CSV data chunks to Google Storage

Properties

Name Type Required Restrictions Description
credentialId string,null false Use the specified credential to access the url
format string false Type of input file format
partitionColumns [string] false maxItems: 100
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/").
type string true Type name for this output type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
format [csv, parquet]
type gcp

GCPOutputAdaptor

{
  "credentialId": "string",
  "format": "csv",
  "partitionColumns": [
    "string"
  ],
  "type": "gcp",
  "url": "string"
}

Save CSV data chunks to Google Storage

Properties

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false Use the specified credential to access the url

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
format string false Type of input file format
partitionColumns [string] false maxItems: 100
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/").
type string true Type name for this output type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
anonymous [redacted]
format [csv, parquet]
type gcp

HTTPDataStreamer

{
  "type": "http",
  "url": "string"
}

Stream CSV data chunks from HTTP

Properties

Name Type Required Restrictions Description
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
type http

HTTPIntake

{
  "type": "http",
  "url": "string"
}

Stream CSV data chunks from HTTP

Properties

Name Type Required Restrictions Description
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
type http

HTTPOutput

{
  "headers": {},
  "method": "POST",
  "type": "http",
  "url": "string"
}

Save CSV data chunks to HTTP data endpoint

Properties

Name Type Required Restrictions Description
headers object false Extra headers to send with the request
method string true Method to use when saving the CSV file
type string true Type name for this output type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
method [POST, PUT]
type http

HttpOutputAdaptor

{
  "headers": {},
  "method": "POST",
  "type": "http",
  "url": "string"
}

Save CSV data chunks to HTTP data endpoint

Properties

Name Type Required Restrictions Description
headers object false Extra headers to send with the request
method string true Method to use when saving the CSV file
type string true Type name for this output type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
method [POST, PUT]
type http

JDBCDataStreamer

{
  "catalog": "string",
  "credentialId": "string",
  "dataStoreId": "string",
  "fetchSize": 1,
  "query": "string",
  "schema": "string",
  "table": "string",
  "type": "jdbc"
}

Stream CSV data chunks from JDBC

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to read input data from.
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false The ID of the credential holding information about a user with read access to the JDBC data source.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataStoreId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false ID of the data store to connect to

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
fetchSize integer false maximum: 1000000
minimum: 1
A user specified fetch size. Changing it can be used to balance throughput and memory usage. Deprecated and ignored since v2.21.
query string false A self-supplied SELECT statement of the dataset you wish to score. Helpful for supplying a more fine-grained selection of data not achievable through specification of "table" and/or "schema" parameters exclusively.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
schema string false The name of the specified database schema to read input data from.
table string false The name of the specified database table to read input data from.
type string true Type name for this intake type

Enumerated Values

Property Value
anonymous [redacted]
anonymous [redacted]
type jdbc

JDBCIntake

{
  "catalog": "string",
  "credentialId": "string",
  "dataStoreId": "string",
  "fetchSize": 1,
  "query": "string",
  "schema": "string",
  "table": "string",
  "type": "jdbc"
}

Stream CSV data chunks from JDBC

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to read input data from.
credentialId string,null false The ID of the credential holding information about a user with read access to the JDBC data source.
dataStoreId string true ID of the data store to connect to
fetchSize integer false maximum: 1000000
minimum: 1
A user specified fetch size. Changing it can be used to balance throughput and memory usage. Deprecated and ignored since v2.21.
query string false A self-supplied SELECT statement of the dataset you wish to score. Helpful for supplying a more fine-grained selection of data not achievable through specification of "table" and/or "schema" parameters exclusively.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
schema string false The name of the specified database schema to read input data from.
table string false The name of the specified database table to read input data from.
type string true Type name for this intake type

Enumerated Values

Property Value
type jdbc

JDBCOutput

{
  "catalog": "string",
  "commitInterval": 600,
  "createTableIfNotExists": false,
  "credentialId": "string",
  "dataStoreId": "string",
  "schema": "string",
  "statementType": "createTable",
  "table": "string",
  "type": "jdbc",
  "updateColumns": [
    "string"
  ],
  "whereColumns": [
    "string"
  ]
}

Save CSV data chunks via JDBC

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to write output data to.
commitInterval integer false maximum: 86400
minimum: 0
Defines a time interval in seconds between each commit is done to the JDBC source. If set to 0, the batch prediction operation will write the entire job before committing.
createTableIfNotExists boolean false Attempt to create the table first if no existing one is detected, before writing data with the strategy defined in the statementType parameter.
credentialId string,null false The ID of the credential holding information about a user with write access to the JDBC data source.
dataStoreId string true ID of the data store to connect to
schema string false The name of the specified database schema to write the results to.
statementType string true The statement type to use when writing the results. Deprecation Warning: Use of create_table is now discouraged. Use one of the other possibilities along with the parameter createTableIfNotExists set to true.
table string true The name of the specified database table to write the results to.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
type string true Type name for this intake type
updateColumns [string] false maxItems: 100
The column names to be updated if statementType is set to either update or upsert.
whereColumns [string] false maxItems: 100
The column names to be used in the where clause if statementType is set to update or upsert.

Enumerated Values

Property Value
statementType [createTable, create_table, insert, insertUpdate, insert_update, update]
type jdbc

JdbcOutputAdaptor

{
  "catalog": "string",
  "commitInterval": 600,
  "createTableIfNotExists": false,
  "credentialId": "string",
  "dataStoreId": "string",
  "schema": "string",
  "statementType": "createTable",
  "table": "string",
  "type": "jdbc",
  "updateColumns": [
    "string"
  ],
  "whereColumns": [
    "string"
  ]
}

Save CSV data chunks via JDBC

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to write output data to.
commitInterval integer false maximum: 86400
minimum: 0
Defines a time interval in seconds between each commit is done to the JDBC source. If set to 0, the batch prediction operation will write the entire job before committing.
createTableIfNotExists boolean false Attempt to create the table first if no existing one is detected, before writing data with the strategy defined in the statementType parameter.
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false The ID of the credential holding information about a user with write access to the JDBC data source.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataStoreId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false ID of the data store to connect to

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
schema string false The name of the specified database schema to write the results to.
statementType string true The statement type to use when writing the results. Deprecation Warning: Use of create_table is now discouraged. Use one of the other possibilities along with the parameter createTableIfNotExists set to true.
table string true The name of the specified database table to write the results to.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
type string true Type name for this intake type
updateColumns [string] false maxItems: 100
The column names to be updated if statementType is set to either update or upsert.
whereColumns [string] false maxItems: 100
The column names to be used in the where clause if statementType is set to update or upsert.

Enumerated Values

Property Value
anonymous [redacted]
anonymous [redacted]
statementType [createTable, create_table, insert, insertUpdate, insert_update, update]
type jdbc

LocalFileDataStreamer

{
  "async": true,
  "multipart": true,
  "type": "local_file"
}

Stream CSV data chunks from local file storage

Properties

Name Type Required Restrictions Description
async boolean,null false The default behavior (async: true) will still submit the job to the queue and start processing as soon as the upload is started.Setting it to false will postpone submitting the job to the queue until all data has been uploaded.This is helpful if the user is on a bad connection and bottlednecked by the upload speed. Instead of blocking the queue this will allow others to submit to the queue until the upload has finished.
multipart boolean false specify if the data will be uploaded in multiple parts instead of a single file
type string true Type name for this intake type

Enumerated Values

Property Value
type [local_file, localFile]

LocalFileIntake

{
  "async": true,
  "multipart": true,
  "type": "local_file"
}

Stream CSV data chunks from local file storage

Properties

Name Type Required Restrictions Description
async boolean,null false The default behavior (async: true) will still submit the job to the queue and start processing as soon as the upload is started.Setting it to false will postpone submitting the job to the queue until all data has been uploaded.This is helpful if the user is on a bad connection and bottlednecked by the upload speed. Instead of blocking the queue this will allow others to submit to the queue until the upload has finished.
multipart boolean false specify if the data will be uploaded in multiple parts instead of a single file
type string true Type name for this intake type

Enumerated Values

Property Value
type [local_file, localFile]

LocalFileOutput

{
  "type": "local_file"
}

Save CSV data chunks to local file storage

Properties

Name Type Required Restrictions Description
type string true Type name for this output type

Enumerated Values

Property Value
type [local_file, localFile]

LocalFileOutputAdaptor

{
  "type": "local_file"
}

Save CSV data chunks to local file storage

Properties

Name Type Required Restrictions Description
type string true Type name for this output type

Enumerated Values

Property Value
type [local_file, localFile]

MonitoringAggregation

{
  "retentionPolicy": "samples",
  "retentionValue": 0
}

Defines the aggregation policy for monitoring jobs.

Properties

Name Type Required Restrictions Description
retentionPolicy string false Monitoring jobs retention policy for aggregation.
retentionValue integer false Amount/percentage of samples to retain.

Enumerated Values

Property Value
retentionPolicy [samples, percentage]

MonitoringColumnsMapping

{
  "actedUponColumn": "string",
  "actualsTimestampColumn": "string",
  "actualsValueColumn": "string",
  "associationIdColumn": "string",
  "customMetricId": "string",
  "customMetricTimestampColumn": "string",
  "customMetricTimestampFormat": "string",
  "customMetricValueColumn": "string",
  "monitoredStatusColumn": "string",
  "predictionsColumns": [
    {
      "className": "string",
      "columnName": "string"
    }
  ],
  "reportDrift": true,
  "reportPredictions": true,
  "uniqueRowIdentifierColumns": [
    "string"
  ]
}

Column names mapping for monitoring

Properties

Name Type Required Restrictions Description
actedUponColumn string false Name of column that contains value for acted_on.
actualsTimestampColumn string false Name of column that contains actual timestamps.
actualsValueColumn string false Name of column that contains actuals value.
associationIdColumn string false Name of column that contains association Id.
customMetricId string false Id of custom metric to process values for.
customMetricTimestampColumn string false Name of column that contains custom metric values timestamps.
customMetricTimestampFormat string false Format of timestamps from customMetricTimestampColumn.
customMetricValueColumn string false Name of column that contains values for custom metric.
monitoredStatusColumn string false Column name used to mark monitored rows.
predictionsColumns any false Name of the column(s) which contain prediction values.

oneOf

Name Type Required Restrictions Description
» anonymous [PredictionColumMap] false maxItems: 100
Map containing column name(s) and class name(s) for multiclass problem.

xor

Name Type Required Restrictions Description
» anonymous string false Column name that contains the prediction for regressions problem.

continued

Name Type Required Restrictions Description
reportDrift boolean false True to report drift, False otherwise.
reportPredictions boolean false True to report prediction, False otherwise.
uniqueRowIdentifierColumns [string] false maxItems: 100
Column(s) name of unique row identifiers.

MonitoringOutputSettings

{
  "monitoredStatusColumn": "string",
  "uniqueRowIdentifierColumns": [
    "string"
  ]
}

Output settings for monitoring jobs

Properties

Name Type Required Restrictions Description
monitoredStatusColumn string true Column name used to mark monitored rows.
uniqueRowIdentifierColumns [string] true maxItems: 100
Column(s) name of unique row identifiers.

PredictionColumMap

{
  "className": "string",
  "columnName": "string"
}

Properties

Name Type Required Restrictions Description
className string true Class name.
columnName string true Column name that contains the prediction for a specific class.

S3DataStreamer

{
  "credentialId": "string",
  "endpointUrl": "string",
  "format": "csv",
  "type": "s3",
  "url": "string"
}

Stream CSV data chunks from Amazon Cloud Storage S3

Properties

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false Use the specified credential to access the url

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
endpointUrl string(url) false Endpoint URL for the S3 connection (omit to use the default)
format string false Type of input file format
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
anonymous [redacted]
format [csv, parquet]
type s3

S3Intake

{
  "credentialId": "string",
  "endpointUrl": "string",
  "format": "csv",
  "type": "s3",
  "url": "string"
}

Stream CSV data chunks from Amazon Cloud Storage S3

Properties

Name Type Required Restrictions Description
credentialId string,null false Use the specified credential to access the url
endpointUrl string(url) false Endpoint URL for the S3 connection (omit to use the default)
format string false Type of input file format
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
format [csv, parquet]
type s3

S3Output

{
  "credentialId": "string",
  "endpointUrl": "string",
  "format": "csv",
  "partitionColumns": [
    "string"
  ],
  "serverSideEncryption": {
    "algorithm": "string",
    "customerAlgorithm": "string",
    "customerKey": "string",
    "kmsEncryptionContext": "string",
    "kmsKeyId": "string"
  },
  "type": "s3",
  "url": "string"
}

Saves CSV data chunks to Amazon Cloud Storage S3

Properties

Name Type Required Restrictions Description
credentialId string,null false Use the specified credential to access the url
endpointUrl string(url) false Endpoint URL for the S3 connection (omit to use the default)
format string false Type of output file format
partitionColumns [string] false maxItems: 100
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/").
serverSideEncryption ServerSideEncryption false Configure Server-Side Encryption for S3 output
type string true Type name for this output type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
format [csv, parquet]
type s3

S3OutputAdaptor

{
  "credentialId": "string",
  "endpointUrl": "string",
  "format": "csv",
  "partitionColumns": [
    "string"
  ],
  "serverSideEncryption": {
    "algorithm": "string",
    "customerAlgorithm": "string",
    "customerKey": "string",
    "kmsEncryptionContext": "string",
    "kmsKeyId": "string"
  },
  "type": "s3",
  "url": "string"
}

Saves CSV data chunks to Amazon Cloud Storage S3

Properties

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false Use the specified credential to access the url

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
endpointUrl string(url) false Endpoint URL for the S3 connection (omit to use the default)
format string false Type of output file format
partitionColumns [string] false maxItems: 100
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/").
serverSideEncryption ServerSideEncryption false Configure Server-Side Encryption for S3 output
type string true Type name for this output type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
anonymous [redacted]
format [csv, parquet]
type s3

Schedule

{
  "dayOfMonth": [
    "*"
  ],
  "dayOfWeek": [
    "*"
  ],
  "hour": [
    "*"
  ],
  "minute": [
    "*"
  ],
  "month": [
    "*"
  ]
}

The scheduling information defining how often and when to execute this job to the Job Scheduling service. Optional if enabled = False.

Properties

Name Type Required Restrictions Description
dayOfMonth [number,string] true maxItems: 31
The date(s) of the month that the job will run. Allowed values are either [1 ... 31] or ["*"] for all days of the month. This field is additive with dayOfWeek, meaning the job will run both on the date(s) defined in this field and the day specified by dayOfWeek (for example, dates 1st, 2nd, 3rd, plus every Tuesday). If dayOfMonth is set to ["*"] and dayOfWeek is defined, the scheduler will trigger on every day of the month that matches dayOfWeek (for example, Tuesday the 2nd, 9th, 16th, 23rd, 30th). Invalid dates such as February 31st are ignored.
dayOfWeek [number,string] true maxItems: 7
The day(s) of the week that the job will run. Allowed values are [0 .. 6], where (Sunday=0), or ["*"], for all days of the week. Strings, either 3-letter abbreviations or the full name of the day, can be used interchangeably (e.g., "sunday", "Sunday", "sun", or "Sun", all map to [0]. This field is additive with dayOfMonth, meaning the job will run both on the date specified by dayOfMonth and the day defined in this field.
hour [number,string] true maxItems: 24
The hour(s) of the day that the job will run. Allowed values are either ["*"] meaning every hour of the day or [0 ... 23].
minute [number,string] true maxItems: 60
The minute(s) of the day that the job will run. Allowed values are either ["*"] meaning every minute of the day or[0 ... 59].
month [number,string] true maxItems: 12
The month(s) of the year that the job will run. Allowed values are either [1 ... 12] or ["*"] for all months of the year. Strings, either 3-letter abbreviations or the full name of the month, can be used interchangeably (e.g., "jan" or "october"). Months that are not compatible with dayOfMonth are ignored, for example {"dayOfMonth": [31], "month":["feb"]}.

ServerSideEncryption

{
  "algorithm": "string",
  "customerAlgorithm": "string",
  "customerKey": "string",
  "kmsEncryptionContext": "string",
  "kmsKeyId": "string"
}

Configure Server-Side Encryption for S3 output

Properties

Name Type Required Restrictions Description
algorithm string false The server-side encryption algorithm used when storing this object in Amazon S3 (for example, AES256, aws:kms).
customerAlgorithm string false Specifies the algorithm to use to when encrypting the object (for example, AES256).
customerKey string false Specifies the customer-provided encryption key for Amazon S3 to use in encrypting data. This value is used to store the object and then it is discarded; Amazon S3 does not store the encryption key. The key must be appropriate for use with the algorithm specified in customerAlgorithm. The key must be sent as an base64 encoded string.
kmsEncryptionContext string false Specifies the Amazon Web Services KMS Encryption Context to use for object encryption. The value of this header is a base64-encoded UTF-8 string holding JSON with the encryption context key-value pairs.
kmsKeyId string false Specifies the ID of the symmetric customer managed key to use for object encryption.

SnowflakeDataStreamer

{
  "catalog": "string",
  "cloudStorageCredentialId": "string",
  "cloudStorageType": "azure",
  "credentialId": "string",
  "dataStoreId": "string",
  "externalStage": "string",
  "query": "string",
  "schema": "string",
  "table": "string",
  "type": "snowflake"
}

Stream CSV data chunks from Snowflake

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to read input data from.
cloudStorageCredentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false The ID of the credential holding information about a user with read access to the cloud storage.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
cloudStorageType string false Type name for cloud storage
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false The ID of the credential holding information about a user with read access to the Snowflake data source.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataStoreId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false ID of the data store to connect to

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
externalStage string true External storage
query string false A self-supplied SELECT statement of the dataset you wish to score. Helpful for supplying a more fine-grained selection of data not achievable through specification of "table" and/or "schema" parameters exclusively.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
schema string false The name of the specified database schema to read input data from.
table string false The name of the specified database table to read input data from.
type string true Type name for this intake type

Enumerated Values

Property Value
anonymous [redacted]
cloudStorageType [azure, gcp, s3]
anonymous [redacted]
anonymous [redacted]
type snowflake

SnowflakeIntake

{
  "catalog": "string",
  "cloudStorageCredentialId": "string",
  "cloudStorageType": "azure",
  "credentialId": "string",
  "dataStoreId": "string",
  "externalStage": "string",
  "query": "string",
  "schema": "string",
  "table": "string",
  "type": "snowflake"
}

Stream CSV data chunks from Snowflake

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to read input data from.
cloudStorageCredentialId string,null false The ID of the credential holding information about a user with read access to the cloud storage.
cloudStorageType string false Type name for cloud storage
credentialId string,null false The ID of the credential holding information about a user with read access to the Snowflake data source.
dataStoreId string true ID of the data store to connect to
externalStage string true External storage
query string false A self-supplied SELECT statement of the dataset you wish to score. Helpful for supplying a more fine-grained selection of data not achievable through specification of "table" and/or "schema" parameters exclusively.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
schema string false The name of the specified database schema to read input data from.
table string false The name of the specified database table to read input data from.
type string true Type name for this intake type

Enumerated Values

Property Value
cloudStorageType [azure, gcp, s3]
type snowflake

SnowflakeOutput

{
  "catalog": "string",
  "cloudStorageCredentialId": "string",
  "cloudStorageType": "azure",
  "createTableIfNotExists": false,
  "credentialId": "string",
  "dataStoreId": "string",
  "externalStage": "string",
  "schema": "string",
  "statementType": "insert",
  "table": "string",
  "type": "snowflake"
}

Save CSV data chunks to Snowflake in bulk

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to write output data to.
cloudStorageCredentialId string,null false The ID of the credential holding information about a user with write access to the cloud storage.
cloudStorageType string false Type name for cloud storage
createTableIfNotExists boolean false Attempt to create the table first if no existing one is detected, before writing data with the strategy defined in the statementType parameter.
credentialId string,null false The ID of the credential holding information about a user with write access to the Snowflake data source.
dataStoreId string true ID of the data store to connect to
externalStage string true External storage
schema string false The name of the specified database schema to write results to.
statementType string true The statement type to use when writing the results.
table string true The name of the specified database table to write results to.
type string true Type name for this output type

Enumerated Values

Property Value
cloudStorageType [azure, gcp, s3]
statementType [insert, create_table, createTable]
type snowflake

SnowflakeOutputAdaptor

{
  "catalog": "string",
  "cloudStorageCredentialId": "string",
  "cloudStorageType": "azure",
  "createTableIfNotExists": false,
  "credentialId": "string",
  "dataStoreId": "string",
  "externalStage": "string",
  "schema": "string",
  "statementType": "insert",
  "table": "string",
  "type": "snowflake"
}

Save CSV data chunks to Snowflake in bulk

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to write output data to.
cloudStorageCredentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false The ID of the credential holding information about a user with write access to the cloud storage.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
cloudStorageType string false Type name for cloud storage
createTableIfNotExists boolean false Attempt to create the table first if no existing one is detected, before writing data with the strategy defined in the statementType parameter.
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false The ID of the credential holding information about a user with write access to the Snowflake data source.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataStoreId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false ID of the data store to connect to

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
externalStage string true External storage
schema string false The name of the specified database schema to write results to.
statementType string true The statement type to use when writing the results.
table string true The name of the specified database table to write results to.
type string true Type name for this output type

Enumerated Values

Property Value
anonymous [redacted]
cloudStorageType [azure, gcp, s3]
anonymous [redacted]
anonymous [redacted]
statementType [insert, create_table, createTable]
type snowflake

SynapseDataStreamer

{
  "cloudStorageCredentialId": "string",
  "credentialId": "string",
  "dataStoreId": "string",
  "externalDataSource": "string",
  "query": "string",
  "schema": "string",
  "table": "string",
  "type": "synapse"
}

Stream CSV data chunks from Azure Synapse

Properties

Name Type Required Restrictions Description
cloudStorageCredentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false The ID of the Azure credential holding information about a user with read access to the cloud storage.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false The ID of the credential holding information about a user with read access to the JDBC data source.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataStoreId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false ID of the data store to connect to

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
externalDataSource string true External datasource name
query string false A self-supplied SELECT statement of the dataset you wish to score. Helpful for supplying a more fine-grained selection of data not achievable through specification of "table" and/or "schema" parameters exclusively.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
schema string false The name of the specified database schema to read input data from.
table string false The name of the specified database table to read input data from.
type string true Type name for this intake type

Enumerated Values

Property Value
anonymous [redacted]
anonymous [redacted]
anonymous [redacted]
type synapse

SynapseIntake

{
  "cloudStorageCredentialId": "string",
  "credentialId": "string",
  "dataStoreId": "string",
  "externalDataSource": "string",
  "query": "string",
  "schema": "string",
  "table": "string",
  "type": "synapse"
}

Stream CSV data chunks from Azure Synapse

Properties

Name Type Required Restrictions Description
cloudStorageCredentialId string,null false The ID of the Azure credential holding information about a user with read access to the cloud storage.
credentialId string,null false The ID of the credential holding information about a user with read access to the JDBC data source.
dataStoreId string true ID of the data store to connect to
externalDataSource string true External datasource name
query string false A self-supplied SELECT statement of the dataset you wish to score. Helpful for supplying a more fine-grained selection of data not achievable through specification of "table" and/or "schema" parameters exclusively.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
schema string false The name of the specified database schema to read input data from.
table string false The name of the specified database table to read input data from.
type string true Type name for this intake type

Enumerated Values

Property Value
type synapse

SynapseOutput

{
  "cloudStorageCredentialId": "string",
  "createTableIfNotExists": false,
  "credentialId": "string",
  "dataStoreId": "string",
  "externalDataSource": "string",
  "schema": "string",
  "statementType": "insert",
  "table": "string",
  "type": "synapse"
}

Save CSV data chunks to Azure Synapse in bulk

Properties

Name Type Required Restrictions Description
cloudStorageCredentialId string,null false The ID of the credential holding information about a user with write access to the cloud storage.
createTableIfNotExists boolean false Attempt to create the table first if no existing one is detected, before writing data with the strategy defined in the statementType parameter.
credentialId string,null false The ID of the credential holding information about a user with write access to the JDBC data source.
dataStoreId string true ID of the data store to connect to
externalDataSource string true External data source name
schema string false The name of the specified database schema to write results to.
statementType string true The statement type to use when writing the results.
table string true The name of the specified database table to write results to.
type string true Type name for this output type

Enumerated Values

Property Value
statementType [insert, create_table, createTable]
type synapse

SynapseOutputAdaptor

{
  "cloudStorageCredentialId": "string",
  "createTableIfNotExists": false,
  "credentialId": "string",
  "dataStoreId": "string",
  "externalDataSource": "string",
  "schema": "string",
  "statementType": "insert",
  "table": "string",
  "type": "synapse"
}

Save CSV data chunks to Azure Synapse in bulk

Properties

Name Type Required Restrictions Description
cloudStorageCredentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false The ID of the credential holding information about a user with write access to the cloud storage.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
createTableIfNotExists boolean false Attempt to create the table first if no existing one is detected, before writing data with the strategy defined in the statementType parameter.
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string,null false The ID of the credential holding information about a user with write access to the JDBC data source.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataStoreId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false ID of the data store to connect to

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
externalDataSource string true External data source name
schema string false The name of the specified database schema to write results to.
statementType string true The statement type to use when writing the results.
table string true The name of the specified database table to write results to.
type string true Type name for this output type

Enumerated Values

Property Value
anonymous [redacted]
anonymous [redacted]
anonymous [redacted]
statementType [insert, create_table, createTable]
type synapse

Updated March 25, 2025