Skip to content

On-premise users: click in-app to access the full platform documentation for your version of DataRobot.

Batch Predictions

This page outlines the operations, endpoints, parameters, and example requests and responses for the Batch Predictions.

GET /api/v2/batchJobs/

Get a collection of batch jobs by statuses

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchJobs/?offset=0&limit=100 \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer true This many results will be skipped
limit query integer true At most this many results are returned
status query any false Includes only jobs that have the status value that matches this flag. Repeat the parameter for filtering on multiple statuses.
source query any false Includes only jobs that have the source value that matches this flag. Repeat the parameter for filtering on multiple statuses.Prefix values with a dash (-) to exclude those sources.
deploymentId query string false Includes only jobs for this particular deployment
modelId query string false ID of leaderboard model which is used in job for processing predictions dataset
jobId query string false Includes only job by specific id
orderBy query string false Sort order which will be applied to batch prediction list. Prefix the attribute name with a dash to sort in descending order, e.g. "-created".
allJobs query boolean false [DEPRECATED - replaced with RBAC permission model] - No effect
cutoffHours query integer false Only list jobs created at most this amount of hours ago.
startDateTime query string(date-time) false ISO-formatted datetime of the earliest time the job was added (inclusive). For example "2008-08-24T12:00:00Z". Will ignore cutoffHours if set.
endDateTime query string(date-time) false ISO-formatted datetime of the latest time the job was added (inclusive). For example "2008-08-24T12:00:00Z".
batchPredictionJobDefinitionId query string false Includes only jobs for this particular definition
hostname query any false Includes only jobs for this particular prediction instance hostname
batchJobType query any false Includes only jobs that have the batch job type that matches this flag. Repeat the parameter for filtering on multiple types.
intakeType query any false Includes only jobs for these particular intakes type
outputType query any false Includes only jobs for these particular outputs type

Enumerated Values

Parameter Value
orderBy [created, -created, status, -status]

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "batchMonitoringJobDefinition": {
        "createdBy": "string",
        "id": "string",
        "name": "string"
      },
      "batchPredictionJobDefinition": {
        "createdBy": "string",
        "id": "string",
        "name": "string"
      },
      "created": "2019-08-24T14:15:22Z",
      "createdBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      },
      "elapsedTimeSec": 0,
      "failedRows": 0,
      "hidden": "2019-08-24T14:15:22Z",
      "id": "string",
      "intakeDatasetDisplayName": "string",
      "jobIntakeSize": 0,
      "jobOutputSize": 0,
      "jobSpec": {
        "abortOnError": true,
        "batchJobType": "monitoring",
        "chunkSize": "auto",
        "columnNamesRemapping": {},
        "csvSettings": {
          "delimiter": ",",
          "encoding": "utf-8",
          "quotechar": "\""
        },
        "deploymentId": "string",
        "disableRowLevelErrorHandling": false,
        "explanationAlgorithm": "shap",
        "explanationClassNames": [
          "string"
        ],
        "explanationNumTopClasses": 1,
        "includePredictionStatus": false,
        "includeProbabilities": true,
        "includeProbabilitiesClasses": [],
        "intakeSettings": {
          "type": "localFile"
        },
        "maxExplanations": 0,
        "maxNgramExplanations": 0,
        "modelId": "string",
        "modelPackageId": "string",
        "monitoringAggregation": {
          "retentionPolicy": "samples",
          "retentionValue": 0
        },
        "monitoringBatchPrefix": "string",
        "monitoringColumns": {
          "actedUponColumn": "string",
          "actualsTimestampColumn": "string",
          "actualsValueColumn": "string",
          "associationIdColumn": "string",
          "customMetricId": "string",
          "customMetricTimestampColumn": "string",
          "customMetricTimestampFormat": "string",
          "customMetricValueColumn": "string",
          "monitoredStatusColumn": "string",
          "predictionsColumns": [
            {
              "className": "string",
              "columnName": "string"
            }
          ],
          "reportDrift": true,
          "reportPredictions": true,
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "monitoringOutputSettings": {
          "monitoredStatusColumn": "string",
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "numConcurrent": 1,
        "outputSettings": {
          "credentialId": "string",
          "format": "csv",
          "partitionColumns": [
            "string"
          ],
          "type": "azure",
          "url": "string"
        },
        "passthroughColumns": [
          "string"
        ],
        "passthroughColumnsSet": "all",
        "pinnedModelId": "string",
        "predictionInstance": {
          "apiKey": "string",
          "datarobotKey": "string",
          "hostName": "string",
          "sslEnabled": true
        },
        "predictionWarningEnabled": true,
        "redactedFields": [
          "string"
        ],
        "skipDriftTracking": false,
        "thresholdHigh": 0,
        "thresholdLow": 0,
        "timeseriesSettings": {
          "forecastPoint": "2019-08-24T14:15:22Z",
          "relaxKnownInAdvanceFeaturesCheck": false,
          "type": "forecast"
        }
      },
      "links": {
        "csvUpload": "string",
        "download": "string",
        "self": "string"
      },
      "logs": [
        "string"
      ],
      "monitoringBatchId": "string",
      "percentageCompleted": 100,
      "queuePosition": 0,
      "queued": true,
      "resultsDeleted": true,
      "scoredRows": 0,
      "skippedRows": 0,
      "source": "string",
      "status": "INITIALIZING",
      "statusDetails": "string"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Responses

Status Meaning Description Schema
200 OK A list of Batch job objects BatchJobListResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/batchJobs/fromJobDefinition/

Launches a one-time batch job based off of the previously supplied definition referring to the job definition ID and puts it on the queue.

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/batchJobs/fromJobDefinition/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{BatchPredictionJobDefinitionId}'

Body parameter

{
  "jobDefinitionId": "string"
}

Parameters

Name In Type Required Description
body body BatchPredictionJobDefinitionId false none

Example responses

202 Response

{
  "batchMonitoringJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "batchPredictionJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "elapsedTimeSec": 0,
  "failedRows": 0,
  "hidden": "2019-08-24T14:15:22Z",
  "id": "string",
  "intakeDatasetDisplayName": "string",
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "jobSpec": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 1,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "links": {
    "csvUpload": "string",
    "download": "string",
    "self": "string"
  },
  "logs": [
    "string"
  ],
  "monitoringBatchId": "string",
  "percentageCompleted": 100,
  "queuePosition": 0,
  "queued": true,
  "resultsDeleted": true,
  "scoredRows": 0,
  "skippedRows": 0,
  "source": "string",
  "status": "INITIALIZING",
  "statusDetails": "string"
}

Responses

Status Meaning Description Schema
202 Accepted Job details for the created Batch Prediction job BatchJobResponse
404 Not Found Job was deleted, never existed or you do not have access to it None
422 Unprocessable Entity Could not create a Batch job. Possible reasons: {} None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

DELETE /api/v2/batchJobs/{batchJobId}/

If the job is running, it will be aborted. Then it will be removed, meaning all underlying data will be deleted and the job is removed from the list of jobs.

Code samples

# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/batchJobs/{batchJobId}/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
batchJobId path string true ID of the Batch job
partNumber path integer true The number of which csv part is being uploaded when using multipart upload

Responses

Status Meaning Description Schema
202 Accepted Job cancelled None
404 Not Found Job does not exist or was not submitted to the queue. None
409 Conflict Job cannot be aborted for some reason. Possible reasons: job is already aborted or completed. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/batchJobs/{batchJobId}/

Retrieve a Batch job.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchJobs/{batchJobId}/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
batchJobId path string true ID of the Batch job
partNumber path integer true The number of which csv part is being uploaded when using multipart upload

Example responses

200 Response

{
  "batchMonitoringJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "batchPredictionJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "elapsedTimeSec": 0,
  "failedRows": 0,
  "hidden": "2019-08-24T14:15:22Z",
  "id": "string",
  "intakeDatasetDisplayName": "string",
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "jobSpec": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 1,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "links": {
    "csvUpload": "string",
    "download": "string",
    "self": "string"
  },
  "logs": [
    "string"
  ],
  "monitoringBatchId": "string",
  "percentageCompleted": 100,
  "queuePosition": 0,
  "queued": true,
  "resultsDeleted": true,
  "scoredRows": 0,
  "skippedRows": 0,
  "source": "string",
  "status": "INITIALIZING",
  "statusDetails": "string"
}

Responses

Status Meaning Description Schema
200 OK Job details for the requested Batch job BatchJobResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

PUT /api/v2/batchJobs/{batchJobId}/csvUpload/

Stream CSV data to the job. Only available for jobs thatuses the localFile intake option.

Code samples

# You can also use wget
curl -X PUT https://app.datarobot.com/api/v2/batchJobs/{batchJobId}/csvUpload/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
batchJobId path string true ID of the Batch job
partNumber path integer true The number of which csv part is being uploaded when using multipart upload

Responses

Status Meaning Description Schema
202 Accepted Job data was successfully submitted None
404 Not Found Job does not exist or does not require data None
409 Conflict Dataset upload has already begun None
415 Unsupported Media Type Not acceptable MIME type None
422 Unprocessable Entity Job was "ABORTED" due to too many errors in the data None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/batchJobs/{batchJobId}/download/

This is only valid for jobs scored using the "localFile" output option

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchJobs/{batchJobId}/download/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
batchJobId path string true ID of the Batch job
partNumber path integer true The number of which csv part is being uploaded when using multipart upload

Responses

Status Meaning Description Schema
200 OK Job was downloaded correctly None
404 Not Found Job does not exist or is not completed None
406 Not Acceptable Not acceptable MIME type None
422 Unprocessable Entity Job was "ABORTED" due to too many errors in the data None

Response Headers

Status Header Type Format Description
200 Content-Disposition string Contains an auto generated filename for this download ("attachment;filename=result-.csv").
200 Content-Type string MIME type of the returned data

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/batchPredictionJobDefinitions/

List all Batch Prediction jobs definitions available

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchPredictionJobDefinitions/?offset=0&limit=100 \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer true This many results will be skipped
limit query integer true At most this many results are returned
searchName query string false A human-readable name for the definition, must be unique across organisations.
deploymentId query string false Includes only definitions for this particular deployment

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "batchPredictionJob": {
        "abortOnError": true,
        "batchJobType": "monitoring",
        "chunkSize": "auto",
        "columnNamesRemapping": {},
        "csvSettings": {
          "delimiter": ",",
          "encoding": "utf-8",
          "quotechar": "\""
        },
        "deploymentId": "string",
        "disableRowLevelErrorHandling": false,
        "explanationAlgorithm": "shap",
        "explanationClassNames": [
          "string"
        ],
        "explanationNumTopClasses": 1,
        "includePredictionStatus": false,
        "includeProbabilities": true,
        "includeProbabilitiesClasses": [],
        "intakeSettings": {
          "type": "localFile"
        },
        "maxExplanations": 0,
        "maxNgramExplanations": 0,
        "modelId": "string",
        "modelPackageId": "string",
        "monitoringAggregation": {
          "retentionPolicy": "samples",
          "retentionValue": 0
        },
        "monitoringBatchPrefix": "string",
        "monitoringColumns": {
          "actedUponColumn": "string",
          "actualsTimestampColumn": "string",
          "actualsValueColumn": "string",
          "associationIdColumn": "string",
          "customMetricId": "string",
          "customMetricTimestampColumn": "string",
          "customMetricTimestampFormat": "string",
          "customMetricValueColumn": "string",
          "monitoredStatusColumn": "string",
          "predictionsColumns": [
            {
              "className": "string",
              "columnName": "string"
            }
          ],
          "reportDrift": true,
          "reportPredictions": true,
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "monitoringOutputSettings": {
          "monitoredStatusColumn": "string",
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "numConcurrent": 0,
        "outputSettings": {
          "credentialId": "string",
          "format": "csv",
          "partitionColumns": [
            "string"
          ],
          "type": "azure",
          "url": "string"
        },
        "passthroughColumns": [
          "string"
        ],
        "passthroughColumnsSet": "all",
        "pinnedModelId": "string",
        "predictionInstance": {
          "apiKey": "string",
          "datarobotKey": "string",
          "hostName": "string",
          "sslEnabled": true
        },
        "predictionWarningEnabled": true,
        "redactedFields": [
          "string"
        ],
        "skipDriftTracking": false,
        "thresholdHigh": 0,
        "thresholdLow": 0,
        "timeseriesSettings": {
          "forecastPoint": "2019-08-24T14:15:22Z",
          "relaxKnownInAdvanceFeaturesCheck": false,
          "type": "forecast"
        }
      },
      "created": "2019-08-24T14:15:22Z",
      "createdBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      },
      "enabled": false,
      "id": "string",
      "lastFailedRunTime": "2019-08-24T14:15:22Z",
      "lastScheduledRunTime": "2019-08-24T14:15:22Z",
      "lastStartedJobStatus": "INITIALIZING",
      "lastStartedJobTime": "2019-08-24T14:15:22Z",
      "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
      "name": "string",
      "nextScheduledRunTime": "2019-08-24T14:15:22Z",
      "schedule": {
        "dayOfMonth": [
          "*"
        ],
        "dayOfWeek": [
          "*"
        ],
        "hour": [
          "*"
        ],
        "minute": [
          "*"
        ],
        "month": [
          "*"
        ]
      },
      "updated": "2019-08-24T14:15:22Z",
      "updatedBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      }
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Responses

Status Meaning Description Schema
200 OK List of all available jobs BatchPredictionJobDefinitionsListResponse
422 Unprocessable Entity Your input data or query arguments did not work together None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/batchPredictionJobDefinitions/

Create a Batch Prediction Job definition. A configuration for a Batch Prediction job which can either be executed manually upon request or on scheduled intervals, if enabled. The API payload is the same as for /batchPredictions along with optional enabled and schedule items.

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/batchPredictionJobDefinitions/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{BatchPredictionJobDefinitionsCreate}'

Body parameter

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "enabled": true,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringBatchPrefix": "string",
  "name": "string",
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Parameters

Name In Type Required Description
body body BatchPredictionJobDefinitionsCreate false none

Example responses

202 Response

{
  "batchPredictionJob": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 0,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "enabled": false,
  "id": "string",
  "lastFailedRunTime": "2019-08-24T14:15:22Z",
  "lastScheduledRunTime": "2019-08-24T14:15:22Z",
  "lastStartedJobStatus": "INITIALIZING",
  "lastStartedJobTime": "2019-08-24T14:15:22Z",
  "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
  "name": "string",
  "nextScheduledRunTime": "2019-08-24T14:15:22Z",
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "updated": "2019-08-24T14:15:22Z",
  "updatedBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  }
}

Responses

Status Meaning Description Schema
202 Accepted Job details for the created Batch Prediction job definition BatchPredictionJobDefinitionsResponse
403 Forbidden You are not authorized to create a job definition on this deployment due to your permissions role None
422 Unprocessable Entity You tried to create a job definition with uncompatible or missing parameters to create a fully functioning job definition None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

DELETE /api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/

Delete a Batch Prediction job definition

Code samples

# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
jobDefinitionId path string true ID of the Batch Prediction job definition

Responses

Status Meaning Description Schema
204 No Content none None
403 Forbidden You are not authorized to delete this job definition due to your permissions role None
404 Not Found Job was deleted, never existed or you do not have access to it None
409 Conflict Job could not be deleted, as there are currently running jobs in the queue. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/

Retrieve a Batch Prediction job definition

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
jobDefinitionId path string true ID of the Batch Prediction job definition

Example responses

200 Response

{
  "batchPredictionJob": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 0,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "enabled": false,
  "id": "string",
  "lastFailedRunTime": "2019-08-24T14:15:22Z",
  "lastScheduledRunTime": "2019-08-24T14:15:22Z",
  "lastStartedJobStatus": "INITIALIZING",
  "lastStartedJobTime": "2019-08-24T14:15:22Z",
  "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
  "name": "string",
  "nextScheduledRunTime": "2019-08-24T14:15:22Z",
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "updated": "2019-08-24T14:15:22Z",
  "updatedBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  }
}

Responses

Status Meaning Description Schema
200 OK Job details for the requested Batch Prediction job definition BatchPredictionJobDefinitionsResponse
404 Not Found Job was deleted, never existed or you do not have access to it None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

PATCH /api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/

Update a Batch Prediction job definition

Code samples

# You can also use wget
curl -X PATCH https://app.datarobot.com/api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "enabled": true,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringBatchPrefix": "string",
  "name": "string",
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Parameters

Name In Type Required Description
jobDefinitionId path string true ID of the Batch Prediction job definition
body body BatchPredictionJobDefinitionsUpdate false none

Example responses

200 Response

{
  "batchPredictionJob": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 0,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "enabled": false,
  "id": "string",
  "lastFailedRunTime": "2019-08-24T14:15:22Z",
  "lastScheduledRunTime": "2019-08-24T14:15:22Z",
  "lastStartedJobStatus": "INITIALIZING",
  "lastStartedJobTime": "2019-08-24T14:15:22Z",
  "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
  "name": "string",
  "nextScheduledRunTime": "2019-08-24T14:15:22Z",
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "updated": "2019-08-24T14:15:22Z",
  "updatedBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  }
}

Responses

Status Meaning Description Schema
200 OK Job details for the updated Batch Prediction job definition BatchPredictionJobDefinitionsResponse
403 Forbidden You are not authorized to alter the contents of this job definition due to your permissions role None
404 Not Found Job was deleted, never existed or you do not have access to it None
409 Conflict You chose a name of your job definition that was already existing within your organization None
422 Unprocessable Entity Could not update the job definition. Possible reasons: {} None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/portable/

Retrieve a Batch Prediction job definition for Portable Batch Predictions

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/portable/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
jobDefinitionId path string true ID of the Batch Prediction job definition

Responses

Status Meaning Description Schema
200 OK Snippet for Portable Batch Predictions None
404 Not Found Job was deleted, never existed or you do not have access to it None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/batchPredictions/

Get a collection of batch prediction jobs by statuses

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchPredictions/?offset=0&limit=100 \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer true This many results will be skipped
limit query integer true At most this many results are returned
status query any false Includes only jobs that have the status value that matches this flag. Repeat the parameter for filtering on multiple statuses.
source query any false Includes only jobs that have the source value that matches this flag. Repeat the parameter for filtering on multiple statuses.Prefix values with a dash (-) to exclude those sources.
deploymentId query string false Includes only jobs for this particular deployment
modelId query string false ID of leaderboard model which is used in job for processing predictions dataset
jobId query string false Includes only job by specific id
orderBy query string false Sort order which will be applied to batch prediction list. Prefix the attribute name with a dash to sort in descending order, e.g. "-created".
allJobs query boolean false [DEPRECATED - replaced with RBAC permission model] - No effect
cutoffHours query integer false Only list jobs created at most this amount of hours ago.
startDateTime query string(date-time) false ISO-formatted datetime of the earliest time the job was added (inclusive). For example "2008-08-24T12:00:00Z". Will ignore cutoffHours if set.
endDateTime query string(date-time) false ISO-formatted datetime of the latest time the job was added (inclusive). For example "2008-08-24T12:00:00Z".
batchPredictionJobDefinitionId query string false Includes only jobs for this particular definition
hostname query any false Includes only jobs for this particular prediction instance hostname
intakeType query any false Includes only jobs for these particular intakes type
outputType query any false Includes only jobs for these particular outputs type

Enumerated Values

Parameter Value
orderBy [created, -created, status, -status]

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "batchPredictionJobDefinition": {
        "createdBy": "string",
        "id": "string",
        "name": "string"
      },
      "created": "2019-08-24T14:15:22Z",
      "createdBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      },
      "elapsedTimeSec": 0,
      "failedRows": 0,
      "hidden": "2019-08-24T14:15:22Z",
      "id": "string",
      "intakeDatasetDisplayName": "string",
      "jobIntakeSize": 0,
      "jobOutputSize": 0,
      "jobSpec": {
        "abortOnError": true,
        "batchJobType": "monitoring",
        "chunkSize": "auto",
        "columnNamesRemapping": {},
        "csvSettings": {
          "delimiter": ",",
          "encoding": "utf-8",
          "quotechar": "\""
        },
        "deploymentId": "string",
        "disableRowLevelErrorHandling": false,
        "explanationAlgorithm": "shap",
        "explanationClassNames": [
          "string"
        ],
        "explanationNumTopClasses": 1,
        "includePredictionStatus": false,
        "includeProbabilities": true,
        "includeProbabilitiesClasses": [],
        "intakeSettings": {
          "type": "localFile"
        },
        "maxExplanations": 0,
        "modelId": "string",
        "modelPackageId": "string",
        "monitoringBatchPrefix": "string",
        "numConcurrent": 1,
        "outputSettings": {
          "credentialId": "string",
          "format": "csv",
          "partitionColumns": [
            "string"
          ],
          "type": "azure",
          "url": "string"
        },
        "passthroughColumns": [
          "string"
        ],
        "passthroughColumnsSet": "all",
        "pinnedModelId": "string",
        "predictionInstance": {
          "apiKey": "string",
          "datarobotKey": "string",
          "hostName": "string",
          "sslEnabled": true
        },
        "predictionThreshold": 1,
        "predictionWarningEnabled": true,
        "redactedFields": [
          "string"
        ],
        "secondaryDatasetsConfigId": "string",
        "skipDriftTracking": false,
        "thresholdHigh": 0,
        "thresholdLow": 0,
        "timeseriesSettings": {
          "forecastPoint": "2019-08-24T14:15:22Z",
          "relaxKnownInAdvanceFeaturesCheck": false,
          "type": "forecast"
        }
      },
      "links": {
        "csvUpload": "string",
        "download": "string",
        "self": "string"
      },
      "logs": [
        "string"
      ],
      "monitoringBatchId": "string",
      "percentageCompleted": 100,
      "queuePosition": 0,
      "queued": true,
      "resultsDeleted": true,
      "scoredRows": 0,
      "skippedRows": 0,
      "source": "string",
      "status": "INITIALIZING",
      "statusDetails": "string"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Responses

Status Meaning Description Schema
200 OK A list of Batch Prediction job objects BatchPredictionJobListResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/batchPredictions/

Submit the configuration for the job and it will be submitted to the queue

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/batchPredictions/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{BatchPredictionJobCreate}'

Body parameter

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringBatchPrefix": "string",
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Parameters

Name In Type Required Description
body body BatchPredictionJobCreate false none

Example responses

202 Response

{
  "batchPredictionJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "elapsedTimeSec": 0,
  "failedRows": 0,
  "hidden": "2019-08-24T14:15:22Z",
  "id": "string",
  "intakeDatasetDisplayName": "string",
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "jobSpec": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringBatchPrefix": "string",
    "numConcurrent": 1,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionThreshold": 1,
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "secondaryDatasetsConfigId": "string",
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "links": {
    "csvUpload": "string",
    "download": "string",
    "self": "string"
  },
  "logs": [
    "string"
  ],
  "monitoringBatchId": "string",
  "percentageCompleted": 100,
  "queuePosition": 0,
  "queued": true,
  "resultsDeleted": true,
  "scoredRows": 0,
  "skippedRows": 0,
  "source": "string",
  "status": "INITIALIZING",
  "statusDetails": "string"
}

Responses

Status Meaning Description Schema
202 Accepted Job details for the created Batch Prediction job BatchPredictionJobResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/batchPredictions/fromExisting/

Copies an existing job and submits it to the queue.

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/batchPredictions/fromExisting/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{BatchPredictionJobId}'

Body parameter

{
  "partNumber": 0,
  "predictionJobId": "string"
}

Parameters

Name In Type Required Description
body body BatchPredictionJobId false none

Example responses

202 Response

{
  "batchPredictionJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "elapsedTimeSec": 0,
  "failedRows": 0,
  "hidden": "2019-08-24T14:15:22Z",
  "id": "string",
  "intakeDatasetDisplayName": "string",
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "jobSpec": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringBatchPrefix": "string",
    "numConcurrent": 1,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionThreshold": 1,
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "secondaryDatasetsConfigId": "string",
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "links": {
    "csvUpload": "string",
    "download": "string",
    "self": "string"
  },
  "logs": [
    "string"
  ],
  "monitoringBatchId": "string",
  "percentageCompleted": 100,
  "queuePosition": 0,
  "queued": true,
  "resultsDeleted": true,
  "scoredRows": 0,
  "skippedRows": 0,
  "source": "string",
  "status": "INITIALIZING",
  "statusDetails": "string"
}

Responses

Status Meaning Description Schema
202 Accepted Job details for the created Batch Prediction job BatchPredictionJobResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/batchPredictions/fromJobDefinition/

Launches a one-time batch prediction job based off of the previously supplied definition referring to the job definition ID and puts it on the queue.

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/batchPredictions/fromJobDefinition/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{BatchPredictionJobDefinitionId}'

Body parameter

{
  "jobDefinitionId": "string"
}

Parameters

Name In Type Required Description
body body BatchPredictionJobDefinitionId false none

Example responses

202 Response

{
  "batchPredictionJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "elapsedTimeSec": 0,
  "failedRows": 0,
  "hidden": "2019-08-24T14:15:22Z",
  "id": "string",
  "intakeDatasetDisplayName": "string",
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "jobSpec": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringBatchPrefix": "string",
    "numConcurrent": 1,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionThreshold": 1,
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "secondaryDatasetsConfigId": "string",
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "links": {
    "csvUpload": "string",
    "download": "string",
    "self": "string"
  },
  "logs": [
    "string"
  ],
  "monitoringBatchId": "string",
  "percentageCompleted": 100,
  "queuePosition": 0,
  "queued": true,
  "resultsDeleted": true,
  "scoredRows": 0,
  "skippedRows": 0,
  "source": "string",
  "status": "INITIALIZING",
  "statusDetails": "string"
}

Responses

Status Meaning Description Schema
202 Accepted Job details for the created Batch Prediction job BatchPredictionJobResponse
404 Not Found Job was deleted, never existed or you do not have access to it None
422 Unprocessable Entity Could not create a Batch job. Possible reasons: {} None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

DELETE /api/v2/batchPredictions/{predictionJobId}/

If the job is running, it will be aborted. Then it will be removed, meaning all underlying data will be deleted and the job is removed from the list of jobs.

Code samples

# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/batchPredictions/{predictionJobId}/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
predictionJobId path string true ID of the Batch Prediction job
partNumber path integer true The number of which csv part is being uploaded when using multipart upload

Responses

Status Meaning Description Schema
202 Accepted Job cancelled None
404 Not Found Job does not exist or was not submitted to the queue. None
409 Conflict Job cannot be aborted for some reason. Possible reasons: job is already aborted or completed. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/batchPredictions/{predictionJobId}/

Retrieve a Batch Prediction job.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchPredictions/{predictionJobId}/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
predictionJobId path string true ID of the Batch Prediction job
partNumber path integer true The number of which csv part is being uploaded when using multipart upload

Example responses

200 Response

{
  "batchPredictionJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "elapsedTimeSec": 0,
  "failedRows": 0,
  "hidden": "2019-08-24T14:15:22Z",
  "id": "string",
  "intakeDatasetDisplayName": "string",
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "jobSpec": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringBatchPrefix": "string",
    "numConcurrent": 1,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionThreshold": 1,
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "secondaryDatasetsConfigId": "string",
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "links": {
    "csvUpload": "string",
    "download": "string",
    "self": "string"
  },
  "logs": [
    "string"
  ],
  "monitoringBatchId": "string",
  "percentageCompleted": 100,
  "queuePosition": 0,
  "queued": true,
  "resultsDeleted": true,
  "scoredRows": 0,
  "skippedRows": 0,
  "source": "string",
  "status": "INITIALIZING",
  "statusDetails": "string"
}

Responses

Status Meaning Description Schema
200 OK Job details for the requested Batch Prediction job BatchPredictionJobResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

PATCH /api/v2/batchPredictions/{predictionJobId}/

If a job has finished execution regardless of the result, it can have parameters changed to ensure better filtering in the job list upon retrieval. Another case: updating job scoring status externally.

Code samples

# You can also use wget
curl -X PATCH https://app.datarobot.com/api/v2/batchPredictions/{predictionJobId}/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "aborted": "2019-08-24T14:15:22Z",
  "completed": "2019-08-24T14:15:22Z",
  "failedRows": 0,
  "hidden": true,
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "logs": [
    "string"
  ],
  "scoredRows": 0,
  "skippedRows": 0,
  "started": "2019-08-24T14:15:22Z",
  "status": "INITIALIZING"
}

Parameters

Name In Type Required Description
predictionJobId path string true ID of the Batch Prediction job
partNumber path integer true The number of which csv part is being uploaded when using multipart upload
body body BatchPredictionJobUpdate false none

Example responses

200 Response

{
  "batchPredictionJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "elapsedTimeSec": 0,
  "failedRows": 0,
  "hidden": "2019-08-24T14:15:22Z",
  "id": "string",
  "intakeDatasetDisplayName": "string",
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "jobSpec": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringBatchPrefix": "string",
    "numConcurrent": 1,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionThreshold": 1,
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "secondaryDatasetsConfigId": "string",
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "links": {
    "csvUpload": "string",
    "download": "string",
    "self": "string"
  },
  "logs": [
    "string"
  ],
  "monitoringBatchId": "string",
  "percentageCompleted": 100,
  "queuePosition": 0,
  "queued": true,
  "resultsDeleted": true,
  "scoredRows": 0,
  "skippedRows": 0,
  "source": "string",
  "status": "INITIALIZING",
  "statusDetails": "string"
}

Responses

Status Meaning Description Schema
200 OK Job updated BatchPredictionJobResponse
404 Not Found Job does not exist or was not submitted to the queue. None
409 Conflict Job cannot be hidden for some reason. Possible reasons: job is not in a deletable state. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

PUT /api/v2/batchPredictions/{predictionJobId}/csvUpload/

Stream CSV data to the prediction job. Only available for jobs thatuses the localFile intake option.

Code samples

# You can also use wget
curl -X PUT https://app.datarobot.com/api/v2/batchPredictions/{predictionJobId}/csvUpload/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
predictionJobId path string true ID of the Batch Prediction job
partNumber path integer true The number of which csv part is being uploaded when using multipart upload

Responses

Status Meaning Description Schema
202 Accepted Job data was successfully submitted None
404 Not Found Job does not exist or does not require data None
406 Not Acceptable Not acceptable MIME type None
409 Conflict Dataset upload has already begun None
422 Unprocessable Entity Job was "ABORTED" due to too many errors in the data None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/batchPredictions/{predictionJobId}/csvUpload/finalizeMultipart/

Finalize a multipart upload, indicating that no further chunks will be sent

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/batchPredictions/{predictionJobId}/csvUpload/finalizeMultipart/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
predictionJobId path string true ID of the Batch Prediction job
partNumber path integer true The number of which csv part is being uploaded when using multipart upload

Responses

Status Meaning Description Schema
202 Accepted Acknowledgement that the request was accepted or an error message None
404 Not Found Job was deleted, never existed or you do not have access to it None
409 Conflict Only multipart jobs can be finalized. None
422 Unprocessable Entity No data was uploaded None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

PUT /api/v2/batchPredictions/{predictionJobId}/csvUpload/part/{partNumber}/

Stream CSV data to the prediction job in many parts.Only available for jobs that uses the localFile intake option.

Code samples

# You can also use wget
curl -X PUT https://app.datarobot.com/api/v2/batchPredictions/{predictionJobId}/csvUpload/part/{partNumber}/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
predictionJobId path string true ID of the Batch Prediction job
partNumber path integer true The number of which csv part is being uploaded when using multipart upload

Responses

Status Meaning Description Schema
202 Accepted Job data was successfully submitted None
404 Not Found Job does not exist or does not require data None
406 Not Acceptable Not acceptable MIME type None
409 Conflict Dataset upload has already begun None
422 Unprocessable Entity Job was "ABORTED" due to too many errors in the data None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/batchPredictions/{predictionJobId}/download/

This is only valid for jobs scored using the "localFile" output option

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchPredictions/{predictionJobId}/download/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
predictionJobId path string true ID of the Batch Prediction job
partNumber path integer true The number of which csv part is being uploaded when using multipart upload

Responses

Status Meaning Description Schema
200 OK Job was downloaded correctly None
404 Not Found Job does not exist or is not completed None
406 Not Acceptable Not acceptable MIME type None
422 Unprocessable Entity Job was "ABORTED" due to too many errors in the data None

Response Headers

Status Header Type Format Description
200 Content-Disposition string Contains an auto generated filename for this download ("attachment;filename=result-.csv").
200 Content-Type string MIME type of the returned data

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

DELETE /api/v2/projects/{projectId}/models/{modelId}/predictionExplanationsInitialization/

Delete an existing PredictionExplanationsInitialization.

Code samples

# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/projects/{projectId}/models/{modelId}/predictionExplanationsInitialization/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
projectId path string true The project ID
modelId path string true The model ID

Responses

Status Meaning Description Schema
204 No Content The deletion was successful. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/projects/{projectId}/models/{modelId}/predictionExplanationsInitialization/

Retrieve the current PredictionExplanationsInitialization. A PredictionExplanationsInitialization is a pre-requisite for successfully computing prediction explanations using a particular model, and can be used to preview the prediction explanations that would be generated for a complete dataset.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/models/{modelId}/predictionExplanationsInitialization/ \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "modelId": "string",
  "predictionExplanationsSample": [
    {
      "adjustedPrediction": 0,
      "adjustedPredictionValues": [
        {
          "label": "string",
          "value": 0
        }
      ],
      "forecastDistance": 0,
      "forecastPoint": "string",
      "prediction": 0,
      "predictionExplanations": [
        {
          "feature": "string",
          "featureValue": "string",
          "imageExplanationUrl": "string",
          "label": "string",
          "perNgramTextExplanations": [
            {
              "isUnknown": true,
              "ngrams": [
                {
                  "label": "string",
                  "value": 0
                }
              ],
              "qualitativateStrength": "string",
              "strength": 0
            }
          ],
          "qualitativateStrength": "string",
          "strength": 0
        }
      ],
      "predictionThreshold": 0,
      "predictionValues": [
        {
          "label": "string",
          "value": 0
        }
      ],
      "rowId": 0,
      "seriesId": "string",
      "timestamp": "string"
    }
  ],
  "projectId": "string"
}

Parameters

Name In Type Required Description
excludeAdjustedPredictions query string false Whether to include adjusted prediction in the PredictionExplanationsSample response.
projectId path string true The project ID
modelId path string true The model ID
body body PredictionExplanationsInitializationRetrieve false none

Enumerated Values

Parameter Value
excludeAdjustedPredictions [false, False, true, True]

Responses

Status Meaning Description Schema
200 OK none None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/projects/{projectId}/models/{modelId}/predictionExplanationsInitialization/

Create a new prediction explanations initialization. This is a necessary prerequisite for generating prediction explanations.

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/models/{modelId}/predictionExplanationsInitialization/ \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "maxExplanations": 3,
  "thresholdHigh": null,
  "thresholdLow": null
}

Parameters

Name In Type Required Description
projectId path string true The project ID
modelId path string true The model ID
body body PredictionExplanationsInitializationCreate false none

Responses

Status Meaning Description Schema
202 Accepted The request was accepted and will be worked on. None

Response Headers

Status Header Type Format Description
202 Location string A url that can be polled to check the status.

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/projects/{projectId}/predictJobs/

List all prediction jobs for a project

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictJobs/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
status query string false If provided, only jobs with the same status will be included in the results; otherwise, queued and inprogress jobs (but not errored jobs) will be returned.
projectId path string true The project ID.

Enumerated Values

Parameter Value
status [queue, inprogress, error]

Example responses

200 Response

[
  {
    "id": "string",
    "isBlocked": true,
    "message": "string",
    "modelId": "string",
    "projectId": "string",
    "status": "queue"
  }
]

Responses

Status Meaning Description Schema
200 OK A list of prediction jobs for a project Inline
404 Not Found Job was not found None

Response Schema

Status Code 200

Name Type Required Restrictions Description
anonymous [PredictJobDetailsResponse] false none
» id string true the job ID of the job
» isBlocked boolean true True if a job is waiting for its dependencies to be resolved first.
» message string true An optional message about the job
» modelId string true The ID of the model
» projectId string true the project the job belongs to
» status string true the status of the job

Enumerated Values

Property Value
status [queue, inprogress, error, ABORTED, COMPLETED]

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

DELETE /api/v2/projects/{projectId}/predictJobs/{jobId}/

Cancel a queued prediction job

Code samples

# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/projects/{projectId}/predictJobs/{jobId}/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
projectId path string true The project ID.
jobId path string true The job ID

Responses

Status Meaning Description Schema
204 No Content The job has been successfully cancelled None
404 Not Found Job was not found or the job has already completed None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/projects/{projectId}/predictJobs/{jobId}/

Look up a particular prediction job

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictJobs/{jobId}/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
projectId path string true The project ID.
jobId path string true The job ID

Example responses

200 Response

{
  "id": "string",
  "isBlocked": true,
  "message": "string",
  "modelId": "string",
  "projectId": "string",
  "status": "queue"
}

Responses

Status Meaning Description Schema
200 OK The job has been successfully retrieved and has not yet finished. PredictJobDetailsResponse
303 See Other The job has been successfully retrieved and has been completed. See Location header. The response json is also included. None

Response Headers

Status Header Type Format Description
200 Location string url present only when the requested job has finished - contains a url from which the completed predictions may be retrieved as with GET /api/v2/projects/{projectId}/predictions/{predictionId}/

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/projects/{projectId}/predictionDatasets/

List predictions datasets uploaded to a project.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictionDatasets/?offset=0&limit=0 \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer true This many results will be skipped.
limit query integer true At most this many results are returned. If 0, all results.
projectId path string true The project ID to query.

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "actualValueColumn": "string",
      "catalogId": "string",
      "catalogVersionId": "string",
      "containsTargetValues": true,
      "created": "2019-08-24T14:15:22Z",
      "dataEndDate": "2019-08-24T14:15:22Z",
      "dataQualityWarnings": {
        "hasKiaMissingValuesInForecastWindow": true,
        "insufficientRowsForEvaluatingModels": true,
        "singleClassActualValueColumn": true
      },
      "dataStartDate": "2019-08-24T14:15:22Z",
      "detectedActualValueColumns": [
        {
          "missingCount": 0,
          "name": "string"
        }
      ],
      "forecastPoint": "string",
      "forecastPointRange": [
        "2019-08-24T14:15:22Z"
      ],
      "id": "string",
      "maxForecastDate": "2019-08-24T14:15:22Z",
      "name": "string",
      "numColumns": 0,
      "numRows": 0,
      "predictionsEndDate": "2019-08-24T14:15:22Z",
      "predictionsStartDate": "2019-08-24T14:15:22Z",
      "projectId": "string",
      "secondaryDatasetsConfigId": "string"
    }
  ],
  "next": "string",
  "previous": "string"
}

Responses

Status Meaning Description Schema
200 OK Request to list the uploaded predictions datasets was successful. PredictionDatasetListControllerResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/projects/{projectId}/predictionDatasets/dataSourceUploads/

Upload a dataset for predictions from a DataSource.

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/predictionDatasets/dataSourceUploads/ \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "actualValueColumn": "string",
  "credentialData": {
    "credentialType": "basic",
    "password": "string",
    "user": "string"
  },
  "credentialId": "string",
  "credentials": [
    {
      "catalogVersionId": "string",
      "password": "string",
      "url": "string",
      "user": "string"
    }
  ],
  "dataSourceId": "string",
  "forecastPoint": "2019-08-24T14:15:22Z",
  "password": "string",
  "predictionsEndDate": "2019-08-24T14:15:22Z",
  "predictionsStartDate": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": true,
  "secondaryDatasetsConfigId": "string",
  "useKerberos": false,
  "user": "string"
}

Parameters

Name In Type Required Description
projectId path string true The project ID to which the data source will be uploaded to.
body body PredictionDataSource false none

Responses

Status Meaning Description Schema
202 Accepted Upload successfully started. See the Location header. None

Response Headers

Status Header Type Format Description
202 Location string A url that can be polled to check the status.

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/projects/{projectId}/predictionDatasets/datasetUploads/

Create a prediction dataset from a Dataset Asset referenced by AI Catalog item/version ID.

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/predictionDatasets/datasetUploads/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "actualValueColumn": "string",
  "credentialData": {
    "credentialType": "basic",
    "password": "string",
    "user": "string"
  },
  "credentialId": "string",
  "credentials": [
    {
      "catalogVersionId": "string",
      "password": "string",
      "url": "string",
      "user": "string"
    }
  ],
  "datasetId": "string",
  "datasetVersionId": "string",
  "forecastPoint": "2019-08-24T14:15:22Z",
  "password": "string",
  "predictionsEndDate": "2019-08-24T14:15:22Z",
  "predictionsStartDate": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": true,
  "secondaryDatasetsConfigId": "string",
  "useKerberos": false,
  "user": "string"
}

Parameters

Name In Type Required Description
projectId path string true The project ID.
body body PredictionFromCatalogDataset false none

Example responses

202 Response

{
  "datasetId": "string"
}

Responses

Status Meaning Description Schema
202 Accepted Creation has successfully started. See the Location header. CreatePredictionDatasetResponse
422 Unprocessable Entity Target not set yet or cannot specify time series options with a non time series project. None

Response Headers

Status Header Type Format Description
202 Location string A url that can be polled to check the status.

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/projects/{projectId}/predictionDatasets/fileUploads/

Upload a file for predictions from an attached file.

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/predictionDatasets/fileUploads/ \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "actualValueColumn": "string",
  "credentials": "string",
  "file": "string",
  "forecastPoint": "2019-08-24T14:15:22Z",
  "predictionsEndDate": "2019-08-24T14:15:22Z",
  "predictionsStartDate": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": "false",
  "secondaryDatasetsConfigId": "string"
}

Parameters

Name In Type Required Description
projectId path string true The project ID to which the data will be uploaded for prediction.
body body PredictionFileUpload false none

Responses

Status Meaning Description Schema
202 Accepted Upload successfully started. See the Location header. None

Response Headers

Status Header Type Format Description
202 Location string A url that can be polled to check the status.

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/projects/{projectId}/predictionDatasets/urlUploads/

Upload a file for predictions from a URL.

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/predictionDatasets/urlUploads/ \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "actualValueColumn": "string",
  "credentials": [
    {
      "catalogVersionId": "string",
      "password": "string",
      "url": "string",
      "user": "string"
    }
  ],
  "forecastPoint": "2019-08-24T14:15:22Z",
  "predictionsEndDate": "2019-08-24T14:15:22Z",
  "predictionsStartDate": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": true,
  "secondaryDatasetsConfigId": "string",
  "url": "string"
}

Parameters

Name In Type Required Description
projectId path string true The project ID to which the data will be uploaded for prediction.
body body PredictionURLUpload false none

Responses

Status Meaning Description Schema
202 Accepted Upload successfully started. See the Location header. None

Response Headers

Status Header Type Format Description
202 Location string A url that can be polled to check the status.

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

DELETE /api/v2/projects/{projectId}/predictionDatasets/{datasetId}/

Delete a dataset that was uploaded for prediction.

Code samples

# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/projects/{projectId}/predictionDatasets/{datasetId}/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
projectId path string true The project ID that owns the data.
datasetId path string true The dataset ID to delete.

Responses

Status Meaning Description Schema
204 No Content The dataset has been successfully deleted. None
404 Not Found No dataset with the specified datasetId found. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/projects/{projectId}/predictionDatasets/{datasetId}/

Get the metadata of a specific dataset. This only works for datasets uploaded to an existing project for prediction.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictionDatasets/{datasetId}/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
projectId path string true The project ID that owns the data.
datasetId path string true The dataset ID to query for.

Example responses

200 Response

{
  "actualValueColumn": "string",
  "catalogId": "string",
  "catalogVersionId": "string",
  "containsTargetValues": true,
  "created": "2019-08-24T14:15:22Z",
  "dataEndDate": "2019-08-24T14:15:22Z",
  "dataQualityWarnings": {
    "hasKiaMissingValuesInForecastWindow": true,
    "insufficientRowsForEvaluatingModels": true,
    "singleClassActualValueColumn": true
  },
  "dataStartDate": "2019-08-24T14:15:22Z",
  "detectedActualValueColumns": [
    {
      "missingCount": 0,
      "name": "string"
    }
  ],
  "forecastPoint": "string",
  "forecastPointRange": [
    "2019-08-24T14:15:22Z"
  ],
  "id": "string",
  "maxForecastDate": "2019-08-24T14:15:22Z",
  "name": "string",
  "numColumns": 0,
  "numRows": 0,
  "predictionsEndDate": "2019-08-24T14:15:22Z",
  "predictionsStartDate": "2019-08-24T14:15:22Z",
  "projectId": "string",
  "secondaryDatasetsConfigId": "string"
}

Responses

Status Meaning Description Schema
200 OK Request to retrieve the metadata of a specified dataset was successful. PredictionDatasetRetrieveResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/projects/{projectId}/predictionExplanations/

Create a new PredictionExplanations object (and its accompanying PredictionExplanationsRecord). In order to successfully create PredictionExplanations for a particular model and dataset, you must first - Compute feature impact for the model via POST /api/v2/projects/{projectId}/models/{modelId}/featureImpact/ - Compute a PredictionExplanationsInitialization for the model via POST /api/v2/projects/{projectId}/models/{modelId}/predictionExplanationsInitialization/ - Compute predictions for the model and dataset via POST /api/v2/projects/{projectId}/predictions/ thresholdHigh and thresholdLow are optional filters applied to speed up computation. When at least one is specified, only the selected outlier rows will have prediction explanations computed. Rows are considered to be outliers if their predicted value (in case of regression projects) or probability of being the positive class (in case of classification projects) isless than thresholdLow or greater than thresholdHigh. If neither is specified, prediction explanations will be computed for all rows.

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/predictionExplanations/ \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "classNames": [
    "string"
  ],
  "datasetId": "string",
  "maxExplanations": 3,
  "modelId": "string",
  "numTopClasses": 1,
  "thresholdHigh": null,
  "thresholdLow": null
}

Parameters

Name In Type Required Description
projectId path string true The project ID
body body PredictionExplanationsCreate false none

Responses

Status Meaning Description Schema
202 Accepted The request was accepted and will be worked on. None

Response Headers

Status Header Type Format Description
202 Location string A url that can be polled to check the status.

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/projects/{projectId}/predictionExplanations/{predictionExplanationsId}/

Retrieve stored Prediction Explanations. Each PredictionExplanationsRow retrieved corresponds to a row of the prediction dataset, although some rows may not have had prediction explanations computed depending on the thresholds selected.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictionExplanations/{predictionExplanationsId}/?limit=3000 \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer false This many results will be skipped.
limit query integer true At most this many results are returned. The default may change and a new maximum limit may be imposed without notice.
excludeAdjustedPredictions query string false Whether to include adjusted prediction in PredictionExplanationsRow response.
projectId path string true The project ID
predictionExplanationsId path string true The ID of the PredictionExplanationsRecord to retrieve.

Enumerated Values

Parameter Value
excludeAdjustedPredictions [false, False, true, True]

Example responses

200 Response

{
  "adjustmentMethod": "string",
  "count": 0,
  "data": [
    {
      "adjustedPrediction": 0,
      "adjustedPredictionValues": [
        {
          "label": "string",
          "value": 0
        }
      ],
      "forecastDistance": 0,
      "forecastPoint": "string",
      "prediction": 0,
      "predictionExplanations": [
        {
          "feature": "string",
          "featureValue": "string",
          "imageExplanationUrl": "string",
          "label": "string",
          "perNgramTextExplanations": [
            {
              "isUnknown": true,
              "ngrams": [
                {
                  "label": "string",
                  "value": 0
                }
              ],
              "qualitativateStrength": "string",
              "strength": 0
            }
          ],
          "qualitativateStrength": "string",
          "strength": 0
        }
      ],
      "predictionThreshold": 0,
      "predictionValues": [
        {
          "label": "string",
          "value": 0
        }
      ],
      "rowId": 0,
      "seriesId": "string",
      "timestamp": "string"
    }
  ],
  "id": "string",
  "next": "http://example.com",
  "predictionExplanationsRecordLocation": "string",
  "previous": "http://example.com"
}

Responses

Status Meaning Description Schema
200 OK none PredictionExplanationsRetrieve

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/projects/{projectId}/predictionExplanationsRecords/

List PredictionExplanationsRecord objects for a project. These contain metadata about the computed prediction explanations and the location at which the PredictionExplanations can be retrieved.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictionExplanationsRecords/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer false This many results will be skipped.
limit query integer false At most this many results are returned.
modelId query string false If specified, only prediction explanations records computed for this model will be returned.
projectId path string true The project ID

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "datasetId": "string",
      "finishTime": 0,
      "id": "string",
      "maxExplanations": 0,
      "modelId": "string",
      "numColumns": 0,
      "predictionExplanationsLocation": "string",
      "predictionThreshold": 0,
      "projectId": "string",
      "thresholdHigh": 0,
      "thresholdLow": 0
    }
  ],
  "next": "string",
  "previous": "string"
}

Responses

Status Meaning Description Schema
200 OK The object was found and returned successfully. PredictionExplanationsRecordList

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

DELETE /api/v2/projects/{projectId}/predictionExplanationsRecords/{predictionExplanationsId}/

Delete saved Prediction Explanations. Deletes both the actual prediction explanations and the corresponding PredictionExplanationsRecord.

Code samples

# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/projects/{projectId}/predictionExplanationsRecords/{predictionExplanationsId}/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
projectId path string true The project ID
predictionExplanationsId path string true The ID of the PredictionExplanationsRecord to retrieve.

Responses

Status Meaning Description Schema
204 No Content The object was deleted successfully. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/projects/{projectId}/predictionExplanationsRecords/{predictionExplanationsId}/

Retrieve a PredictionExplanationsRecord object. A PredictionExplanationsRecord contains metadata about the computed prediction explanations and the location at which the PredictionExplanations can be retrieved.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictionExplanationsRecords/{predictionExplanationsId}/ \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "datasetId": "string",
  "finishTime": 0,
  "id": "string",
  "maxExplanations": 0,
  "modelId": "string",
  "numColumns": 0,
  "predictionExplanationsLocation": "string",
  "predictionThreshold": 0,
  "projectId": "string",
  "thresholdHigh": 0,
  "thresholdLow": 0
}

Parameters

Name In Type Required Description
projectId path string true The project ID
predictionExplanationsId path string true The ID of the PredictionExplanationsRecord to retrieve.
body body PredictionExplanationsRecord false none

Responses

Status Meaning Description Schema
200 OK The object was found and returned successfully. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/projects/{projectId}/predictions/

Get a list of prediction records.

.. deprecated:: v2.21 Use GET /api/v2/projects/{projectId}/predictionsMetadata/ instead. The only difference is that parameter datasetId is renamed to predictionDatasetId both in request and response.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictions/?offset=0&limit=1000 \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer true This many results will be skipped
limit query integer true At most this many results are returned. To specify no limit, use 0. The default may change and a maximum limit may be imposed without notice.
datasetId query string false Dataset id used to create the predictions
modelId query string false Model id
projectId path string true The project of the predictions.

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "actualValueColumn": "string",
      "datasetId": "string",
      "explanationAlgorithm": "string",
      "featureDerivationWindowCounts": 0,
      "forecastPoint": "2019-08-24T14:15:22Z",
      "id": "string",
      "includesPredictionIntervals": true,
      "maxExplanations": 0,
      "modelId": "string",
      "predictionDatasetId": "string",
      "predictionIntervalsSize": 0,
      "predictionThreshold": 0,
      "predictionsEndDate": "2019-08-24T14:15:22Z",
      "predictionsStartDate": "2019-08-24T14:15:22Z",
      "projectId": "string",
      "shapWarnings": {
        "maxNormalizedMismatch": 0,
        "mismatchRowCount": 0
      },
      "url": "string"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com"
}

Responses

Status Meaning Description Schema
200 OK The json array of prediction metadata objects. RetrieveListPredictionMetadataObjectsResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/projects/{projectId}/predictions/

There are two ways of making predictions. The recommended way is to first upload your dataset to the project, and then using the corresponding datasetId, predict against that dataset. To follow that pattern, send the json request body.

Note that requesting prediction intervals will automatically trigger backtesting if backtests were not already completed for this model.

The legacy method which is deprecated is to send the file directly with the predictions request. If you need to predict against a file 10MB in size or larger, you will be required to use the above workflow for uploaded datasets. However, the following multipart/form-data can be used with small files:

:form file: a dataset to make predictions on :form modelId: the model to use to make predictions

.. note:: If using the legacy method of uploading data to this endpoint, a new dataset will be created behind the scenes. For performance reasons, it would be much better to utilize the workflow of creating the dataset first and using the supported method of making predictions of this endpoint. However, to preserve the functionality of existing workflows, the legacy method still exists.

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/predictions/ \
  -H "Content-Type: application/json" \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "actualValueColumn": "string",
  "datasetId": "string",
  "explanationAlgorithm": "shap",
  "forecastPoint": "2019-08-24T14:15:22Z",
  "includeFdwCounts": false,
  "includePredictionIntervals": true,
  "maxExplanations": 1,
  "modelId": "string",
  "predictionIntervalsSize": 1,
  "predictionThreshold": 1,
  "predictionsEndDate": "2019-08-24T14:15:22Z",
  "predictionsStartDate": "2019-08-24T14:15:22Z"
}

Parameters

Name In Type Required Description
projectId path string true The project to make predictions within.
Content-Type header string true Content types available for making request. multipart/form-data is the legacy deprecated method to send the small file with the prediction request.
body body CreatePredictionFromDataset false none

Enumerated Values

Parameter Value
Content-Type [application/json, multipart/form-data]

Responses

Status Meaning Description Schema
202 Accepted Prediction has successfully been requested. See Location header. None
422 Unprocessable Entity The request cannot be processed. None

Response Headers

Status Header Type Format Description
202 Location string A url that can be polled to check the status of the predictions as with GET /api/v2/projects/{projectId}/predictJobs/{jobId}/

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/projects/{projectId}/predictions/{predictionId}/

Retrieve predictions that have previously been computed. Training predictions encoded either as JSON or CSV. If CSV output was requested, the returned CSV data will contain the following columns:

  • For regression projects: row_id and prediction.
  • For binary classification projects: row_id, prediction, class_<positive_class_label> and class_<negative_class_label>.
  • For multiclass projects: row_id, prediction and a class_<class_label> for each class.
  • For multilabel projects: row_id and for each class prediction_<class_label> and class_<class_label>.
  • For time-series, these additional columns will be added: forecast_point, forecast_distance, timestamp, and series_id.

.. minversion:: v2.21

* If `explanationAlgorithm` = 'shap', these additional columns will be added:
  triplets of (`Explanation_<i>_feature_name`,
  `Explanation_<i>_feature_value`, and `Explanation_<i>_strength`) for `i` ranging
  from 1 to `maxExplanations`, `shap_remaining_total` and `shap_base_value`. Binary
  classification projects will also have `explained_class`, the class for which
  positive SHAP values imply an increased probability.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictions/{predictionId}/ \
  -H "Accept: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
predictionId path string true The id of the prediction record to retrieve. If you have the jobId, you can retrieve the predictionId using GET /api/v2/projects/{projectId}/predictJobs/{jobId}/.
projectId path string true The id of the project the prediction belongs to.
Accept header string false Requested MIME type for the returned data

Enumerated Values

Parameter Value
Accept [application/json, text/csv]

Example responses

200 Response

{
  "actualValueColumn": "string",
  "explanationAlgorithm": "string",
  "featureDerivationWindowCounts": 0,
  "includesPredictionIntervals": true,
  "maxExplanations": 0,
  "positiveClass": "string",
  "predictionIntervalsSize": 0,
  "predictions": [
    {
      "actualValue": "string",
      "forecastDistance": 0,
      "forecastPoint": "2019-08-24T14:15:22Z",
      "originalFormatTimestamp": "string",
      "positiveProbability": 0,
      "prediction": 0,
      "predictionExplanationMetadata": [
        {
          "shapRemainingTotal": 0
        }
      ],
      "predictionExplanations": [
        {
          "feature": "string",
          "featureValue": 0,
          "label": "string",
          "strength": 0
        }
      ],
      "predictionIntervalLowerBound": 0,
      "predictionIntervalUpperBound": 0,
      "predictionThreshold": 1,
      "predictionValues": [
        {
          "label": "string",
          "threshold": 1,
          "value": 0
        }
      ],
      "rowId": 0,
      "segmentId": "string",
      "seriesId": "string",
      "target": "string",
      "timestamp": "2019-08-24T14:15:22Z"
    }
  ],
  "shapBaseValue": 0,
  "shapWarnings": [
    {
      "maxNormalizedMismatch": 0,
      "mismatchRowCount": 0
    }
  ],
  "task": "Regression"
}

Responses

Status Meaning Description Schema
200 OK Predictions that have previously been computed. PredictionRetrieveResponse
404 Not Found No prediction data found. None

Response Headers

Status Header Type Format Description
200 Content-Type string MIME type of the returned data

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/projects/{projectId}/predictionsMetadata/

Use the ID of a metadata object to get the complete set of predictions.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictionsMetadata/?offset=0&limit=1000 \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer true This many results will be skipped
limit query integer true At most this many results are returned. To specify no limit, use 0. The default may change and a maximum limit may be imposed without notice.
predictionDatasetId query string false Dataset id used to create the predictions
modelId query string false Model id
projectId path string true The project of the predictions.

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "actualValueColumn": "string",
      "datasetId": "string",
      "explanationAlgorithm": "string",
      "featureDerivationWindowCounts": 0,
      "forecastPoint": "2019-08-24T14:15:22Z",
      "id": "string",
      "includesPredictionIntervals": true,
      "maxExplanations": 0,
      "modelId": "string",
      "predictionDatasetId": "string",
      "predictionIntervalsSize": 0,
      "predictionThreshold": 0,
      "predictionsEndDate": "2019-08-24T14:15:22Z",
      "predictionsStartDate": "2019-08-24T14:15:22Z",
      "projectId": "string",
      "shapWarnings": {
        "maxNormalizedMismatch": 0,
        "mismatchRowCount": 0
      },
      "url": "string"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com"
}

Responses

Status Meaning Description Schema
200 OK The json array of prediction metadata objects. RetrieveListPredictionMetadataObjectsResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/projects/{projectId}/predictionsMetadata/{predictionId}/

Use the ID of a metadata object to get the complete set of predictions.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictionsMetadata/{predictionId}/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
predictionId path string true The id of the prediction record to retrieve. If you have the jobId, you can retrieve the predictionId using GET /api/v2/projects/{projectId}/predictJobs/{jobId}/.
projectId path string true The id of the project the prediction belongs to.

Example responses

200 Response

{
  "actualValueColumn": "string",
  "datasetId": "string",
  "explanationAlgorithm": "string",
  "featureDerivationWindowCounts": 0,
  "forecastPoint": "2019-08-24T14:15:22Z",
  "id": "string",
  "includesPredictionIntervals": true,
  "maxExplanations": 0,
  "modelId": "string",
  "predictionDatasetId": "string",
  "predictionIntervalsSize": 0,
  "predictionThreshold": 0,
  "predictionsEndDate": "2019-08-24T14:15:22Z",
  "predictionsStartDate": "2019-08-24T14:15:22Z",
  "projectId": "string",
  "shapWarnings": {
    "maxNormalizedMismatch": 0,
    "mismatchRowCount": 0
  },
  "url": "string"
}

Responses

Status Meaning Description Schema
200 OK Prediction metadata object. RetrievePredictionMetadataObject
404 Not Found Training predictions not found. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/projects/{projectId}/trainingPredictions/

Get a list of training prediction records

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/trainingPredictions/?offset=0&limit=0 \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer true This many results will be skipped
limit query integer true At most this many results are returned
projectId path string true Project ID to retrieve training predictions for

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "dataSubset": "all",
      "explanationAlgorithm": "shap",
      "id": "string",
      "maxExplanations": 100,
      "modelId": "string",
      "shapWarnings": [
        {
          "partitionName": "string",
          "value": {
            "maxNormalizedMismatch": 0,
            "mismatchRowCount": 0
          }
        }
      ],
      "url": "http://example.com"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com"
}

Responses

Status Meaning Description Schema
200 OK A list of training prediction jobs TrainingPredictionsListResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/projects/{projectId}/trainingPredictions/

Create training data predictions

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/trainingPredictions/ \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "dataSubset": "all",
  "explanationAlgorithm": "string",
  "maxExplanations": 1,
  "modelId": "string"
}

Parameters

Name In Type Required Description
projectId path string true Project ID to compute training predictions for
body body CreateTrainingPrediction false none

Responses

Status Meaning Description Schema
202 Accepted Submitted successfully. See Location header. None
422 Unprocessable Entity - Model/Timeseries/Blender does not support shap based prediction explanations
- Error message from StackedPredictionRequestValidationError
- Could not create training predictions job. Request with same parameters already submitted. None

Response Headers

Status Header Type Format Description
202 Location string URL for tracking async job status.

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/projects/{projectId}/trainingPredictions/{predictionId}/

Retrieve training predictions that have previously been computed

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/trainingPredictions/{predictionId}/?offset=0&limit=0 \
  -H "Accept: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer true This many results will be skipped
limit query integer true At most this many results are returned
projectId path string true Project ID to retrieve training predictions for
predictionId path string true Prediction ID to retrieve training predictions for
Accept header string false Requested MIME type for the returned data

Enumerated Values

Parameter Value
Accept [application/json, text/csv]

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "forecastDistance": 0,
      "forecastPoint": "2019-08-24T14:15:22Z",
      "partitionId": "string",
      "prediction": 0,
      "predictionExplanations": [
        {
          "feature": "string",
          "featureValue": 0,
          "label": "string",
          "strength": 0
        }
      ],
      "predictionThreshold": 1,
      "predictionValues": [
        {
          "label": "string",
          "threshold": 1,
          "value": 0
        }
      ],
      "rowId": 0,
      "seriesId": "string",
      "shapMetadata": {
        "shapBaseValue": 0,
        "shapRemainingTotal": 0,
        "warnings": [
          {
            "maxNormalizedMismatch": 0,
            "mismatchRowCount": 0
          }
        ]
      },
      "timestamp": "2019-08-24T14:15:22Z"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com"
}

Responses

Status Meaning Description Schema
200 OK Training predictions encoded either as JSON or CSV string
404 Not Found Job does not exist or is not completed None

Response Headers

Status Header Type Format Description
200 Content-Type string MIME type of the returned data

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/scheduledJobs/

Get a list of scheduled batch prediction jobs a user can view

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/scheduledJobs/?offset=0&limit=20 \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer true The number of scheduled jobs to skip. Defaults to 0.
limit query integer true The number of scheduled jobs (max 100) to return. Defaults to 20
orderBy query string false The order to sort the scheduled jobs. Defaults to order by last successful run timestamp in descending order.
search query string false Case insensitive search against scheduled jobs name or type name.
deploymentId query string false Filter by the prediction integration deployment ID. Ignored for non prediction integration type ID.
typeId query string false filter by scheduled job type ID.
queryByUser query string false Which user field to filter with.
filterEnabled query string false Filter jobs using the enabled field. If true, only enabled jobs are returned, otherwise if false, only disabled jobs are returned. The default returns both enabled and disabled jobs.

Enumerated Values

Parameter Value
typeId datasetRefresh
queryByUser [createdBy, updatedBy]
filterEnabled [false, False, true, True]

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "createdBy": "string",
      "deploymentId": "string",
      "enabled": true,
      "id": "string",
      "name": "string",
      "schedule": {
        "dayOfMonth": [
          "*"
        ],
        "dayOfWeek": [
          "*"
        ],
        "hour": [
          "*"
        ],
        "minute": [
          "*"
        ],
        "month": [
          "*"
        ]
      },
      "scheduledJobId": "string",
      "status": {
        "lastFailedRun": "2019-08-24T14:15:22Z",
        "lastSuccessfulRun": "2019-08-24T14:15:22Z",
        "nextRunTime": "2019-08-24T14:15:22Z",
        "queuePosition": 0,
        "running": true
      },
      "typeId": "string",
      "updatedAt": "2019-08-24T14:15:22Z"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0,
  "updatedAt": "2019-08-24T14:15:22Z",
  "updatedBy": "string"
}

Responses

Status Meaning Description Schema
200 OK A list of scheduled batch prediction jobs ScheduledJobsListResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

Schemas

ActualValueColumnInfo

{
  "missingCount": 0,
  "name": "string"
}

Properties

Name Type Required Restrictions Description
missingCount integer true Count of the missing values in the column.
name string true Name of the column.

AzureDataStreamer

{
  "credentialId": "string",
  "format": "csv",
  "type": "azure",
  "url": "string"
}

Stream CSV data chunks from Azure

Properties

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false Use the specified credential to access the url

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
format string false Type of input file format
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
anonymous [redacted]
format [csv, parquet]
type azure

AzureIntake

{
  "credentialId": "string",
  "format": "csv",
  "type": "azure",
  "url": "string"
}

Stream CSV data chunks from Azure

Properties

Name Type Required Restrictions Description
credentialId string¦null false Use the specified credential to access the url
format string false Type of input file format
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
format [csv, parquet]
type azure

AzureOutput

{
  "credentialId": "string",
  "format": "csv",
  "partitionColumns": [
    "string"
  ],
  "type": "azure",
  "url": "string"
}

Save CSV data chunks to Azure Blob Storage

Properties

Name Type Required Restrictions Description
credentialId string¦null false Use the specified credential to access the url
format string false Type of output file format
partitionColumns [string] false maxItems: 100
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/").
type string true Type name for this output type
url string(url) true URL for the file or directory

Enumerated Values

Property Value
format [csv, parquet]
type azure

AzureOutputAdaptor

{
  "credentialId": "string",
  "format": "csv",
  "partitionColumns": [
    "string"
  ],
  "type": "azure",
  "url": "string"
}

Save CSV data chunks to Azure Blob Storage

Properties

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false Use the specified credential to access the url

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
format string false Type of output file format
partitionColumns [string] false maxItems: 100
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/").
type string true Type name for this output type
url string(url) true URL for the file or directory

Enumerated Values

Property Value
anonymous [redacted]
format [csv, parquet]
type azure

AzureServicePrincipalCredentials

{
  "azureTenantId": "string",
  "clientId": "string",
  "clientSecret": "string",
  "configId": "string",
  "credentialType": "azure_service_principal"
}

Properties

Name Type Required Restrictions Description
azureTenantId string false Tenant ID of the Azure AD service principal.
clientId string false Client ID of the Azure AD service principal.
clientSecret string false Client Secret of the Azure AD service principal.
configId string false ID of secure configurations of credentials shared by admin.
credentialType string true The type of these credentials, 'azure_service_principal' here.

Enumerated Values

Property Value
credentialType azure_service_principal

BasicCredentials

{
  "credentialType": "basic",
  "password": "string",
  "user": "string"
}

Properties

Name Type Required Restrictions Description
credentialType string true The type of these credentials, 'basic' here.
password string true The password for database authentication. The password is encrypted at rest and never saved / stored.
user string true The username for database authentication.

Enumerated Values

Property Value
credentialType basic

BatchJobCSVSettings

{
  "delimiter": ",",
  "encoding": "utf-8",
  "quotechar": "\""
}

The CSV settings used for this job

Properties

Name Type Required Restrictions Description
delimiter any true CSV fields are delimited by this character. Use the string "tab" to denote TSV (TAB separated values).

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous string false maxLength: 1
minLength: 1
minLength: 1
none

continued

Name Type Required Restrictions Description
encoding string true The encoding to be used for intake and output. For example (but not limited to): "shift_jis", "latin_1" or "mskanji".
quotechar string true maxLength: 1
minLength: 1
minLength: 1
Fields containing the delimiter or newlines must be quoted using this character.

Enumerated Values

Property Value
anonymous tab

BatchJobCreatedBy

{
  "fullName": "string",
  "userId": "string",
  "username": "string"
}

Who created this job

Properties

Name Type Required Restrictions Description
fullName string¦null true The full name of the user who created this job (if defined by the user)
userId string true The User ID of the user who created this job
username string true The username (e-mail address) of the user who created this job

BatchJobDefinitionResponse

{
  "createdBy": "string",
  "id": "string",
  "name": "string"
}

The Batch Prediction Job Definition linking to this job, if any.

Properties

Name Type Required Restrictions Description
createdBy string true The ID of creator of this job definition
id string true The ID of the Batch Prediction job definition
name string true A human-readable name for the definition, must be unique across organisations

{
  "csvUpload": "string",
  "download": "string",
  "self": "string"
}

Links useful for this job

Properties

Name Type Required Restrictions Description
csvUpload string(url) false The URL used to upload the dataset for this job. Only available for localFile intake.
download string¦null false The URL used to download the results from this job. Only available for localFile outputs. Will be null if the download is not yet available.
self string(url) true The URL used access this job.

BatchJobListResponse

{
  "count": 0,
  "data": [
    {
      "batchMonitoringJobDefinition": {
        "createdBy": "string",
        "id": "string",
        "name": "string"
      },
      "batchPredictionJobDefinition": {
        "createdBy": "string",
        "id": "string",
        "name": "string"
      },
      "created": "2019-08-24T14:15:22Z",
      "createdBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      },
      "elapsedTimeSec": 0,
      "failedRows": 0,
      "hidden": "2019-08-24T14:15:22Z",
      "id": "string",
      "intakeDatasetDisplayName": "string",
      "jobIntakeSize": 0,
      "jobOutputSize": 0,
      "jobSpec": {
        "abortOnError": true,
        "batchJobType": "monitoring",
        "chunkSize": "auto",
        "columnNamesRemapping": {},
        "csvSettings": {
          "delimiter": ",",
          "encoding": "utf-8",
          "quotechar": "\""
        },
        "deploymentId": "string",
        "disableRowLevelErrorHandling": false,
        "explanationAlgorithm": "shap",
        "explanationClassNames": [
          "string"
        ],
        "explanationNumTopClasses": 1,
        "includePredictionStatus": false,
        "includeProbabilities": true,
        "includeProbabilitiesClasses": [],
        "intakeSettings": {
          "type": "localFile"
        },
        "maxExplanations": 0,
        "maxNgramExplanations": 0,
        "modelId": "string",
        "modelPackageId": "string",
        "monitoringAggregation": {
          "retentionPolicy": "samples",
          "retentionValue": 0
        },
        "monitoringBatchPrefix": "string",
        "monitoringColumns": {
          "actedUponColumn": "string",
          "actualsTimestampColumn": "string",
          "actualsValueColumn": "string",
          "associationIdColumn": "string",
          "customMetricId": "string",
          "customMetricTimestampColumn": "string",
          "customMetricTimestampFormat": "string",
          "customMetricValueColumn": "string",
          "monitoredStatusColumn": "string",
          "predictionsColumns": [
            {
              "className": "string",
              "columnName": "string"
            }
          ],
          "reportDrift": true,
          "reportPredictions": true,
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "monitoringOutputSettings": {
          "monitoredStatusColumn": "string",
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "numConcurrent": 1,
        "outputSettings": {
          "credentialId": "string",
          "format": "csv",
          "partitionColumns": [
            "string"
          ],
          "type": "azure",
          "url": "string"
        },
        "passthroughColumns": [
          "string"
        ],
        "passthroughColumnsSet": "all",
        "pinnedModelId": "string",
        "predictionInstance": {
          "apiKey": "string",
          "datarobotKey": "string",
          "hostName": "string",
          "sslEnabled": true
        },
        "predictionWarningEnabled": true,
        "redactedFields": [
          "string"
        ],
        "skipDriftTracking": false,
        "thresholdHigh": 0,
        "thresholdLow": 0,
        "timeseriesSettings": {
          "forecastPoint": "2019-08-24T14:15:22Z",
          "relaxKnownInAdvanceFeaturesCheck": false,
          "type": "forecast"
        }
      },
      "links": {
        "csvUpload": "string",
        "download": "string",
        "self": "string"
      },
      "logs": [
        "string"
      ],
      "monitoringBatchId": "string",
      "percentageCompleted": 100,
      "queuePosition": 0,
      "queued": true,
      "resultsDeleted": true,
      "scoredRows": 0,
      "skippedRows": 0,
      "source": "string",
      "status": "INITIALIZING",
      "statusDetails": "string"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Properties

Name Type Required Restrictions Description
count integer false Number of items returned on this page.
data [BatchJobResponse] true maxItems: 10000
An array of jobs
next string(uri)¦null true URL pointing to the next page (if null, there is no next page).
previous string(uri)¦null true URL pointing to the previous page (if null, there is no previous page).
totalCount integer true The total number of items across all pages.

BatchJobPredictionInstance

{
  "apiKey": "string",
  "datarobotKey": "string",
  "hostName": "string",
  "sslEnabled": true
}

Override the default prediction instance from the deployment when scoring this job.

Properties

Name Type Required Restrictions Description
apiKey string false By default, prediction requests will use the API key of the user that created the job. This allows you to make requests on behalf of other users.
datarobotKey string false If running a job against a prediction instance in the Managed AI Cloud, you must provide the organization level DataRobot-Key.
hostName string true Override the default host name of the deployment with this.
sslEnabled boolean true Use SSL (HTTPS) when communicating with the overriden prediction server.

BatchJobRemapping

{
  "inputName": "string",
  "outputName": "string"
}

Properties

Name Type Required Restrictions Description
inputName string true Rename column with this name
outputName string¦null true Rename column to this name (leave as null to remove from the output)

BatchJobResponse

{
  "batchMonitoringJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "batchPredictionJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "elapsedTimeSec": 0,
  "failedRows": 0,
  "hidden": "2019-08-24T14:15:22Z",
  "id": "string",
  "intakeDatasetDisplayName": "string",
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "jobSpec": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 1,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "links": {
    "csvUpload": "string",
    "download": "string",
    "self": "string"
  },
  "logs": [
    "string"
  ],
  "monitoringBatchId": "string",
  "percentageCompleted": 100,
  "queuePosition": 0,
  "queued": true,
  "resultsDeleted": true,
  "scoredRows": 0,
  "skippedRows": 0,
  "source": "string",
  "status": "INITIALIZING",
  "statusDetails": "string"
}

Properties

Name Type Required Restrictions Description
batchMonitoringJobDefinition BatchJobDefinitionResponse false The Batch Prediction Job Definition linking to this job, if any.
batchPredictionJobDefinition BatchJobDefinitionResponse false The Batch Prediction Job Definition linking to this job, if any.
created string(date-time) true When was this job created
createdBy BatchJobCreatedBy true Who created this job
elapsedTimeSec integer true minimum: 0
Number of seconds the job has been processing for
failedRows integer true minimum: 0
Number of rows that have failed scoring
hidden string(date-time) false When was this job was hidden last, blank if visible
id string true The ID of the Batch job
intakeDatasetDisplayName string¦null false If applicable (e.g. for AI catalog), will contain the dataset name used for the intake dataset.
jobIntakeSize integer¦null true minimum: 0
Number of bytes in the intake dataset for this job
jobOutputSize integer¦null true minimum: 0
Number of bytes in the output dataset for this job
jobSpec BatchJobSpecResponse true The job configuration used to create this job
links BatchJobLinks true Links useful for this job
logs [string] true The job log.
monitoringBatchId string¦null true Id of the monitoring batch created by this job. Only present if the job runs on a deployment with batch monitoring enabled.
percentageCompleted number true maximum: 100
minimum: 0
Indicates job progress which is based on number of already processed rows in dataset
queuePosition integer¦null false minimum: 0
To ensure a dedicated prediction instance is not overloaded, only one job will be run against it at a time. This is the number of jobs that are awaiting processing before this job start running. May not be available in all environments.
queued boolean true The job has been put on the queue for execution.
resultsDeleted boolean false Indicates if the job was subject to garbage collection and had its artifacts deleted (output files, if any, and scoring data on local storage)
scoredRows integer true minimum: 0
Number of rows that have been used in prediction computation
skippedRows integer true minimum: 0
Number of rows that have been skipped during scoring. May contain non-zero value only in time-series predictions case if provided dataset contains more than required historical rows.
source string false Source from which batch job was started
status string true The current job status
statusDetails string true Explanation for current status

Enumerated Values

Property Value
status [INITIALIZING, RUNNING, COMPLETED, ABORTED, FAILED]

BatchJobSpecResponse

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "maxNgramExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionWarningEnabled": true,
  "redactedFields": [
    "string"
  ],
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

The job configuration used to create this job

Properties

Name Type Required Restrictions Description
abortOnError boolean true Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchJobCSVSettings true The CSV settings used for this job
deploymentId string false ID of deployment which is used in job for processing predictions dataset
disableRowLevelErrorHandling boolean true Skip row by row error handling
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean true Include prediction status column in the output
includeProbabilities boolean true Include probabilities for all classes
includeProbabilitiesClasses [string] true maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any true The response option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureDataStreamer false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous DataStageDataStreamer false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous CatalogDataStreamer false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous GCPDataStreamer false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryDataStreamer false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous S3DataStreamer false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeDataStreamer false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseDataStreamer false Stream CSV data chunks from Azure Synapse

xor

Name Type Required Restrictions Description
» anonymous DSSDataStreamer false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemDataStreamer false none

xor

Name Type Required Restrictions Description
» anonymous HTTPDataStreamer false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCDataStreamer false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileDataStreamer false Stream CSV data chunks from local file storage

continued

Name Type Required Restrictions Description
maxExplanations integer true maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
maxNgramExplanations any false The maximum number of text ngram explanations to supply per row of the dataset. The default recommended maxNgramExplanations is all (no limit)

oneOf

Name Type Required Restrictions Description
» anonymous integer false minimum: 0
none

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringAggregation MonitoringAggregation false Defines the aggregation policy for monitoring jobs.
monitoringBatchPrefix string¦null false Name of the batch to create with this job
monitoringColumns MonitoringColumnsMapping false Column names mapping for monitoring
monitoringOutputSettings MonitoringOutputSettings false Output settings for monitoring jobs
numConcurrent integer false minimum: 1
Number of simultaneous requests to run against the prediction instance
outputSettings any false The response option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutputAdaptor false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous GCPOutputAdaptor false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutputAdaptor false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous S3OutputAdaptor false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutputAdaptor false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutputAdaptor false Save CSV data chunks to Azure Synapse in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutputAdaptor false none

xor

Name Type Required Restrictions Description
» anonymous HttpOutputAdaptor false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JdbcOutputAdaptor false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutputAdaptor false Save CSV data chunks to local file storage

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionWarningEnabled boolean¦null false Enable prediction warnings.
redactedFields [string] true A list of qualified field names from intake- and/or outputSettings that was redacted due to permissions and sharing settings. For example: intakeSettings.dataStoreId
skipDriftTracking boolean true Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsHistorical false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
anonymous all
passthroughColumnsSet all

BatchJobTimeSeriesSettingsForecast

{
  "forecastPoint": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "forecast"
}

Properties

Name Type Required Restrictions Description
forecastPoint string(date-time) false Used for forecast predictions in order to override the inferred forecast point from the dataset.
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Forecast mode makes predictions using forecastPoint or rows in the dataset without target.

Enumerated Values

Property Value
type forecast

BatchJobTimeSeriesSettingsHistorical

{
  "predictionsEndDate": "2019-08-24T14:15:22Z",
  "predictionsStartDate": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "historical"
}

Properties

Name Type Required Restrictions Description
predictionsEndDate string(date-time) false Used for historical predictions in order to override date to which predictions should be calculated. By default value will be inferred automatically from the dataset.
predictionsStartDate string(date-time) false Used for historical predictions in order to override date from which predictions should be calculated. By default value will be inferred automatically from the dataset.
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Historical mode enables bulk predictions which calculates predictions for all possible forecast points and forecast distances in the dataset within the predictionsStartDate/predictionsEndDate range.

Enumerated Values

Property Value
type historical

BatchPredictionCreatedBy

{
  "fullName": "string",
  "userId": "string",
  "username": "string"
}

Who created this job

Properties

Name Type Required Restrictions Description
fullName string¦null true The full name of the user who created this job (if defined by the user)
userId string true The User ID of the user who created this job
username string true The username (e-mail address) of the user who created this job

BatchPredictionJobCSVSettings

{
  "delimiter": ",",
  "encoding": "utf-8",
  "quotechar": "\""
}

The CSV settings used for this job

Properties

Name Type Required Restrictions Description
delimiter any true CSV fields are delimited by this character. Use the string "tab" to denote TSV (TAB separated values).

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous string false maxLength: 1
minLength: 1
minLength: 1
none

continued

Name Type Required Restrictions Description
encoding string true The encoding to be used for intake and output. For example (but not limited to): "shift_jis", "latin_1" or "mskanji".
quotechar string true maxLength: 1
minLength: 1
minLength: 1
Fields containing the delimiter or newlines must be quoted using this character.

Enumerated Values

Property Value
anonymous tab

BatchPredictionJobCreate

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringBatchPrefix": "string",
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Properties

Name Type Required Restrictions Description
abortOnError boolean true Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchPredictionJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchPredictionJobCSVSettings true The CSV settings used for this job
deploymentId string false ID of deployment which is used in job for processing predictions dataset
disableRowLevelErrorHandling boolean true Skip row by row error handling
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean true Include prediction status column in the output
includeProbabilities boolean true Include probabilities for all classes
includeProbabilitiesClasses [string] true maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any true The intake option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureIntake false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous BigQueryIntake false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous DataStageIntake false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous Catalog false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous DSS false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemIntake false none

xor

Name Type Required Restrictions Description
» anonymous GCPIntake false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPIntake false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCIntake false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileIntake false Stream CSV data chunks from local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Intake false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeIntake false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseIntake false Stream CSV data chunks from Azure Synapse

continued

Name Type Required Restrictions Description
maxExplanations integer true maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringBatchPrefix string¦null false Name of the batch to create with this job
numConcurrent integer false minimum: 1
Number of simultaneous requests to run against the prediction instance
outputSettings any false The output option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutput false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutput false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutput false none

xor

Name Type Required Restrictions Description
» anonymous GCPOutput false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPOutput false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JDBCOutput false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutput false Save CSV data chunks to local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Output false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutput false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutput false Save CSV data chunks to Azure Synapse in bulk

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchPredictionJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionThreshold number false maximum: 1
minimum: 0
Threshold is the point that sets the class boundary for a predicted value. The model classifies an observation below the threshold as FALSE, and an observation above the threshold as TRUE. In other words, DataRobot automatically assigns the positive class label to any prediction exceeding the threshold. This value can be set between 0.0 and 1.0.
predictionWarningEnabled boolean¦null false Enable prediction warnings.
secondaryDatasetsConfigId string false Configuration id for secondary datasets to use when making a prediction.
skipDriftTracking boolean true Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsHistorical false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsTraining false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
passthroughColumnsSet all

BatchPredictionJobDefinitionId

{
  "jobDefinitionId": "string"
}

Properties

Name Type Required Restrictions Description
jobDefinitionId string true ID of the Batch Prediction job definition

BatchPredictionJobDefinitionJobSpecResponse

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "maxNgramExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "numConcurrent": 0,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionWarningEnabled": true,
  "redactedFields": [
    "string"
  ],
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

The Batch Prediction Job specification to be put on the queue in intervals

Properties

Name Type Required Restrictions Description
abortOnError boolean true Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchJobCSVSettings true The CSV settings used for this job
deploymentId string false ID of deployment which is used in job for processing predictions dataset
disableRowLevelErrorHandling boolean true Skip row by row error handling
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean true Include prediction status column in the output
includeProbabilities boolean true Include probabilities for all classes
includeProbabilitiesClasses [string] true maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any true The response option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureDataStreamer false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous DataStageDataStreamer false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous CatalogDataStreamer false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous GCPDataStreamer false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryDataStreamer false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous S3DataStreamer false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeDataStreamer false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseDataStreamer false Stream CSV data chunks from Azure Synapse

xor

Name Type Required Restrictions Description
» anonymous DSSDataStreamer false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemDataStreamer false none

xor

Name Type Required Restrictions Description
» anonymous HTTPDataStreamer false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCDataStreamer false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileDataStreamer false Stream CSV data chunks from local file storage

continued

Name Type Required Restrictions Description
maxExplanations integer true maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
maxNgramExplanations any false The maximum number of text ngram explanations to supply per row of the dataset. The default recommended maxNgramExplanations is all (no limit)

oneOf

Name Type Required Restrictions Description
» anonymous integer false minimum: 0
none

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringAggregation MonitoringAggregation false Defines the aggregation policy for monitoring jobs.
monitoringBatchPrefix string¦null false Name of the batch to create with this job
monitoringColumns MonitoringColumnsMapping false Column names mapping for monitoring
monitoringOutputSettings MonitoringOutputSettings false Output settings for monitoring jobs
numConcurrent integer true minimum: 0
Number of simultaneous requests to run against the prediction instance
outputSettings any false The response option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutputAdaptor false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous GCPOutputAdaptor false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutputAdaptor false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous S3OutputAdaptor false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutputAdaptor false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutputAdaptor false Save CSV data chunks to Azure Synapse in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutputAdaptor false none

xor

Name Type Required Restrictions Description
» anonymous HttpOutputAdaptor false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JdbcOutputAdaptor false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutputAdaptor false Save CSV data chunks to local file storage

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionWarningEnabled boolean¦null false Enable prediction warnings.
redactedFields [string] true A list of qualified field names from intake- and/or outputSettings that was redacted due to permissions and sharing settings. For example: intakeSettings.dataStoreId
skipDriftTracking boolean true Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsForecastWithPolicy false none

xor

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsHistorical false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
anonymous all
passthroughColumnsSet all

BatchPredictionJobDefinitionResponse

{
  "createdBy": "string",
  "id": "string",
  "name": "string"
}

The Batch Prediction Job Definition linking to this job, if any.

Properties

Name Type Required Restrictions Description
createdBy string true The ID of creator of this job definition
id string true The ID of the Batch Prediction job definition
name string true A human-readable name for the definition, must be unique across organisations

BatchPredictionJobDefinitionsCreate

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "enabled": true,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringBatchPrefix": "string",
  "name": "string",
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Properties

Name Type Required Restrictions Description
abortOnError boolean true Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchPredictionJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchPredictionJobCSVSettings true The CSV settings used for this job
deploymentId string true ID of deployment which is used in job for processing predictions dataset
disableRowLevelErrorHandling boolean true Skip row by row error handling
enabled boolean false If this job definition is enabled as a scheduled job. Optional if no schedule is supplied.
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean true Include prediction status column in the output
includeProbabilities boolean true Include probabilities for all classes
includeProbabilitiesClasses [string] true maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any true The intake option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureIntake false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous BigQueryIntake false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous DataStageIntake false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous Catalog false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous DSS false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemIntake false none

xor

Name Type Required Restrictions Description
» anonymous GCPIntake false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPIntake false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCIntake false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileIntake false Stream CSV data chunks from local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Intake false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeIntake false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseIntake false Stream CSV data chunks from Azure Synapse

continued

Name Type Required Restrictions Description
maxExplanations integer true maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringBatchPrefix string¦null false Name of the batch to create with this job
name string false maxLength: 100
minLength: 1
minLength: 1
A human-readable name for the definition, must be unique across organisations, if left out the backend will generate one for you.
numConcurrent integer false minimum: 1
Number of simultaneous requests to run against the prediction instance
outputSettings any false The output option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutput false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutput false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutput false none

xor

Name Type Required Restrictions Description
» anonymous GCPOutput false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPOutput false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JDBCOutput false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutput false Save CSV data chunks to local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Output false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutput false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutput false Save CSV data chunks to Azure Synapse in bulk

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchPredictionJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionThreshold number false maximum: 1
minimum: 0
Threshold is the point that sets the class boundary for a predicted value. The model classifies an observation below the threshold as FALSE, and an observation above the threshold as TRUE. In other words, DataRobot automatically assigns the positive class label to any prediction exceeding the threshold. This value can be set between 0.0 and 1.0.
predictionWarningEnabled boolean¦null false Enable prediction warnings.
schedule Schedule false The scheduling information defining how often and when to execute this job to the Job Scheduling service. Optional if enabled = False.
secondaryDatasetsConfigId string false Configuration id for secondary datasets to use when making a prediction.
skipDriftTracking boolean true Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsForecastWithPolicy false none

xor

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsHistorical false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
passthroughColumnsSet all

BatchPredictionJobDefinitionsListResponse

{
  "count": 0,
  "data": [
    {
      "batchPredictionJob": {
        "abortOnError": true,
        "batchJobType": "monitoring",
        "chunkSize": "auto",
        "columnNamesRemapping": {},
        "csvSettings": {
          "delimiter": ",",
          "encoding": "utf-8",
          "quotechar": "\""
        },
        "deploymentId": "string",
        "disableRowLevelErrorHandling": false,
        "explanationAlgorithm": "shap",
        "explanationClassNames": [
          "string"
        ],
        "explanationNumTopClasses": 1,
        "includePredictionStatus": false,
        "includeProbabilities": true,
        "includeProbabilitiesClasses": [],
        "intakeSettings": {
          "type": "localFile"
        },
        "maxExplanations": 0,
        "maxNgramExplanations": 0,
        "modelId": "string",
        "modelPackageId": "string",
        "monitoringAggregation": {
          "retentionPolicy": "samples",
          "retentionValue": 0
        },
        "monitoringBatchPrefix": "string",
        "monitoringColumns": {
          "actedUponColumn": "string",
          "actualsTimestampColumn": "string",
          "actualsValueColumn": "string",
          "associationIdColumn": "string",
          "customMetricId": "string",
          "customMetricTimestampColumn": "string",
          "customMetricTimestampFormat": "string",
          "customMetricValueColumn": "string",
          "monitoredStatusColumn": "string",
          "predictionsColumns": [
            {
              "className": "string",
              "columnName": "string"
            }
          ],
          "reportDrift": true,
          "reportPredictions": true,
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "monitoringOutputSettings": {
          "monitoredStatusColumn": "string",
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "numConcurrent": 0,
        "outputSettings": {
          "credentialId": "string",
          "format": "csv",
          "partitionColumns": [
            "string"
          ],
          "type": "azure",
          "url": "string"
        },
        "passthroughColumns": [
          "string"
        ],
        "passthroughColumnsSet": "all",
        "pinnedModelId": "string",
        "predictionInstance": {
          "apiKey": "string",
          "datarobotKey": "string",
          "hostName": "string",
          "sslEnabled": true
        },
        "predictionWarningEnabled": true,
        "redactedFields": [
          "string"
        ],
        "skipDriftTracking": false,
        "thresholdHigh": 0,
        "thresholdLow": 0,
        "timeseriesSettings": {
          "forecastPoint": "2019-08-24T14:15:22Z",
          "relaxKnownInAdvanceFeaturesCheck": false,
          "type": "forecast"
        }
      },
      "created": "2019-08-24T14:15:22Z",
      "createdBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      },
      "enabled": false,
      "id": "string",
      "lastFailedRunTime": "2019-08-24T14:15:22Z",
      "lastScheduledRunTime": "2019-08-24T14:15:22Z",
      "lastStartedJobStatus": "INITIALIZING",
      "lastStartedJobTime": "2019-08-24T14:15:22Z",
      "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
      "name": "string",
      "nextScheduledRunTime": "2019-08-24T14:15:22Z",
      "schedule": {
        "dayOfMonth": [
          "*"
        ],
        "dayOfWeek": [
          "*"
        ],
        "hour": [
          "*"
        ],
        "minute": [
          "*"
        ],
        "month": [
          "*"
        ]
      },
      "updated": "2019-08-24T14:15:22Z",
      "updatedBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      }
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Properties

Name Type Required Restrictions Description
count integer false Number of items returned on this page.
data [BatchPredictionJobDefinitionsResponse] true An array of scheduled jobs
next string(uri)¦null true URL pointing to the next page (if null, there is no next page).
previous string(uri)¦null true URL pointing to the previous page (if null, there is no previous page).
totalCount integer true The total number of items across all pages.

BatchPredictionJobDefinitionsResponse

{
  "batchPredictionJob": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 0,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "enabled": false,
  "id": "string",
  "lastFailedRunTime": "2019-08-24T14:15:22Z",
  "lastScheduledRunTime": "2019-08-24T14:15:22Z",
  "lastStartedJobStatus": "INITIALIZING",
  "lastStartedJobTime": "2019-08-24T14:15:22Z",
  "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
  "name": "string",
  "nextScheduledRunTime": "2019-08-24T14:15:22Z",
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "updated": "2019-08-24T14:15:22Z",
  "updatedBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  }
}

Properties

Name Type Required Restrictions Description
batchPredictionJob BatchPredictionJobDefinitionJobSpecResponse true The Batch Prediction Job specification to be put on the queue in intervals
created string(date-time) true When was this job created
createdBy BatchJobCreatedBy true Who created this job
enabled boolean true If this job definition is enabled as a scheduled job.
id string true The ID of the Batch job definition
lastFailedRunTime string(date-time)¦null false Last time this job had a failed run
lastScheduledRunTime string(date-time)¦null false Last time this job was scheduled to run (though not guaranteed it actually ran at that time)
lastStartedJobStatus string¦null true The status of the latest job launched to the queue (if any).
lastStartedJobTime string(date-time)¦null true The last time (if any) a job was launched.
lastSuccessfulRunTime string(date-time)¦null false Last time this job had a successful run
name string true A human-readable name for the definition, must be unique across organisations
nextScheduledRunTime string(date-time)¦null false Next time this job is scheduled to run
schedule Schedule false The scheduling information defining how often and when to execute this job to the Job Scheduling service. Optional if enabled = False.
updated string(date-time) true When was this job last updated
updatedBy BatchJobCreatedBy true Who created this job

Enumerated Values

Property Value
lastStartedJobStatus [INITIALIZING, RUNNING, COMPLETED, ABORTED, FAILED]

BatchPredictionJobDefinitionsUpdate

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "enabled": true,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringBatchPrefix": "string",
  "name": "string",
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Properties

Name Type Required Restrictions Description
abortOnError boolean false Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchPredictionJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchPredictionJobCSVSettings false The CSV settings used for this job
deploymentId string false ID of deployment which is used in job for processing predictions dataset
disableRowLevelErrorHandling boolean false Skip row by row error handling
enabled boolean false If this job definition is enabled as a scheduled job. Optional if no schedule is supplied.
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean false Include prediction status column in the output
includeProbabilities boolean false Include probabilities for all classes
includeProbabilitiesClasses [string] false maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any false The intake option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureIntake false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous BigQueryIntake false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous DataStageIntake false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous Catalog false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous DSS false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemIntake false none

xor

Name Type Required Restrictions Description
» anonymous GCPIntake false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPIntake false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCIntake false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileIntake false Stream CSV data chunks from local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Intake false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeIntake false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseIntake false Stream CSV data chunks from Azure Synapse

continued

Name Type Required Restrictions Description
maxExplanations integer false maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringBatchPrefix string¦null false Name of the batch to create with this job
name string false maxLength: 100
minLength: 1
minLength: 1
A human-readable name for the definition, must be unique across organisations, if left out the backend will generate one for you.
numConcurrent integer false minimum: 1
Number of simultaneous requests to run against the prediction instance
outputSettings any false The output option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutput false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutput false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutput false none

xor

Name Type Required Restrictions Description
» anonymous GCPOutput false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPOutput false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JDBCOutput false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutput false Save CSV data chunks to local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Output false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutput false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutput false Save CSV data chunks to Azure Synapse in bulk

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchPredictionJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionThreshold number false maximum: 1
minimum: 0
Threshold is the point that sets the class boundary for a predicted value. The model classifies an observation below the threshold as FALSE, and an observation above the threshold as TRUE. In other words, DataRobot automatically assigns the positive class label to any prediction exceeding the threshold. This value can be set between 0.0 and 1.0.
predictionWarningEnabled boolean¦null false Enable prediction warnings.
schedule Schedule false The scheduling information defining how often and when to execute this job to the Job Scheduling service. Optional if enabled = False.
secondaryDatasetsConfigId string false Configuration id for secondary datasets to use when making a prediction.
skipDriftTracking boolean false Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsForecastWithPolicy false none

xor

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsHistorical false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
passthroughColumnsSet all

BatchPredictionJobId

{
  "partNumber": 0,
  "predictionJobId": "string"
}

Properties

Name Type Required Restrictions Description
partNumber integer true minimum: 0
The number of which csv part is being uploaded when using multipart upload
predictionJobId string true ID of the Batch Prediction job

{
  "csvUpload": "string",
  "download": "string",
  "self": "string"
}

Links useful for this job

Properties

Name Type Required Restrictions Description
csvUpload string(url) false The URL used to upload the dataset for this job. Only available for localFile intake.
download string¦null false The URL used to download the results from this job. Only available for localFile outputs. Will be null if the download is not yet available.
self string(url) true The URL used access this job.

BatchPredictionJobListResponse

{
  "count": 0,
  "data": [
    {
      "batchPredictionJobDefinition": {
        "createdBy": "string",
        "id": "string",
        "name": "string"
      },
      "created": "2019-08-24T14:15:22Z",
      "createdBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      },
      "elapsedTimeSec": 0,
      "failedRows": 0,
      "hidden": "2019-08-24T14:15:22Z",
      "id": "string",
      "intakeDatasetDisplayName": "string",
      "jobIntakeSize": 0,
      "jobOutputSize": 0,
      "jobSpec": {
        "abortOnError": true,
        "batchJobType": "monitoring",
        "chunkSize": "auto",
        "columnNamesRemapping": {},
        "csvSettings": {
          "delimiter": ",",
          "encoding": "utf-8",
          "quotechar": "\""
        },
        "deploymentId": "string",
        "disableRowLevelErrorHandling": false,
        "explanationAlgorithm": "shap",
        "explanationClassNames": [
          "string"
        ],
        "explanationNumTopClasses": 1,
        "includePredictionStatus": false,
        "includeProbabilities": true,
        "includeProbabilitiesClasses": [],
        "intakeSettings": {
          "type": "localFile"
        },
        "maxExplanations": 0,
        "modelId": "string",
        "modelPackageId": "string",
        "monitoringBatchPrefix": "string",
        "numConcurrent": 1,
        "outputSettings": {
          "credentialId": "string",
          "format": "csv",
          "partitionColumns": [
            "string"
          ],
          "type": "azure",
          "url": "string"
        },
        "passthroughColumns": [
          "string"
        ],
        "passthroughColumnsSet": "all",
        "pinnedModelId": "string",
        "predictionInstance": {
          "apiKey": "string",
          "datarobotKey": "string",
          "hostName": "string",
          "sslEnabled": true
        },
        "predictionThreshold": 1,
        "predictionWarningEnabled": true,
        "redactedFields": [
          "string"
        ],
        "secondaryDatasetsConfigId": "string",
        "skipDriftTracking": false,
        "thresholdHigh": 0,
        "thresholdLow": 0,
        "timeseriesSettings": {
          "forecastPoint": "2019-08-24T14:15:22Z",
          "relaxKnownInAdvanceFeaturesCheck": false,
          "type": "forecast"
        }
      },
      "links": {
        "csvUpload": "string",
        "download": "string",
        "self": "string"
      },
      "logs": [
        "string"
      ],
      "monitoringBatchId": "string",
      "percentageCompleted": 100,
      "queuePosition": 0,
      "queued": true,
      "resultsDeleted": true,
      "scoredRows": 0,
      "skippedRows": 0,
      "source": "string",
      "status": "INITIALIZING",
      "statusDetails": "string"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Properties

Name Type Required Restrictions Description
count integer false Number of items returned on this page.
data [BatchPredictionJobResponse] true An array of jobs
next string(uri)¦null true URL pointing to the next page (if null, there is no next page).
previous string(uri)¦null true URL pointing to the previous page (if null, there is no previous page).
totalCount integer true The total number of items across all pages.

BatchPredictionJobPredictionInstance

{
  "apiKey": "string",
  "datarobotKey": "string",
  "hostName": "string",
  "sslEnabled": true
}

Override the default prediction instance from the deployment when scoring this job.

Properties

Name Type Required Restrictions Description
apiKey string false By default, prediction requests will use the API key of the user that created the job. This allows you to make requests on behalf of other users.
datarobotKey string false If running a job against a prediction instance in the Managed AI Cloud, you must provide the organization level DataRobot-Key.
hostName string true Override the default host name of the deployment with this.
sslEnabled boolean true Use SSL (HTTPS) when communicating with the overriden prediction server.

BatchPredictionJobRemapping

{
  "inputName": "string",
  "outputName": "string"
}

Properties

Name Type Required Restrictions Description
inputName string true Rename column with this name
outputName string¦null true Rename column to this name (leave as null to remove from the output)

BatchPredictionJobResponse

{
  "batchPredictionJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "elapsedTimeSec": 0,
  "failedRows": 0,
  "hidden": "2019-08-24T14:15:22Z",
  "id": "string",
  "intakeDatasetDisplayName": "string",
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "jobSpec": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringBatchPrefix": "string",
    "numConcurrent": 1,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionThreshold": 1,
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "secondaryDatasetsConfigId": "string",
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "links": {
    "csvUpload": "string",
    "download": "string",
    "self": "string"
  },
  "logs": [
    "string"
  ],
  "monitoringBatchId": "string",
  "percentageCompleted": 100,
  "queuePosition": 0,
  "queued": true,
  "resultsDeleted": true,
  "scoredRows": 0,
  "skippedRows": 0,
  "source": "string",
  "status": "INITIALIZING",
  "statusDetails": "string"
}

Properties

Name Type Required Restrictions Description
batchPredictionJobDefinition BatchPredictionJobDefinitionResponse false The Batch Prediction Job Definition linking to this job, if any.
created string(date-time) true When was this job created
createdBy BatchPredictionCreatedBy true Who created this job
elapsedTimeSec integer true minimum: 0
Number of seconds the job has been processing for
failedRows integer true minimum: 0
Number of rows that have failed scoring
hidden string(date-time) false When was this job was hidden last, blank if visible
id string true The ID of the Batch Prediction job
intakeDatasetDisplayName string¦null false If applicable (e.g. for AI catalog), will contain the dataset name used for the intake dataset.
jobIntakeSize integer¦null true minimum: 0
Number of bytes in the intake dataset for this job
jobOutputSize integer¦null true minimum: 0
Number of bytes in the output dataset for this job
jobSpec BatchPredictionJobSpecResponse true The job configuration used to create this job
links BatchPredictionJobLinks true Links useful for this job
logs [string] true The job log.
monitoringBatchId string¦null true Id of the monitoring batch created by this job. Only present if the job runs on a deployment with batch monitoring enabled.
percentageCompleted number true maximum: 100
minimum: 0
Indicates job progress which is based on number of already processed rows in dataset
queuePosition integer¦null false minimum: 0
To ensure a dedicated prediction instance is not overloaded, only one job will be run against it at a time. This is the number of jobs that are awaiting processing before this job start running. May not be available in all environments.
queued boolean true The job has been put on the queue for execution.
resultsDeleted boolean false Indicates if the job was subject to garbage collection and had its artifacts deleted (output files, if any, and scoring data on local storage)
scoredRows integer true minimum: 0
Number of rows that have been used in prediction computation
skippedRows integer true minimum: 0
Number of rows that have been skipped during scoring. May contain non-zero value only in time-series predictions case if provided dataset contains more than required historical rows.
source string false Source from which batch job was started
status string true The current job status
statusDetails string true Explanation for current status

Enumerated Values

Property Value
status [INITIALIZING, RUNNING, COMPLETED, ABORTED, FAILED]

BatchPredictionJobSpecResponse

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringBatchPrefix": "string",
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "redactedFields": [
    "string"
  ],
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

The job configuration used to create this job

Properties

Name Type Required Restrictions Description
abortOnError boolean true Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchPredictionJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchPredictionJobCSVSettings true The CSV settings used for this job
deploymentId string false ID of deployment which is used in job for processing predictions dataset
disableRowLevelErrorHandling boolean true Skip row by row error handling
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean true Include prediction status column in the output
includeProbabilities boolean true Include probabilities for all classes
includeProbabilitiesClasses [string] true maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any true The response option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureDataStreamer false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous DataStageDataStreamer false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous CatalogDataStreamer false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous GCPDataStreamer false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryDataStreamer false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous S3DataStreamer false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeDataStreamer false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseDataStreamer false Stream CSV data chunks from Azure Synapse

xor

Name Type Required Restrictions Description
» anonymous DSSDataStreamer false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemDataStreamer false none

xor

Name Type Required Restrictions Description
» anonymous HTTPDataStreamer false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCDataStreamer false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileDataStreamer false Stream CSV data chunks from local file storage

continued

Name Type Required Restrictions Description
maxExplanations integer true maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringBatchPrefix string¦null false Name of the batch to create with this job
numConcurrent integer false minimum: 1
Number of simultaneous requests to run against the prediction instance
outputSettings any false The response option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutputAdaptor false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous GCPOutputAdaptor false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutputAdaptor false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous S3OutputAdaptor false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutputAdaptor false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutputAdaptor false Save CSV data chunks to Azure Synapse in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutputAdaptor false none

xor

Name Type Required Restrictions Description
» anonymous HttpOutputAdaptor false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JdbcOutputAdaptor false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutputAdaptor false Save CSV data chunks to local file storage

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchPredictionJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionThreshold number false maximum: 1
minimum: 0
Threshold is the point that sets the class boundary for a predicted value. The model classifies an observation below the threshold as FALSE, and an observation above the threshold as TRUE. In other words, DataRobot automatically assigns the positive class label to any prediction exceeding the threshold. This value can be set between 0.0 and 1.0.
predictionWarningEnabled boolean¦null false Enable prediction warnings.
redactedFields [string] true A list of qualified field names from intake- and/or outputSettings that was redacted due to permissions and sharing settings. For example: intakeSettings.dataStoreId
secondaryDatasetsConfigId string false Configuration id for secondary datasets to use when making a prediction.
skipDriftTracking boolean true Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsHistorical false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsTraining false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
passthroughColumnsSet all

BatchPredictionJobTimeSeriesSettingsForecast

{
  "forecastPoint": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "forecast"
}

Properties

Name Type Required Restrictions Description
forecastPoint string(date-time) false Used for forecast predictions in order to override the inferred forecast point from the dataset.
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Forecast mode makes predictions using forecastPoint or rows in the dataset without target.

Enumerated Values

Property Value
type forecast

BatchPredictionJobTimeSeriesSettingsForecastWithPolicy

{
  "forecastPointPolicy": {
    "configuration": {
      "offset": "string"
    },
    "type": "jobRunTimeBased"
  },
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "forecast"
}

Properties

Name Type Required Restrictions Description
forecastPointPolicy JobRunTimeBasedForecastPointPolicy true Forecast point policy
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Forecast mode makes predictions using forecastPoint or rows in the dataset without target.

Enumerated Values

Property Value
type forecast

BatchPredictionJobTimeSeriesSettingsHistorical

{
  "predictionsEndDate": "2019-08-24T14:15:22Z",
  "predictionsStartDate": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "historical"
}

Properties

Name Type Required Restrictions Description
predictionsEndDate string(date-time) false Used for historical predictions in order to override date to which predictions should be calculated. By default value will be inferred automatically from the dataset.
predictionsStartDate string(date-time) false Used for historical predictions in order to override date from which predictions should be calculated. By default value will be inferred automatically from the dataset.
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Historical mode enables bulk predictions which calculates predictions for all possible forecast points and forecast distances in the dataset within the predictionsStartDate/predictionsEndDate range.

Enumerated Values

Property Value
type historical

BatchPredictionJobTimeSeriesSettingsTraining

{
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "training"
}

Properties

Name Type Required Restrictions Description
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Forecast mode used for making predictions on subsets of training data.

Enumerated Values

Property Value
type training

BatchPredictionJobUpdate

{
  "aborted": "2019-08-24T14:15:22Z",
  "completed": "2019-08-24T14:15:22Z",
  "failedRows": 0,
  "hidden": true,
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "logs": [
    "string"
  ],
  "scoredRows": 0,
  "skippedRows": 0,
  "started": "2019-08-24T14:15:22Z",
  "status": "INITIALIZING"
}

Properties

Name Type Required Restrictions Description
aborted string(date-time)¦null false Time when job abortion happened
completed string(date-time)¦null false Time when job completed scoring
failedRows integer false Number of rows that have failed scoring
hidden boolean false Hides or unhides the job from the job list
jobIntakeSize integer¦null false Number of bytes in the intake dataset for this job
jobOutputSize integer¦null false Number of bytes in the output dataset for this job
logs [string] false The job log.
scoredRows integer false Number of rows that have been used in prediction computation
skippedRows integer false Number of rows that have been skipped during scoring. May contain non-zero value only in time-series predictions case if provided dataset contains more than required historical rows.
started string(date-time)¦null false Time when job scoring begin
status string false The current job status

Enumerated Values

Property Value
status [INITIALIZING, RUNNING, COMPLETED, ABORTED, FAILED]

BigQueryDataStreamer

{
  "bucket": "string",
  "credentialId": "string",
  "dataset": "string",
  "table": "string",
  "type": "bigquery"
}

Stream CSV data chunks from Big Query using GCS

Properties

Name Type Required Restrictions Description
bucket string true The name of gcs bucket for data export
credentialId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false The ID of the GCP credentials

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataset string true The name of the specified big query dataset to read input data from
table string true The name of the specified big query table to read input data from
type string true Type name for this intake type

Enumerated Values

Property Value
anonymous [redacted]
type bigquery

BigQueryIntake

{
  "bucket": "string",
  "credentialId": "string",
  "dataset": "string",
  "table": "string",
  "type": "bigquery"
}

Stream CSV data chunks from Big Query using GCS

Properties

Name Type Required Restrictions Description
bucket string true The name of gcs bucket for data export
credentialId string true The ID of the GCP credentials
dataset string true The name of the specified big query dataset to read input data from
table string true The name of the specified big query table to read input data from
type string true Type name for this intake type

Enumerated Values

Property Value
type bigquery

BigQueryOutput

{
  "bucket": "string",
  "credentialId": "string",
  "dataset": "string",
  "table": "string",
  "type": "bigquery"
}

Save CSV data chunks to Google BigQuery in bulk

Properties

Name Type Required Restrictions Description
bucket string true The name of gcs bucket for data loading
credentialId string true The ID of the GCP credentials
dataset string true The name of the specified big query dataset to write data back
table string true The name of the specified big query table to write data back
type string true Type name for this output type

Enumerated Values

Property Value
type bigquery

BigQueryOutputAdaptor

{
  "bucket": "string",
  "credentialId": "string",
  "dataset": "string",
  "table": "string",
  "type": "bigquery"
}

Save CSV data chunks to Google BigQuery in bulk

Properties

Name Type Required Restrictions Description
bucket string true The name of gcs bucket for data loading
credentialId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false The ID of the GCP credentials

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataset string true The name of the specified big query dataset to write data back
table string true The name of the specified big query table to write data back
type string true Type name for this output type

Enumerated Values

Property Value
anonymous [redacted]
type bigquery

Catalog

{
  "datasetId": "string",
  "datasetVersionId": "string",
  "type": "dataset"
}

Stream CSV data chunks from AI catalog dataset

Properties

Name Type Required Restrictions Description
datasetId string true The ID of the AI catalog dataset
datasetVersionId string false The ID of the AI catalog dataset version
type string true Type name for this intake type

Enumerated Values

Property Value
type dataset

CatalogDataStreamer

{
  "datasetId": "string",
  "datasetVersionId": "string",
  "type": "dataset"
}

Stream CSV data chunks from AI catalog dataset

Properties

Name Type Required Restrictions Description
datasetId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false The ID of the AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
datasetVersionId string false The ID of the AI catalog dataset version
type string true Type name for this intake type

Enumerated Values

Property Value
anonymous [redacted]
type dataset

CreatePredictionDatasetResponse

{
  "datasetId": "string"
}

Properties

Name Type Required Restrictions Description
datasetId string true The ID of the newly created prediction dataset.

CreatePredictionFromDataset

{
  "actualValueColumn": "string",
  "datasetId": "string",
  "explanationAlgorithm": "shap",
  "forecastPoint": "2019-08-24T14:15:22Z",
  "includeFdwCounts": false,
  "includePredictionIntervals": true,
  "maxExplanations": 1,
  "modelId": "string",
  "predictionIntervalsSize": 1,
  "predictionThreshold": 1,
  "predictionsEndDate": "2019-08-24T14:15:22Z",
  "predictionsStartDate": "2019-08-24T14:15:22Z"
}

Properties

Name Type Required Restrictions Description
actualValueColumn string false For time series projects only. Actual value column name, valid for the prediction files if the project is unsupervised and the dataset is considered as bulk predictions dataset. This value is optional.
datasetId string true The dataset to compute predictions for - must have previously been uploaded.
explanationAlgorithm string false If set to shap, the response will include prediction explanations based on the SHAP explainer (SHapley Additive exPlanations). Defaults to null (no prediction explanations).
forecastPoint string(date-time) false For time series projects only. The time in the dataset relative to which predictions are generated. This value is optional. If not specified the default value is the value in the row with the latest specified timestamp. Specifying this value for a project that is not a time series project will result in an error.
includeFdwCounts boolean false For time series projects with partial history only. Indicates if feature derivation window counts featureDerivationWindowCounts will be part of the response.
includePredictionIntervals boolean false Specifies whether prediction intervals should be calculated for this request. Defaults to True if predictionIntervalsSize is specified, otherwise defaults to False.
maxExplanations integer false maximum: 100
minimum: 1
Specifies the maximum number of explanation values that should be returned for each row, ordered by absolute value, greatest to least. In the case of 'shap': If not set, explanations are returned for all features. If the number of features is greater than the 'maxExplanations', the sum of remaining values will also be returned as 'shapRemainingTotal'. Defaults to null for datasets narrower than 100 columns, defaults to 100 for datasets wider than 100 columns. Cannot be set if 'explanationAlgorithm' is omitted.
modelId string true The model to make predictions on.
predictionIntervalsSize integer false maximum: 100
minimum: 1
Represents the percentile to use for the size of the prediction intervals. Defaults to 80 if includePredictionIntervals is True.
predictionThreshold number false maximum: 1
minimum: 0
Threshold used for binary classification in predictions. Accepts values from 0.0 to 1.0. If not specified, model default prediction threshold will be used.
predictionsEndDate string(date-time) false The end date for bulk predictions, exclusive. Used for time series projects only. Note that this parameter is used for generating historical predictions using the training data, not for future predictions. If not specified, the dataset is not considered as a bulk predictions dataset. This parameter should be provided in conjunction with a predictionsStartDate, and cannot be provided with the forecastPoint parameter.
predictionsStartDate string(date-time) false The start date for bulk predictions. Used for time series projects only. Note that this parameter is used for generating historical predictions using the training data, not for future predictions. If not specified, the dataset is not considered as a bulk predictions dataset. This parameter should be provided in conjunction with a predictionsEndDate, and cannot be provided with the forecastPoint parameter.

Enumerated Values

Property Value
explanationAlgorithm shap

CreateTrainingPrediction

{
  "dataSubset": "all",
  "explanationAlgorithm": "string",
  "maxExplanations": 1,
  "modelId": "string"
}

Properties

Name Type Required Restrictions Description
dataSubset string true Subset of data predicted on: The value "all" returns predictions for all rows in the dataset including data used for training, validation, holdout and any rows discarded. This is not available for large datasets or projects created with Date/Time partitioning. The value "validationAndHoldout" returns predictions for the rows used to calculate the validation score and the holdout score. Not available for large projects or Date/Time projects for models trained into the validation set. The value "holdout" returns predictions for the rows used to calculate the holdout score. Not available for projects created without a holdout or for models trained into holdout for large datasets or created with Date/Time partitioning. The value "allBacktests" returns predictions for the rows used to calculate the backtesting scores for Date/Time projects. The value "validation" returns predictions for the rows used to calculate the validation score.
explanationAlgorithm string false If set to "shap", the response will include prediction explanations based on the SHAP explainer (SHapley Additive exPlanations). Defaults to null (no prediction explanations)
maxExplanations integer false maximum: 100
minimum: 1
Specifies the maximum number of explanation values that should be returned for each row, ordered by absolute value, greatest to least. In the case of "shap": If not set, explanations are returned for all features. If the number of features is greater than the "maxExplanations", the sum of remaining values will also be returned as "shapRemainingTotal". Defaults to null for datasets narrower than 100 columns, defaults to 100 for datasets wider than 100 columns. Cannot be set if "explanationAlgorithm" is omitted.
modelId string true The model to make predictions on

Enumerated Values

Property Value
dataSubset [all, validationAndHoldout, holdout, allBacktests, validation, crossValidation]

CredentialId

{
  "catalogVersionId": "string",
  "credentialId": "string",
  "url": "string"
}

Properties

Name Type Required Restrictions Description
catalogVersionId string false The ID of the latest version of the catalog entry.
credentialId string true The ID of the set of credentials to use instead of user and password. Note that with this change, username and password will become optional.
url string false The link to retrieve more detailed information about the entity that uses this catalog dataset.

DSS

{
  "datasetId": "string",
  "partition": "holdout",
  "projectId": "string",
  "type": "dss"
}

Stream CSV data chunks from DSS dataset

Properties

Name Type Required Restrictions Description
datasetId string false The ID of the dataset
partition string false Partition used to predict
projectId string true The ID of the project
type string true Type name for this intake type

Enumerated Values

Property Value
partition [holdout, validation, allBacktests, None]
type dss

DSSDataStreamer

{
  "datasetId": "string",
  "partition": "holdout",
  "projectId": "string",
  "type": "dss"
}

Stream CSV data chunks from DSS dataset

Properties

Name Type Required Restrictions Description
datasetId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false The ID of the dataset

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
partition string false Partition used to predict
projectId string true The ID of the project
type string true Type name for this intake type

Enumerated Values

Property Value
anonymous [redacted]
partition [holdout, validation, allBacktests, None]
type dss

DataQualityWarningsRecord

{
  "hasKiaMissingValuesInForecastWindow": true,
  "insufficientRowsForEvaluatingModels": true,
  "singleClassActualValueColumn": true
}

A Json object of available warnings about potential problems in this prediction dataset. Empty if no warnings.

Properties

Name Type Required Restrictions Description
hasKiaMissingValuesInForecastWindow boolean false If true, known-in-advance features in this dataset have missing values in the forecast window. Absence of the known-in-advance values can negatively impact prediction quality. Only applies for time series projects.
insufficientRowsForEvaluatingModels boolean false If true, the dataset has a target column present indicating it can be used to evaluate model performance but too few rows to be trustworthy in so doing. If false, either it has no target column at all or it has sufficient rows for model evaluation. Only applies for regression, binary classification, multiclass classification projects and time series unsupervised projects.
singleClassActualValueColumn boolean false If true, actual value column has only one class and such insights as ROC curve can not be calculated. Only applies for binary classification projects or unsupervised projects.

DataStageDataStreamer

{
  "dataStageId": "string",
  "type": "dataStage"
}

Stream CSV data chunks from data stage storage

Properties

Name Type Required Restrictions Description
dataStageId string true The ID of the data stage
type string true Type name for this intake type

Enumerated Values

Property Value
type dataStage

DataStageIntake

{
  "dataStageId": "string",
  "type": "dataStage"
}

Stream CSV data chunks from data stage storage

Properties

Name Type Required Restrictions Description
dataStageId string true The ID of the data stage
type string true Type name for this intake type

Enumerated Values

Property Value
type dataStage

DatabricksAccessTokenCredentials

{
  "credentialType": "databricks_access_token_account",
  "databricksAccessToken": "string"
}

Properties

Name Type Required Restrictions Description
credentialType string true The type of these credentials, 'databricks_access_token_account' here.
databricksAccessToken string true minLength: 1
minLength: 1
Databricks personal access token.

Enumerated Values

Property Value
credentialType databricks_access_token_account

FileSystemDataStreamer

{
  "path": "string",
  "type": "filesystem"
}

Properties

Name Type Required Restrictions Description
path string true Path to data on host filesystem
type string true Type name for this intake type

Enumerated Values

Property Value
type filesystem

FileSystemIntake

{
  "path": "string",
  "type": "filesystem"
}

Properties

Name Type Required Restrictions Description
path string true Path to data on host filesystem
type string true Type name for this intake type

Enumerated Values

Property Value
type filesystem

FileSystemOutput

{
  "path": "string",
  "type": "filesystem"
}

Properties

Name Type Required Restrictions Description
path string true Path to results on host filesystem
type string true Type name for this output type

Enumerated Values

Property Value
type filesystem

FileSystemOutputAdaptor

{
  "path": "string",
  "type": "filesystem"
}

Properties

Name Type Required Restrictions Description
path string true Path to results on host filesystem
type string true Type name for this output type

Enumerated Values

Property Value
type filesystem

GCPDataStreamer

{
  "credentialId": "string",
  "format": "csv",
  "type": "gcp",
  "url": "string"
}

Stream CSV data chunks from Google Storage

Properties

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false Use the specified credential to access the url

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
format string false Type of input file format
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
anonymous [redacted]
format [csv, parquet]
type gcp

GCPIntake