Batch Predictions¶
This page outlines the operations, endpoints, parameters, and example requests and responses for the Batch Predictions.
GET /api/v2/batchJobs/¶
Get a collection of batch jobs by statuses
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchJobs/?offset=0&limit=100 \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
offset | query | integer | true | This many results will be skipped |
limit | query | integer | true | At most this many results are returned |
status | query | any | false | Includes only jobs that have the status value that matches this flag. Repeat the parameter for filtering on multiple statuses. |
source | query | any | false | Includes only jobs that have the source value that matches this flag. Repeat the parameter for filtering on multiple statuses.Prefix values with a dash (- ) to exclude those sources. |
deploymentId | query | string | false | Includes only jobs for this particular deployment |
modelId | query | string | false | ID of leaderboard model which is used in job for processing predictions dataset |
jobId | query | string | false | Includes only job by specific id |
orderBy | query | string | false | Sort order which will be applied to batch prediction list. Prefix the attribute name with a dash to sort in descending order, e.g. "-created". |
allJobs | query | boolean | false | [DEPRECATED - replaced with RBAC permission model] - No effect |
cutoffHours | query | integer | false | Only list jobs created at most this amount of hours ago. |
startDateTime | query | string(date-time) | false | ISO-formatted datetime of the earliest time the job was added (inclusive). For example "2008-08-24T12:00:00Z". Will ignore cutoffHours if set. |
endDateTime | query | string(date-time) | false | ISO-formatted datetime of the latest time the job was added (inclusive). For example "2008-08-24T12:00:00Z". |
batchPredictionJobDefinitionId | query | string | false | Includes only jobs for this particular definition |
hostname | query | any | false | Includes only jobs for this particular prediction instance hostname |
batchJobType | query | any | false | Includes only jobs that have the batch job type that matches this flag. Repeat the parameter for filtering on multiple types. |
intakeType | query | any | false | Includes only jobs for these particular intakes type |
outputType | query | any | false | Includes only jobs for these particular outputs type |
Enumerated Values¶
Parameter | Value |
---|---|
orderBy | [created , -created , status , -status ] |
Example responses¶
200 Response
{
"count": 0,
"data": [
{
"batchMonitoringJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"batchPredictionJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"elapsedTimeSec": 0,
"failedRows": 0,
"hidden": "2019-08-24T14:15:22Z",
"id": "string",
"intakeDatasetDisplayName": "string",
"jobIntakeSize": 0,
"jobOutputSize": 0,
"jobSpec": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"maxNgramExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringAggregation": {
"retentionPolicy": "samples",
"retentionValue": 0
},
"monitoringBatchPrefix": "string",
"monitoringColumns": {
"actedUponColumn": "string",
"actualsTimestampColumn": "string",
"actualsValueColumn": "string",
"associationIdColumn": "string",
"customMetricId": "string",
"customMetricTimestampColumn": "string",
"customMetricTimestampFormat": "string",
"customMetricValueColumn": "string",
"monitoredStatusColumn": "string",
"predictionsColumns": [
{
"className": "string",
"columnName": "string"
}
],
"reportDrift": true,
"reportPredictions": true,
"uniqueRowIdentifierColumns": [
"string"
]
},
"monitoringOutputSettings": {
"monitoredStatusColumn": "string",
"uniqueRowIdentifierColumns": [
"string"
]
},
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"links": {
"csvUpload": "string",
"download": "string",
"self": "string"
},
"logs": [
"string"
],
"monitoringBatchId": "string",
"percentageCompleted": 100,
"queuePosition": 0,
"queued": true,
"resultsDeleted": true,
"scoredRows": 0,
"skippedRows": 0,
"source": "string",
"status": "INITIALIZING",
"statusDetails": "string"
}
],
"next": "http://example.com",
"previous": "http://example.com",
"totalCount": 0
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | A list of Batch job objects | BatchJobListResponse |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/batchJobs/fromJobDefinition/¶
Launches a one-time batch job based off of the previously supplied definition referring to the job definition ID and puts it on the queue.
Code samples¶
# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/batchJobs/fromJobDefinition/ \
-H "Content-Type: application/json" \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{BatchPredictionJobDefinitionId}'
Body parameter¶
{
"jobDefinitionId": "string"
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
body | body | BatchPredictionJobDefinitionId | false | none |
Example responses¶
202 Response
{
"batchMonitoringJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"batchPredictionJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"elapsedTimeSec": 0,
"failedRows": 0,
"hidden": "2019-08-24T14:15:22Z",
"id": "string",
"intakeDatasetDisplayName": "string",
"jobIntakeSize": 0,
"jobOutputSize": 0,
"jobSpec": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"maxNgramExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringAggregation": {
"retentionPolicy": "samples",
"retentionValue": 0
},
"monitoringBatchPrefix": "string",
"monitoringColumns": {
"actedUponColumn": "string",
"actualsTimestampColumn": "string",
"actualsValueColumn": "string",
"associationIdColumn": "string",
"customMetricId": "string",
"customMetricTimestampColumn": "string",
"customMetricTimestampFormat": "string",
"customMetricValueColumn": "string",
"monitoredStatusColumn": "string",
"predictionsColumns": [
{
"className": "string",
"columnName": "string"
}
],
"reportDrift": true,
"reportPredictions": true,
"uniqueRowIdentifierColumns": [
"string"
]
},
"monitoringOutputSettings": {
"monitoredStatusColumn": "string",
"uniqueRowIdentifierColumns": [
"string"
]
},
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"links": {
"csvUpload": "string",
"download": "string",
"self": "string"
},
"logs": [
"string"
],
"monitoringBatchId": "string",
"percentageCompleted": 100,
"queuePosition": 0,
"queued": true,
"resultsDeleted": true,
"scoredRows": 0,
"skippedRows": 0,
"source": "string",
"status": "INITIALIZING",
"statusDetails": "string"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Job details for the created Batch Prediction job | BatchJobResponse |
404 | Not Found | Job was deleted, never existed or you do not have access to it | None |
422 | Unprocessable Entity | Could not create a Batch job. Possible reasons: {} | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
DELETE /api/v2/batchJobs/{batchJobId}/¶
If the job is running, it will be aborted. Then it will be removed, meaning all underlying data will be deleted and the job is removed from the list of jobs.
Code samples¶
# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/batchJobs/{batchJobId}/ \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
batchJobId | path | string | true | ID of the Batch job |
partNumber | path | integer | true | The number of which csv part is being uploaded when using multipart upload |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Job cancelled | None |
404 | Not Found | Job does not exist or was not submitted to the queue. | None |
409 | Conflict | Job cannot be aborted for some reason. Possible reasons: job is already aborted or completed. | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/batchJobs/{batchJobId}/¶
Retrieve a Batch job.
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchJobs/{batchJobId}/ \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
batchJobId | path | string | true | ID of the Batch job |
partNumber | path | integer | true | The number of which csv part is being uploaded when using multipart upload |
Example responses¶
200 Response
{
"batchMonitoringJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"batchPredictionJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"elapsedTimeSec": 0,
"failedRows": 0,
"hidden": "2019-08-24T14:15:22Z",
"id": "string",
"intakeDatasetDisplayName": "string",
"jobIntakeSize": 0,
"jobOutputSize": 0,
"jobSpec": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"maxNgramExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringAggregation": {
"retentionPolicy": "samples",
"retentionValue": 0
},
"monitoringBatchPrefix": "string",
"monitoringColumns": {
"actedUponColumn": "string",
"actualsTimestampColumn": "string",
"actualsValueColumn": "string",
"associationIdColumn": "string",
"customMetricId": "string",
"customMetricTimestampColumn": "string",
"customMetricTimestampFormat": "string",
"customMetricValueColumn": "string",
"monitoredStatusColumn": "string",
"predictionsColumns": [
{
"className": "string",
"columnName": "string"
}
],
"reportDrift": true,
"reportPredictions": true,
"uniqueRowIdentifierColumns": [
"string"
]
},
"monitoringOutputSettings": {
"monitoredStatusColumn": "string",
"uniqueRowIdentifierColumns": [
"string"
]
},
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"links": {
"csvUpload": "string",
"download": "string",
"self": "string"
},
"logs": [
"string"
],
"monitoringBatchId": "string",
"percentageCompleted": 100,
"queuePosition": 0,
"queued": true,
"resultsDeleted": true,
"scoredRows": 0,
"skippedRows": 0,
"source": "string",
"status": "INITIALIZING",
"statusDetails": "string"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | Job details for the requested Batch job | BatchJobResponse |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
PUT /api/v2/batchJobs/{batchJobId}/csvUpload/¶
Stream CSV data to the job. Only available for jobs thatuses the localFile intake option.
Code samples¶
# You can also use wget
curl -X PUT https://app.datarobot.com/api/v2/batchJobs/{batchJobId}/csvUpload/ \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
batchJobId | path | string | true | ID of the Batch job |
partNumber | path | integer | true | The number of which csv part is being uploaded when using multipart upload |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Job data was successfully submitted | None |
404 | Not Found | Job does not exist or does not require data | None |
409 | Conflict | Dataset upload has already begun | None |
415 | Unsupported Media Type | Not acceptable MIME type | None |
422 | Unprocessable Entity | Job was "ABORTED" due to too many errors in the data | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/batchJobs/{batchJobId}/download/¶
This is only valid for jobs scored using the "localFile" output option
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchJobs/{batchJobId}/download/ \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
batchJobId | path | string | true | ID of the Batch job |
partNumber | path | integer | true | The number of which csv part is being uploaded when using multipart upload |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | Job was downloaded correctly | None |
404 | Not Found | Job does not exist or is not completed | None |
406 | Not Acceptable | Not acceptable MIME type | None |
422 | Unprocessable Entity | Job was "ABORTED" due to too many errors in the data | None |
Response Headers¶
Status | Header | Type | Format | Description |
---|---|---|---|---|
200 | Content-Disposition | string | Contains an auto generated filename for this download ("attachment;filename=result- |
|
200 | Content-Type | string | MIME type of the returned data |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/batchPredictionJobDefinitions/¶
List all Batch Prediction jobs definitions available
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchPredictionJobDefinitions/?offset=0&limit=100 \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
offset | query | integer | true | This many results will be skipped |
limit | query | integer | true | At most this many results are returned |
searchName | query | string | false | A human-readable name for the definition, must be unique across organisations. |
deploymentId | query | string | false | Includes only definitions for this particular deployment |
Example responses¶
200 Response
{
"count": 0,
"data": [
{
"batchPredictionJob": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"maxNgramExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringAggregation": {
"retentionPolicy": "samples",
"retentionValue": 0
},
"monitoringBatchPrefix": "string",
"monitoringColumns": {
"actedUponColumn": "string",
"actualsTimestampColumn": "string",
"actualsValueColumn": "string",
"associationIdColumn": "string",
"customMetricId": "string",
"customMetricTimestampColumn": "string",
"customMetricTimestampFormat": "string",
"customMetricValueColumn": "string",
"monitoredStatusColumn": "string",
"predictionsColumns": [
{
"className": "string",
"columnName": "string"
}
],
"reportDrift": true,
"reportPredictions": true,
"uniqueRowIdentifierColumns": [
"string"
]
},
"monitoringOutputSettings": {
"monitoredStatusColumn": "string",
"uniqueRowIdentifierColumns": [
"string"
]
},
"numConcurrent": 0,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"enabled": false,
"id": "string",
"lastFailedRunTime": "2019-08-24T14:15:22Z",
"lastScheduledRunTime": "2019-08-24T14:15:22Z",
"lastStartedJobStatus": "INITIALIZING",
"lastStartedJobTime": "2019-08-24T14:15:22Z",
"lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
"name": "string",
"nextScheduledRunTime": "2019-08-24T14:15:22Z",
"schedule": {
"dayOfMonth": [
"*"
],
"dayOfWeek": [
"*"
],
"hour": [
"*"
],
"minute": [
"*"
],
"month": [
"*"
]
},
"updated": "2019-08-24T14:15:22Z",
"updatedBy": {
"fullName": "string",
"userId": "string",
"username": "string"
}
}
],
"next": "http://example.com",
"previous": "http://example.com",
"totalCount": 0
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | List of all available jobs | BatchPredictionJobDefinitionsListResponse |
422 | Unprocessable Entity | Your input data or query arguments did not work together | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/batchPredictionJobDefinitions/¶
Create a Batch Prediction Job definition. A configuration for a Batch Prediction job which can either be executed manually upon request or on scheduled intervals, if enabled. The API payload is the same as for /batchPredictions
along with optional enabled
and schedule
items.
Code samples¶
# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/batchPredictionJobDefinitions/ \
-H "Content-Type: application/json" \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{BatchPredictionJobDefinitionsCreate}'
Body parameter¶
{
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"enabled": true,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringBatchPrefix": "string",
"name": "string",
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionThreshold": 1,
"predictionWarningEnabled": true,
"schedule": {
"dayOfMonth": [
"*"
],
"dayOfWeek": [
"*"
],
"hour": [
"*"
],
"minute": [
"*"
],
"month": [
"*"
]
},
"secondaryDatasetsConfigId": "string",
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
body | body | BatchPredictionJobDefinitionsCreate | false | none |
Example responses¶
202 Response
{
"batchPredictionJob": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"maxNgramExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringAggregation": {
"retentionPolicy": "samples",
"retentionValue": 0
},
"monitoringBatchPrefix": "string",
"monitoringColumns": {
"actedUponColumn": "string",
"actualsTimestampColumn": "string",
"actualsValueColumn": "string",
"associationIdColumn": "string",
"customMetricId": "string",
"customMetricTimestampColumn": "string",
"customMetricTimestampFormat": "string",
"customMetricValueColumn": "string",
"monitoredStatusColumn": "string",
"predictionsColumns": [
{
"className": "string",
"columnName": "string"
}
],
"reportDrift": true,
"reportPredictions": true,
"uniqueRowIdentifierColumns": [
"string"
]
},
"monitoringOutputSettings": {
"monitoredStatusColumn": "string",
"uniqueRowIdentifierColumns": [
"string"
]
},
"numConcurrent": 0,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"enabled": false,
"id": "string",
"lastFailedRunTime": "2019-08-24T14:15:22Z",
"lastScheduledRunTime": "2019-08-24T14:15:22Z",
"lastStartedJobStatus": "INITIALIZING",
"lastStartedJobTime": "2019-08-24T14:15:22Z",
"lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
"name": "string",
"nextScheduledRunTime": "2019-08-24T14:15:22Z",
"schedule": {
"dayOfMonth": [
"*"
],
"dayOfWeek": [
"*"
],
"hour": [
"*"
],
"minute": [
"*"
],
"month": [
"*"
]
},
"updated": "2019-08-24T14:15:22Z",
"updatedBy": {
"fullName": "string",
"userId": "string",
"username": "string"
}
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Job details for the created Batch Prediction job definition | BatchPredictionJobDefinitionsResponse |
403 | Forbidden | You are not authorized to create a job definition on this deployment due to your permissions role | None |
422 | Unprocessable Entity | You tried to create a job definition with uncompatible or missing parameters to create a fully functioning job definition | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
DELETE /api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/¶
Delete a Batch Prediction job definition
Code samples¶
# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/ \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
jobDefinitionId | path | string | true | ID of the Batch Prediction job definition |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
204 | No Content | none | None |
403 | Forbidden | You are not authorized to delete this job definition due to your permissions role | None |
404 | Not Found | Job was deleted, never existed or you do not have access to it | None |
409 | Conflict | Job could not be deleted, as there are currently running jobs in the queue. | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/¶
Retrieve a Batch Prediction job definition
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/ \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
jobDefinitionId | path | string | true | ID of the Batch Prediction job definition |
Example responses¶
200 Response
{
"batchPredictionJob": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"maxNgramExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringAggregation": {
"retentionPolicy": "samples",
"retentionValue": 0
},
"monitoringBatchPrefix": "string",
"monitoringColumns": {
"actedUponColumn": "string",
"actualsTimestampColumn": "string",
"actualsValueColumn": "string",
"associationIdColumn": "string",
"customMetricId": "string",
"customMetricTimestampColumn": "string",
"customMetricTimestampFormat": "string",
"customMetricValueColumn": "string",
"monitoredStatusColumn": "string",
"predictionsColumns": [
{
"className": "string",
"columnName": "string"
}
],
"reportDrift": true,
"reportPredictions": true,
"uniqueRowIdentifierColumns": [
"string"
]
},
"monitoringOutputSettings": {
"monitoredStatusColumn": "string",
"uniqueRowIdentifierColumns": [
"string"
]
},
"numConcurrent": 0,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"enabled": false,
"id": "string",
"lastFailedRunTime": "2019-08-24T14:15:22Z",
"lastScheduledRunTime": "2019-08-24T14:15:22Z",
"lastStartedJobStatus": "INITIALIZING",
"lastStartedJobTime": "2019-08-24T14:15:22Z",
"lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
"name": "string",
"nextScheduledRunTime": "2019-08-24T14:15:22Z",
"schedule": {
"dayOfMonth": [
"*"
],
"dayOfWeek": [
"*"
],
"hour": [
"*"
],
"minute": [
"*"
],
"month": [
"*"
]
},
"updated": "2019-08-24T14:15:22Z",
"updatedBy": {
"fullName": "string",
"userId": "string",
"username": "string"
}
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | Job details for the requested Batch Prediction job definition | BatchPredictionJobDefinitionsResponse |
404 | Not Found | Job was deleted, never existed or you do not have access to it | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
PATCH /api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/¶
Update a Batch Prediction job definition
Code samples¶
# You can also use wget
curl -X PATCH https://app.datarobot.com/api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/ \
-H "Content-Type: application/json" \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{undefined}'
Body parameter¶
{
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"enabled": true,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringBatchPrefix": "string",
"name": "string",
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionThreshold": 1,
"predictionWarningEnabled": true,
"schedule": {
"dayOfMonth": [
"*"
],
"dayOfWeek": [
"*"
],
"hour": [
"*"
],
"minute": [
"*"
],
"month": [
"*"
]
},
"secondaryDatasetsConfigId": "string",
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
jobDefinitionId | path | string | true | ID of the Batch Prediction job definition |
body | body | BatchPredictionJobDefinitionsUpdate | false | none |
Example responses¶
200 Response
{
"batchPredictionJob": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"maxNgramExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringAggregation": {
"retentionPolicy": "samples",
"retentionValue": 0
},
"monitoringBatchPrefix": "string",
"monitoringColumns": {
"actedUponColumn": "string",
"actualsTimestampColumn": "string",
"actualsValueColumn": "string",
"associationIdColumn": "string",
"customMetricId": "string",
"customMetricTimestampColumn": "string",
"customMetricTimestampFormat": "string",
"customMetricValueColumn": "string",
"monitoredStatusColumn": "string",
"predictionsColumns": [
{
"className": "string",
"columnName": "string"
}
],
"reportDrift": true,
"reportPredictions": true,
"uniqueRowIdentifierColumns": [
"string"
]
},
"monitoringOutputSettings": {
"monitoredStatusColumn": "string",
"uniqueRowIdentifierColumns": [
"string"
]
},
"numConcurrent": 0,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"enabled": false,
"id": "string",
"lastFailedRunTime": "2019-08-24T14:15:22Z",
"lastScheduledRunTime": "2019-08-24T14:15:22Z",
"lastStartedJobStatus": "INITIALIZING",
"lastStartedJobTime": "2019-08-24T14:15:22Z",
"lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
"name": "string",
"nextScheduledRunTime": "2019-08-24T14:15:22Z",
"schedule": {
"dayOfMonth": [
"*"
],
"dayOfWeek": [
"*"
],
"hour": [
"*"
],
"minute": [
"*"
],
"month": [
"*"
]
},
"updated": "2019-08-24T14:15:22Z",
"updatedBy": {
"fullName": "string",
"userId": "string",
"username": "string"
}
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | Job details for the updated Batch Prediction job definition | BatchPredictionJobDefinitionsResponse |
403 | Forbidden | You are not authorized to alter the contents of this job definition due to your permissions role | None |
404 | Not Found | Job was deleted, never existed or you do not have access to it | None |
409 | Conflict | You chose a name of your job definition that was already existing within your organization | None |
422 | Unprocessable Entity | Could not update the job definition. Possible reasons: {} | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/portable/¶
Retrieve a Batch Prediction job definition for Portable Batch Predictions
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchPredictionJobDefinitions/{jobDefinitionId}/portable/ \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
jobDefinitionId | path | string | true | ID of the Batch Prediction job definition |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | Snippet for Portable Batch Predictions | None |
404 | Not Found | Job was deleted, never existed or you do not have access to it | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/batchPredictions/¶
Get a collection of batch prediction jobs by statuses
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchPredictions/?offset=0&limit=100 \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
offset | query | integer | true | This many results will be skipped |
limit | query | integer | true | At most this many results are returned |
status | query | any | false | Includes only jobs that have the status value that matches this flag. Repeat the parameter for filtering on multiple statuses. |
source | query | any | false | Includes only jobs that have the source value that matches this flag. Repeat the parameter for filtering on multiple statuses.Prefix values with a dash (- ) to exclude those sources. |
deploymentId | query | string | false | Includes only jobs for this particular deployment |
modelId | query | string | false | ID of leaderboard model which is used in job for processing predictions dataset |
jobId | query | string | false | Includes only job by specific id |
orderBy | query | string | false | Sort order which will be applied to batch prediction list. Prefix the attribute name with a dash to sort in descending order, e.g. "-created". |
allJobs | query | boolean | false | [DEPRECATED - replaced with RBAC permission model] - No effect |
cutoffHours | query | integer | false | Only list jobs created at most this amount of hours ago. |
startDateTime | query | string(date-time) | false | ISO-formatted datetime of the earliest time the job was added (inclusive). For example "2008-08-24T12:00:00Z". Will ignore cutoffHours if set. |
endDateTime | query | string(date-time) | false | ISO-formatted datetime of the latest time the job was added (inclusive). For example "2008-08-24T12:00:00Z". |
batchPredictionJobDefinitionId | query | string | false | Includes only jobs for this particular definition |
hostname | query | any | false | Includes only jobs for this particular prediction instance hostname |
intakeType | query | any | false | Includes only jobs for these particular intakes type |
outputType | query | any | false | Includes only jobs for these particular outputs type |
Enumerated Values¶
Parameter | Value |
---|---|
orderBy | [created , -created , status , -status ] |
Example responses¶
200 Response
{
"count": 0,
"data": [
{
"batchPredictionJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"elapsedTimeSec": 0,
"failedRows": 0,
"hidden": "2019-08-24T14:15:22Z",
"id": "string",
"intakeDatasetDisplayName": "string",
"jobIntakeSize": 0,
"jobOutputSize": 0,
"jobSpec": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringBatchPrefix": "string",
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionThreshold": 1,
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"secondaryDatasetsConfigId": "string",
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"links": {
"csvUpload": "string",
"download": "string",
"self": "string"
},
"logs": [
"string"
],
"monitoringBatchId": "string",
"percentageCompleted": 100,
"queuePosition": 0,
"queued": true,
"resultsDeleted": true,
"scoredRows": 0,
"skippedRows": 0,
"source": "string",
"status": "INITIALIZING",
"statusDetails": "string"
}
],
"next": "http://example.com",
"previous": "http://example.com",
"totalCount": 0
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | A list of Batch Prediction job objects | BatchPredictionJobListResponse |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/batchPredictions/¶
Submit the configuration for the job and it will be submitted to the queue
Code samples¶
# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/batchPredictions/ \
-H "Content-Type: application/json" \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{BatchPredictionJobCreate}'
Body parameter¶
{
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringBatchPrefix": "string",
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionThreshold": 1,
"predictionWarningEnabled": true,
"secondaryDatasetsConfigId": "string",
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
body | body | BatchPredictionJobCreate | false | none |
Example responses¶
202 Response
{
"batchPredictionJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"elapsedTimeSec": 0,
"failedRows": 0,
"hidden": "2019-08-24T14:15:22Z",
"id": "string",
"intakeDatasetDisplayName": "string",
"jobIntakeSize": 0,
"jobOutputSize": 0,
"jobSpec": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringBatchPrefix": "string",
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionThreshold": 1,
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"secondaryDatasetsConfigId": "string",
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"links": {
"csvUpload": "string",
"download": "string",
"self": "string"
},
"logs": [
"string"
],
"monitoringBatchId": "string",
"percentageCompleted": 100,
"queuePosition": 0,
"queued": true,
"resultsDeleted": true,
"scoredRows": 0,
"skippedRows": 0,
"source": "string",
"status": "INITIALIZING",
"statusDetails": "string"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Job details for the created Batch Prediction job | BatchPredictionJobResponse |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/batchPredictions/fromExisting/¶
Copies an existing job and submits it to the queue.
Code samples¶
# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/batchPredictions/fromExisting/ \
-H "Content-Type: application/json" \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{BatchPredictionJobId}'
Body parameter¶
{
"partNumber": 0,
"predictionJobId": "string"
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
body | body | BatchPredictionJobId | false | none |
Example responses¶
202 Response
{
"batchPredictionJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"elapsedTimeSec": 0,
"failedRows": 0,
"hidden": "2019-08-24T14:15:22Z",
"id": "string",
"intakeDatasetDisplayName": "string",
"jobIntakeSize": 0,
"jobOutputSize": 0,
"jobSpec": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringBatchPrefix": "string",
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionThreshold": 1,
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"secondaryDatasetsConfigId": "string",
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"links": {
"csvUpload": "string",
"download": "string",
"self": "string"
},
"logs": [
"string"
],
"monitoringBatchId": "string",
"percentageCompleted": 100,
"queuePosition": 0,
"queued": true,
"resultsDeleted": true,
"scoredRows": 0,
"skippedRows": 0,
"source": "string",
"status": "INITIALIZING",
"statusDetails": "string"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Job details for the created Batch Prediction job | BatchPredictionJobResponse |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/batchPredictions/fromJobDefinition/¶
Launches a one-time batch prediction job based off of the previously supplied definition referring to the job definition ID and puts it on the queue.
Code samples¶
# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/batchPredictions/fromJobDefinition/ \
-H "Content-Type: application/json" \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{BatchPredictionJobDefinitionId}'
Body parameter¶
{
"jobDefinitionId": "string"
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
body | body | BatchPredictionJobDefinitionId | false | none |
Example responses¶
202 Response
{
"batchPredictionJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"elapsedTimeSec": 0,
"failedRows": 0,
"hidden": "2019-08-24T14:15:22Z",
"id": "string",
"intakeDatasetDisplayName": "string",
"jobIntakeSize": 0,
"jobOutputSize": 0,
"jobSpec": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringBatchPrefix": "string",
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionThreshold": 1,
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"secondaryDatasetsConfigId": "string",
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"links": {
"csvUpload": "string",
"download": "string",
"self": "string"
},
"logs": [
"string"
],
"monitoringBatchId": "string",
"percentageCompleted": 100,
"queuePosition": 0,
"queued": true,
"resultsDeleted": true,
"scoredRows": 0,
"skippedRows": 0,
"source": "string",
"status": "INITIALIZING",
"statusDetails": "string"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Job details for the created Batch Prediction job | BatchPredictionJobResponse |
404 | Not Found | Job was deleted, never existed or you do not have access to it | None |
422 | Unprocessable Entity | Could not create a Batch job. Possible reasons: {} | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
DELETE /api/v2/batchPredictions/{predictionJobId}/¶
If the job is running, it will be aborted. Then it will be removed, meaning all underlying data will be deleted and the job is removed from the list of jobs.
Code samples¶
# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/batchPredictions/{predictionJobId}/ \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
predictionJobId | path | string | true | ID of the Batch Prediction job |
partNumber | path | integer | true | The number of which csv part is being uploaded when using multipart upload |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Job cancelled | None |
404 | Not Found | Job does not exist or was not submitted to the queue. | None |
409 | Conflict | Job cannot be aborted for some reason. Possible reasons: job is already aborted or completed. | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/batchPredictions/{predictionJobId}/¶
Retrieve a Batch Prediction job.
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchPredictions/{predictionJobId}/ \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
predictionJobId | path | string | true | ID of the Batch Prediction job |
partNumber | path | integer | true | The number of which csv part is being uploaded when using multipart upload |
Example responses¶
200 Response
{
"batchPredictionJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"elapsedTimeSec": 0,
"failedRows": 0,
"hidden": "2019-08-24T14:15:22Z",
"id": "string",
"intakeDatasetDisplayName": "string",
"jobIntakeSize": 0,
"jobOutputSize": 0,
"jobSpec": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringBatchPrefix": "string",
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionThreshold": 1,
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"secondaryDatasetsConfigId": "string",
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"links": {
"csvUpload": "string",
"download": "string",
"self": "string"
},
"logs": [
"string"
],
"monitoringBatchId": "string",
"percentageCompleted": 100,
"queuePosition": 0,
"queued": true,
"resultsDeleted": true,
"scoredRows": 0,
"skippedRows": 0,
"source": "string",
"status": "INITIALIZING",
"statusDetails": "string"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | Job details for the requested Batch Prediction job | BatchPredictionJobResponse |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
PATCH /api/v2/batchPredictions/{predictionJobId}/¶
If a job has finished execution regardless of the result, it can have parameters changed to ensure better filtering in the job list upon retrieval. Another case: updating job scoring status externally.
Code samples¶
# You can also use wget
curl -X PATCH https://app.datarobot.com/api/v2/batchPredictions/{predictionJobId}/ \
-H "Content-Type: application/json" \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{undefined}'
Body parameter¶
{
"aborted": "2019-08-24T14:15:22Z",
"completed": "2019-08-24T14:15:22Z",
"failedRows": 0,
"hidden": true,
"jobIntakeSize": 0,
"jobOutputSize": 0,
"logs": [
"string"
],
"scoredRows": 0,
"skippedRows": 0,
"started": "2019-08-24T14:15:22Z",
"status": "INITIALIZING"
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
predictionJobId | path | string | true | ID of the Batch Prediction job |
partNumber | path | integer | true | The number of which csv part is being uploaded when using multipart upload |
body | body | BatchPredictionJobUpdate | false | none |
Example responses¶
200 Response
{
"batchPredictionJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"elapsedTimeSec": 0,
"failedRows": 0,
"hidden": "2019-08-24T14:15:22Z",
"id": "string",
"intakeDatasetDisplayName": "string",
"jobIntakeSize": 0,
"jobOutputSize": 0,
"jobSpec": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringBatchPrefix": "string",
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionThreshold": 1,
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"secondaryDatasetsConfigId": "string",
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"links": {
"csvUpload": "string",
"download": "string",
"self": "string"
},
"logs": [
"string"
],
"monitoringBatchId": "string",
"percentageCompleted": 100,
"queuePosition": 0,
"queued": true,
"resultsDeleted": true,
"scoredRows": 0,
"skippedRows": 0,
"source": "string",
"status": "INITIALIZING",
"statusDetails": "string"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | Job updated | BatchPredictionJobResponse |
404 | Not Found | Job does not exist or was not submitted to the queue. | None |
409 | Conflict | Job cannot be hidden for some reason. Possible reasons: job is not in a deletable state. | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
PUT /api/v2/batchPredictions/{predictionJobId}/csvUpload/¶
Stream CSV data to the prediction job. Only available for jobs thatuses the localFile intake option.
Code samples¶
# You can also use wget
curl -X PUT https://app.datarobot.com/api/v2/batchPredictions/{predictionJobId}/csvUpload/ \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
predictionJobId | path | string | true | ID of the Batch Prediction job |
partNumber | path | integer | true | The number of which csv part is being uploaded when using multipart upload |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Job data was successfully submitted | None |
404 | Not Found | Job does not exist or does not require data | None |
406 | Not Acceptable | Not acceptable MIME type | None |
409 | Conflict | Dataset upload has already begun | None |
422 | Unprocessable Entity | Job was "ABORTED" due to too many errors in the data | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/batchPredictions/{predictionJobId}/csvUpload/finalizeMultipart/¶
Finalize a multipart upload, indicating that no further chunks will be sent
Code samples¶
# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/batchPredictions/{predictionJobId}/csvUpload/finalizeMultipart/ \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
predictionJobId | path | string | true | ID of the Batch Prediction job |
partNumber | path | integer | true | The number of which csv part is being uploaded when using multipart upload |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Acknowledgement that the request was accepted or an error message | None |
404 | Not Found | Job was deleted, never existed or you do not have access to it | None |
409 | Conflict | Only multipart jobs can be finalized. | None |
422 | Unprocessable Entity | No data was uploaded | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
PUT /api/v2/batchPredictions/{predictionJobId}/csvUpload/part/{partNumber}/¶
Stream CSV data to the prediction job in many parts.Only available for jobs that uses the localFile intake option.
Code samples¶
# You can also use wget
curl -X PUT https://app.datarobot.com/api/v2/batchPredictions/{predictionJobId}/csvUpload/part/{partNumber}/ \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
predictionJobId | path | string | true | ID of the Batch Prediction job |
partNumber | path | integer | true | The number of which csv part is being uploaded when using multipart upload |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Job data was successfully submitted | None |
404 | Not Found | Job does not exist or does not require data | None |
406 | Not Acceptable | Not acceptable MIME type | None |
409 | Conflict | Dataset upload has already begun | None |
422 | Unprocessable Entity | Job was "ABORTED" due to too many errors in the data | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/batchPredictions/{predictionJobId}/download/¶
This is only valid for jobs scored using the "localFile" output option
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchPredictions/{predictionJobId}/download/ \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
predictionJobId | path | string | true | ID of the Batch Prediction job |
partNumber | path | integer | true | The number of which csv part is being uploaded when using multipart upload |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | Job was downloaded correctly | None |
404 | Not Found | Job does not exist or is not completed | None |
406 | Not Acceptable | Not acceptable MIME type | None |
422 | Unprocessable Entity | Job was "ABORTED" due to too many errors in the data | None |
Response Headers¶
Status | Header | Type | Format | Description |
---|---|---|---|---|
200 | Content-Disposition | string | Contains an auto generated filename for this download ("attachment;filename=result- |
|
200 | Content-Type | string | MIME type of the returned data |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
DELETE /api/v2/projects/{projectId}/models/{modelId}/predictionExplanationsInitialization/¶
Delete an existing PredictionExplanationsInitialization.
Code samples¶
# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/projects/{projectId}/models/{modelId}/predictionExplanationsInitialization/ \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
projectId | path | string | true | The project ID |
modelId | path | string | true | The model ID |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
204 | No Content | The deletion was successful. | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/projects/{projectId}/models/{modelId}/predictionExplanationsInitialization/¶
Retrieve the current PredictionExplanationsInitialization. A PredictionExplanationsInitialization is a pre-requisite for successfully computing prediction explanations using a particular model, and can be used to preview the prediction explanations that would be generated for a complete dataset.
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/models/{modelId}/predictionExplanationsInitialization/ \
-H "Content-Type: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{undefined}'
Body parameter¶
{
"modelId": "string",
"predictionExplanationsSample": [
{
"adjustedPrediction": 0,
"adjustedPredictionValues": [
{
"label": "string",
"value": 0
}
],
"forecastDistance": 0,
"forecastPoint": "string",
"prediction": 0,
"predictionExplanations": [
{
"feature": "string",
"featureValue": "string",
"imageExplanationUrl": "string",
"label": "string",
"perNgramTextExplanations": [
{
"isUnknown": true,
"ngrams": [
{
"label": "string",
"value": 0
}
],
"qualitativateStrength": "string",
"strength": 0
}
],
"qualitativateStrength": "string",
"strength": 0
}
],
"predictionThreshold": 0,
"predictionValues": [
{
"label": "string",
"value": 0
}
],
"rowId": 0,
"seriesId": "string",
"timestamp": "string"
}
],
"projectId": "string"
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
excludeAdjustedPredictions | query | string | false | Whether to include adjusted prediction in the PredictionExplanationsSample response. |
projectId | path | string | true | The project ID |
modelId | path | string | true | The model ID |
body | body | PredictionExplanationsInitializationRetrieve | false | none |
Enumerated Values¶
Parameter | Value |
---|---|
excludeAdjustedPredictions | [false , False , true , True ] |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | none | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/projects/{projectId}/models/{modelId}/predictionExplanationsInitialization/¶
Create a new prediction explanations initialization. This is a necessary prerequisite for generating prediction explanations.
Code samples¶
# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/models/{modelId}/predictionExplanationsInitialization/ \
-H "Content-Type: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{undefined}'
Body parameter¶
{
"maxExplanations": 3,
"thresholdHigh": null,
"thresholdLow": null
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
projectId | path | string | true | The project ID |
modelId | path | string | true | The model ID |
body | body | PredictionExplanationsInitializationCreate | false | none |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | The request was accepted and will be worked on. | None |
Response Headers¶
Status | Header | Type | Format | Description |
---|---|---|---|---|
202 | Location | string | A url that can be polled to check the status. |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/projects/{projectId}/predictJobs/¶
List all prediction jobs for a project
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictJobs/ \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
status | query | string | false | If provided, only jobs with the same status will be included in the results; otherwise, queued and inprogress jobs (but not errored jobs) will be returned. |
projectId | path | string | true | The project ID. |
Enumerated Values¶
Parameter | Value |
---|---|
status | [queue , inprogress , error ] |
Example responses¶
200 Response
[
{
"id": "string",
"isBlocked": true,
"message": "string",
"modelId": "string",
"projectId": "string",
"status": "queue"
}
]
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | A list of prediction jobs for a project | Inline |
404 | Not Found | Job was not found | None |
Response Schema¶
Status Code 200
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
anonymous | [PredictJobDetailsResponse] | false | none | |
» id | string | true | the job ID of the job | |
» isBlocked | boolean | true | True if a job is waiting for its dependencies to be resolved first. | |
» message | string | true | An optional message about the job | |
» modelId | string | true | The ID of the model | |
» projectId | string | true | the project the job belongs to | |
» status | string | true | the status of the job |
Enumerated Values¶
Property | Value |
---|---|
status | [queue , inprogress , error , ABORTED , COMPLETED ] |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
DELETE /api/v2/projects/{projectId}/predictJobs/{jobId}/¶
Cancel a queued prediction job
Code samples¶
# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/projects/{projectId}/predictJobs/{jobId}/ \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
projectId | path | string | true | The project ID. |
jobId | path | string | true | The job ID |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
204 | No Content | The job has been successfully cancelled | None |
404 | Not Found | Job was not found or the job has already completed | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/projects/{projectId}/predictJobs/{jobId}/¶
Look up a particular prediction job
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictJobs/{jobId}/ \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
projectId | path | string | true | The project ID. |
jobId | path | string | true | The job ID |
Example responses¶
200 Response
{
"id": "string",
"isBlocked": true,
"message": "string",
"modelId": "string",
"projectId": "string",
"status": "queue"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | The job has been successfully retrieved and has not yet finished. | PredictJobDetailsResponse |
303 | See Other | The job has been successfully retrieved and has been completed. See Location header. The response json is also included. | None |
Response Headers¶
Status | Header | Type | Format | Description |
---|---|---|---|---|
200 | Location | string | url | present only when the requested job has finished - contains a url from which the completed predictions may be retrieved as with GET /api/v2/projects/{projectId}/predictions/{predictionId}/ |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/projects/{projectId}/predictionDatasets/¶
List predictions datasets uploaded to a project.
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictionDatasets/?offset=0&limit=0 \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
offset | query | integer | true | This many results will be skipped. |
limit | query | integer | true | At most this many results are returned. If 0, all results. |
projectId | path | string | true | The project ID to query. |
Example responses¶
200 Response
{
"count": 0,
"data": [
{
"actualValueColumn": "string",
"catalogId": "string",
"catalogVersionId": "string",
"containsTargetValues": true,
"created": "2019-08-24T14:15:22Z",
"dataEndDate": "2019-08-24T14:15:22Z",
"dataQualityWarnings": {
"hasKiaMissingValuesInForecastWindow": true,
"insufficientRowsForEvaluatingModels": true,
"singleClassActualValueColumn": true
},
"dataStartDate": "2019-08-24T14:15:22Z",
"detectedActualValueColumns": [
{
"missingCount": 0,
"name": "string"
}
],
"forecastPoint": "string",
"forecastPointRange": [
"2019-08-24T14:15:22Z"
],
"id": "string",
"maxForecastDate": "2019-08-24T14:15:22Z",
"name": "string",
"numColumns": 0,
"numRows": 0,
"predictionsEndDate": "2019-08-24T14:15:22Z",
"predictionsStartDate": "2019-08-24T14:15:22Z",
"projectId": "string",
"secondaryDatasetsConfigId": "string"
}
],
"next": "string",
"previous": "string"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | Request to list the uploaded predictions datasets was successful. | PredictionDatasetListControllerResponse |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/projects/{projectId}/predictionDatasets/dataSourceUploads/¶
Upload a dataset for predictions from a DataSource
.
Code samples¶
# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/predictionDatasets/dataSourceUploads/ \
-H "Content-Type: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{undefined}'
Body parameter¶
{
"actualValueColumn": "string",
"credentialData": {
"credentialType": "basic",
"password": "string",
"user": "string"
},
"credentialId": "string",
"credentials": [
{
"catalogVersionId": "string",
"password": "string",
"url": "string",
"user": "string"
}
],
"dataSourceId": "string",
"forecastPoint": "2019-08-24T14:15:22Z",
"password": "string",
"predictionsEndDate": "2019-08-24T14:15:22Z",
"predictionsStartDate": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": true,
"secondaryDatasetsConfigId": "string",
"useKerberos": false,
"user": "string"
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
projectId | path | string | true | The project ID to which the data source will be uploaded to. |
body | body | PredictionDataSource | false | none |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Upload successfully started. See the Location header. | None |
Response Headers¶
Status | Header | Type | Format | Description |
---|---|---|---|---|
202 | Location | string | A url that can be polled to check the status. |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/projects/{projectId}/predictionDatasets/datasetUploads/¶
Create a prediction dataset from a Dataset Asset referenced by AI Catalog item/version ID.
Code samples¶
# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/predictionDatasets/datasetUploads/ \
-H "Content-Type: application/json" \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{undefined}'
Body parameter¶
{
"actualValueColumn": "string",
"credentialData": {
"credentialType": "basic",
"password": "string",
"user": "string"
},
"credentialId": "string",
"credentials": [
{
"catalogVersionId": "string",
"password": "string",
"url": "string",
"user": "string"
}
],
"datasetId": "string",
"datasetVersionId": "string",
"forecastPoint": "2019-08-24T14:15:22Z",
"password": "string",
"predictionsEndDate": "2019-08-24T14:15:22Z",
"predictionsStartDate": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": true,
"secondaryDatasetsConfigId": "string",
"useKerberos": false,
"user": "string"
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
projectId | path | string | true | The project ID. |
body | body | PredictionFromCatalogDataset | false | none |
Example responses¶
202 Response
{
"datasetId": "string"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Creation has successfully started. See the Location header. | CreatePredictionDatasetResponse |
422 | Unprocessable Entity | Target not set yet or cannot specify time series options with a non time series project. | None |
Response Headers¶
Status | Header | Type | Format | Description |
---|---|---|---|---|
202 | Location | string | A url that can be polled to check the status. |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/projects/{projectId}/predictionDatasets/fileUploads/¶
Upload a file for predictions from an attached file.
Code samples¶
# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/predictionDatasets/fileUploads/ \
-H "Content-Type: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{undefined}'
Body parameter¶
{
"actualValueColumn": "string",
"credentials": "string",
"file": "string",
"forecastPoint": "2019-08-24T14:15:22Z",
"predictionsEndDate": "2019-08-24T14:15:22Z",
"predictionsStartDate": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": "false",
"secondaryDatasetsConfigId": "string"
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
projectId | path | string | true | The project ID to which the data will be uploaded for prediction. |
body | body | PredictionFileUpload | false | none |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Upload successfully started. See the Location header. | None |
Response Headers¶
Status | Header | Type | Format | Description |
---|---|---|---|---|
202 | Location | string | A url that can be polled to check the status. |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/projects/{projectId}/predictionDatasets/urlUploads/¶
Upload a file for predictions from a URL.
Code samples¶
# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/predictionDatasets/urlUploads/ \
-H "Content-Type: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{undefined}'
Body parameter¶
{
"actualValueColumn": "string",
"credentials": [
{
"catalogVersionId": "string",
"password": "string",
"url": "string",
"user": "string"
}
],
"forecastPoint": "2019-08-24T14:15:22Z",
"predictionsEndDate": "2019-08-24T14:15:22Z",
"predictionsStartDate": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": true,
"secondaryDatasetsConfigId": "string",
"url": "string"
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
projectId | path | string | true | The project ID to which the data will be uploaded for prediction. |
body | body | PredictionURLUpload | false | none |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Upload successfully started. See the Location header. | None |
Response Headers¶
Status | Header | Type | Format | Description |
---|---|---|---|---|
202 | Location | string | A url that can be polled to check the status. |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
DELETE /api/v2/projects/{projectId}/predictionDatasets/{datasetId}/¶
Delete a dataset that was uploaded for prediction.
Code samples¶
# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/projects/{projectId}/predictionDatasets/{datasetId}/ \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
projectId | path | string | true | The project ID that owns the data. |
datasetId | path | string | true | The dataset ID to delete. |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
204 | No Content | The dataset has been successfully deleted. | None |
404 | Not Found | No dataset with the specified datasetId found. |
None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/projects/{projectId}/predictionDatasets/{datasetId}/¶
Get the metadata of a specific dataset. This only works for datasets uploaded to an existing project for prediction.
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictionDatasets/{datasetId}/ \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
projectId | path | string | true | The project ID that owns the data. |
datasetId | path | string | true | The dataset ID to query for. |
Example responses¶
200 Response
{
"actualValueColumn": "string",
"catalogId": "string",
"catalogVersionId": "string",
"containsTargetValues": true,
"created": "2019-08-24T14:15:22Z",
"dataEndDate": "2019-08-24T14:15:22Z",
"dataQualityWarnings": {
"hasKiaMissingValuesInForecastWindow": true,
"insufficientRowsForEvaluatingModels": true,
"singleClassActualValueColumn": true
},
"dataStartDate": "2019-08-24T14:15:22Z",
"detectedActualValueColumns": [
{
"missingCount": 0,
"name": "string"
}
],
"forecastPoint": "string",
"forecastPointRange": [
"2019-08-24T14:15:22Z"
],
"id": "string",
"maxForecastDate": "2019-08-24T14:15:22Z",
"name": "string",
"numColumns": 0,
"numRows": 0,
"predictionsEndDate": "2019-08-24T14:15:22Z",
"predictionsStartDate": "2019-08-24T14:15:22Z",
"projectId": "string",
"secondaryDatasetsConfigId": "string"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | Request to retrieve the metadata of a specified dataset was successful. | PredictionDatasetRetrieveResponse |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/projects/{projectId}/predictionExplanations/¶
Create a new PredictionExplanations object (and its accompanying PredictionExplanationsRecord).
In order to successfully create PredictionExplanations for a particular model and dataset, you must first
- Compute feature impact for the model via POST /api/v2/projects/{projectId}/models/{modelId}/featureImpact/
- Compute a PredictionExplanationsInitialization for the model via POST /api/v2/projects/{projectId}/models/{modelId}/predictionExplanationsInitialization/
- Compute predictions for the model and dataset via POST /api/v2/projects/{projectId}/predictions/
thresholdHigh
and thresholdLow
are optional filters applied to speed up computation. When at least one is specified, only the selected outlier rows will have prediction explanations computed. Rows are considered to be outliers if their predicted value (in case of regression projects) or probability of being the positive class (in case of classification projects) isless than thresholdLow
or greater than thresholdHigh
. If neither is specified, prediction explanations will be computed for all rows.
Code samples¶
# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/predictionExplanations/ \
-H "Content-Type: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{undefined}'
Body parameter¶
{
"classNames": [
"string"
],
"datasetId": "string",
"maxExplanations": 3,
"modelId": "string",
"numTopClasses": 1,
"thresholdHigh": null,
"thresholdLow": null
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
projectId | path | string | true | The project ID |
body | body | PredictionExplanationsCreate | false | none |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | The request was accepted and will be worked on. | None |
Response Headers¶
Status | Header | Type | Format | Description |
---|---|---|---|---|
202 | Location | string | A url that can be polled to check the status. |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/projects/{projectId}/predictionExplanations/{predictionExplanationsId}/¶
Retrieve stored Prediction Explanations. Each PredictionExplanationsRow retrieved corresponds to a row of the prediction dataset, although some rows may not have had prediction explanations computed depending on the thresholds selected.
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictionExplanations/{predictionExplanationsId}/?limit=3000 \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
offset | query | integer | false | This many results will be skipped. |
limit | query | integer | true | At most this many results are returned. The default may change and a new maximum limit may be imposed without notice. |
excludeAdjustedPredictions | query | string | false | Whether to include adjusted prediction in PredictionExplanationsRow response. |
projectId | path | string | true | The project ID |
predictionExplanationsId | path | string | true | The ID of the PredictionExplanationsRecord to retrieve. |
Enumerated Values¶
Parameter | Value |
---|---|
excludeAdjustedPredictions | [false , False , true , True ] |
Example responses¶
200 Response
{
"adjustmentMethod": "string",
"count": 0,
"data": [
{
"adjustedPrediction": 0,
"adjustedPredictionValues": [
{
"label": "string",
"value": 0
}
],
"forecastDistance": 0,
"forecastPoint": "string",
"prediction": 0,
"predictionExplanations": [
{
"feature": "string",
"featureValue": "string",
"imageExplanationUrl": "string",
"label": "string",
"perNgramTextExplanations": [
{
"isUnknown": true,
"ngrams": [
{
"label": "string",
"value": 0
}
],
"qualitativateStrength": "string",
"strength": 0
}
],
"qualitativateStrength": "string",
"strength": 0
}
],
"predictionThreshold": 0,
"predictionValues": [
{
"label": "string",
"value": 0
}
],
"rowId": 0,
"seriesId": "string",
"timestamp": "string"
}
],
"id": "string",
"next": "http://example.com",
"predictionExplanationsRecordLocation": "string",
"previous": "http://example.com"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | none | PredictionExplanationsRetrieve |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/projects/{projectId}/predictionExplanationsRecords/¶
List PredictionExplanationsRecord objects for a project. These contain metadata about the computed prediction explanations and the location at which the PredictionExplanations can be retrieved.
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictionExplanationsRecords/ \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
offset | query | integer | false | This many results will be skipped. |
limit | query | integer | false | At most this many results are returned. |
modelId | query | string | false | If specified, only prediction explanations records computed for this model will be returned. |
projectId | path | string | true | The project ID |
Example responses¶
200 Response
{
"count": 0,
"data": [
{
"datasetId": "string",
"finishTime": 0,
"id": "string",
"maxExplanations": 0,
"modelId": "string",
"numColumns": 0,
"predictionExplanationsLocation": "string",
"predictionThreshold": 0,
"projectId": "string",
"thresholdHigh": 0,
"thresholdLow": 0
}
],
"next": "string",
"previous": "string"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | The object was found and returned successfully. | PredictionExplanationsRecordList |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
DELETE /api/v2/projects/{projectId}/predictionExplanationsRecords/{predictionExplanationsId}/¶
Delete saved Prediction Explanations. Deletes both the actual prediction explanations and the corresponding PredictionExplanationsRecord.
Code samples¶
# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/projects/{projectId}/predictionExplanationsRecords/{predictionExplanationsId}/ \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
projectId | path | string | true | The project ID |
predictionExplanationsId | path | string | true | The ID of the PredictionExplanationsRecord to retrieve. |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
204 | No Content | The object was deleted successfully. | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/projects/{projectId}/predictionExplanationsRecords/{predictionExplanationsId}/¶
Retrieve a PredictionExplanationsRecord object. A PredictionExplanationsRecord contains metadata about the computed prediction explanations and the location at which the PredictionExplanations can be retrieved.
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictionExplanationsRecords/{predictionExplanationsId}/ \
-H "Content-Type: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{undefined}'
Body parameter¶
{
"datasetId": "string",
"finishTime": 0,
"id": "string",
"maxExplanations": 0,
"modelId": "string",
"numColumns": 0,
"predictionExplanationsLocation": "string",
"predictionThreshold": 0,
"projectId": "string",
"thresholdHigh": 0,
"thresholdLow": 0
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
projectId | path | string | true | The project ID |
predictionExplanationsId | path | string | true | The ID of the PredictionExplanationsRecord to retrieve. |
body | body | PredictionExplanationsRecord | false | none |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | The object was found and returned successfully. | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/projects/{projectId}/predictions/¶
Get a list of prediction records.
.. deprecated:: v2.21
Use GET /api/v2/projects/{projectId}/predictionsMetadata/ instead. The only
difference is that parameter datasetId
is renamed to predictionDatasetId
both in request and response.
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictions/?offset=0&limit=1000 \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
offset | query | integer | true | This many results will be skipped |
limit | query | integer | true | At most this many results are returned. To specify no limit, use 0. The default may change and a maximum limit may be imposed without notice. |
datasetId | query | string | false | Dataset id used to create the predictions |
modelId | query | string | false | Model id |
projectId | path | string | true | The project of the predictions. |
Example responses¶
200 Response
{
"count": 0,
"data": [
{
"actualValueColumn": "string",
"datasetId": "string",
"explanationAlgorithm": "string",
"featureDerivationWindowCounts": 0,
"forecastPoint": "2019-08-24T14:15:22Z",
"id": "string",
"includesPredictionIntervals": true,
"maxExplanations": 0,
"modelId": "string",
"predictionDatasetId": "string",
"predictionIntervalsSize": 0,
"predictionThreshold": 0,
"predictionsEndDate": "2019-08-24T14:15:22Z",
"predictionsStartDate": "2019-08-24T14:15:22Z",
"projectId": "string",
"shapWarnings": {
"maxNormalizedMismatch": 0,
"mismatchRowCount": 0
},
"url": "string"
}
],
"next": "http://example.com",
"previous": "http://example.com"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | The json array of prediction metadata objects. | RetrieveListPredictionMetadataObjectsResponse |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/projects/{projectId}/predictions/¶
There are two ways of making predictions. The recommended way is to first upload your dataset to the project, and then using the corresponding datasetId, predict against that dataset. To follow that pattern, send the json request body.
Note that requesting prediction intervals will automatically trigger backtesting if backtests were not already completed for this model.
The legacy method which is deprecated is to send the file directly with the predictions request. If you need to predict against a file 10MB in size or larger, you will be required to use the above workflow for uploaded datasets. However, the following multipart/form-data can be used with small files:
:form file: a dataset to make predictions on :form modelId: the model to use to make predictions
.. note:: If using the legacy method of uploading data to this endpoint, a new dataset will be created behind the scenes. For performance reasons, it would be much better to utilize the workflow of creating the dataset first and using the supported method of making predictions of this endpoint. However, to preserve the functionality of existing workflows, the legacy method still exists.
Code samples¶
# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/predictions/ \
-H "Content-Type: application/json" \
-H "Content-Type: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{undefined}'
Body parameter¶
{
"actualValueColumn": "string",
"datasetId": "string",
"explanationAlgorithm": "shap",
"forecastPoint": "2019-08-24T14:15:22Z",
"includeFdwCounts": false,
"includePredictionIntervals": true,
"maxExplanations": 1,
"modelId": "string",
"predictionIntervalsSize": 1,
"predictionThreshold": 1,
"predictionsEndDate": "2019-08-24T14:15:22Z",
"predictionsStartDate": "2019-08-24T14:15:22Z"
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
projectId | path | string | true | The project to make predictions within. |
Content-Type | header | string | true | Content types available for making request. multipart/form-data is the legacy deprecated method to send the small file with the prediction request. |
body | body | CreatePredictionFromDataset | false | none |
Enumerated Values¶
Parameter | Value |
---|---|
Content-Type | [application/json , multipart/form-data ] |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Prediction has successfully been requested. See Location header. | None |
422 | Unprocessable Entity | The request cannot be processed. | None |
Response Headers¶
Status | Header | Type | Format | Description |
---|---|---|---|---|
202 | Location | string | A url that can be polled to check the status of the predictions as with GET /api/v2/projects/{projectId}/predictJobs/{jobId}/ |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/projects/{projectId}/predictions/{predictionId}/¶
Retrieve predictions that have previously been computed. Training predictions encoded either as JSON or CSV. If CSV output was requested, the returned CSV data will contain the following columns:
- For regression projects:
row_id
andprediction
. - For binary classification projects:
row_id
,prediction
,class_<positive_class_label>
andclass_<negative_class_label>
. - For multiclass projects:
row_id
,prediction
and aclass_<class_label>
for each class. - For multilabel projects:
row_id
and for each classprediction_<class_label>
andclass_<class_label>
. - For time-series, these additional columns will be added:
forecast_point
,forecast_distance
,timestamp
, andseries_id
.
.. minversion:: v2.21
* If `explanationAlgorithm` = 'shap', these additional columns will be added:
triplets of (`Explanation_<i>_feature_name`,
`Explanation_<i>_feature_value`, and `Explanation_<i>_strength`) for `i` ranging
from 1 to `maxExplanations`, `shap_remaining_total` and `shap_base_value`. Binary
classification projects will also have `explained_class`, the class for which
positive SHAP values imply an increased probability.
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictions/{predictionId}/ \
-H "Accept: application/json" \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
predictionId | path | string | true | The id of the prediction record to retrieve. If you have the jobId, you can retrieve the predictionId using GET /api/v2/projects/{projectId}/predictJobs/{jobId}/. |
projectId | path | string | true | The id of the project the prediction belongs to. |
Accept | header | string | false | Requested MIME type for the returned data |
Enumerated Values¶
Parameter | Value |
---|---|
Accept | [application/json , text/csv ] |
Example responses¶
200 Response
{
"actualValueColumn": "string",
"explanationAlgorithm": "string",
"featureDerivationWindowCounts": 0,
"includesPredictionIntervals": true,
"maxExplanations": 0,
"positiveClass": "string",
"predictionIntervalsSize": 0,
"predictions": [
{
"actualValue": "string",
"forecastDistance": 0,
"forecastPoint": "2019-08-24T14:15:22Z",
"originalFormatTimestamp": "string",
"positiveProbability": 0,
"prediction": 0,
"predictionExplanationMetadata": [
{
"shapRemainingTotal": 0
}
],
"predictionExplanations": [
{
"feature": "string",
"featureValue": 0,
"label": "string",
"strength": 0
}
],
"predictionIntervalLowerBound": 0,
"predictionIntervalUpperBound": 0,
"predictionThreshold": 1,
"predictionValues": [
{
"label": "string",
"threshold": 1,
"value": 0
}
],
"rowId": 0,
"segmentId": "string",
"seriesId": "string",
"target": "string",
"timestamp": "2019-08-24T14:15:22Z"
}
],
"shapBaseValue": 0,
"shapWarnings": [
{
"maxNormalizedMismatch": 0,
"mismatchRowCount": 0
}
],
"task": "Regression"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | Predictions that have previously been computed. | PredictionRetrieveResponse |
404 | Not Found | No prediction data found. | None |
Response Headers¶
Status | Header | Type | Format | Description |
---|---|---|---|---|
200 | Content-Type | string | MIME type of the returned data |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/projects/{projectId}/predictionsMetadata/¶
Use the ID of a metadata object to get the complete set of predictions.
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictionsMetadata/?offset=0&limit=1000 \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
offset | query | integer | true | This many results will be skipped |
limit | query | integer | true | At most this many results are returned. To specify no limit, use 0. The default may change and a maximum limit may be imposed without notice. |
predictionDatasetId | query | string | false | Dataset id used to create the predictions |
modelId | query | string | false | Model id |
projectId | path | string | true | The project of the predictions. |
Example responses¶
200 Response
{
"count": 0,
"data": [
{
"actualValueColumn": "string",
"datasetId": "string",
"explanationAlgorithm": "string",
"featureDerivationWindowCounts": 0,
"forecastPoint": "2019-08-24T14:15:22Z",
"id": "string",
"includesPredictionIntervals": true,
"maxExplanations": 0,
"modelId": "string",
"predictionDatasetId": "string",
"predictionIntervalsSize": 0,
"predictionThreshold": 0,
"predictionsEndDate": "2019-08-24T14:15:22Z",
"predictionsStartDate": "2019-08-24T14:15:22Z",
"projectId": "string",
"shapWarnings": {
"maxNormalizedMismatch": 0,
"mismatchRowCount": 0
},
"url": "string"
}
],
"next": "http://example.com",
"previous": "http://example.com"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | The json array of prediction metadata objects. | RetrieveListPredictionMetadataObjectsResponse |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/projects/{projectId}/predictionsMetadata/{predictionId}/¶
Use the ID of a metadata object to get the complete set of predictions.
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/predictionsMetadata/{predictionId}/ \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
predictionId | path | string | true | The id of the prediction record to retrieve. If you have the jobId, you can retrieve the predictionId using GET /api/v2/projects/{projectId}/predictJobs/{jobId}/. |
projectId | path | string | true | The id of the project the prediction belongs to. |
Example responses¶
200 Response
{
"actualValueColumn": "string",
"datasetId": "string",
"explanationAlgorithm": "string",
"featureDerivationWindowCounts": 0,
"forecastPoint": "2019-08-24T14:15:22Z",
"id": "string",
"includesPredictionIntervals": true,
"maxExplanations": 0,
"modelId": "string",
"predictionDatasetId": "string",
"predictionIntervalsSize": 0,
"predictionThreshold": 0,
"predictionsEndDate": "2019-08-24T14:15:22Z",
"predictionsStartDate": "2019-08-24T14:15:22Z",
"projectId": "string",
"shapWarnings": {
"maxNormalizedMismatch": 0,
"mismatchRowCount": 0
},
"url": "string"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | Prediction metadata object. | RetrievePredictionMetadataObject |
404 | Not Found | Training predictions not found. | None |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/projects/{projectId}/trainingPredictions/¶
Get a list of training prediction records
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/trainingPredictions/?offset=0&limit=0 \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
offset | query | integer | true | This many results will be skipped |
limit | query | integer | true | At most this many results are returned |
projectId | path | string | true | Project ID to retrieve training predictions for |
Example responses¶
200 Response
{
"count": 0,
"data": [
{
"dataSubset": "all",
"explanationAlgorithm": "shap",
"id": "string",
"maxExplanations": 100,
"modelId": "string",
"shapWarnings": [
{
"partitionName": "string",
"value": {
"maxNormalizedMismatch": 0,
"mismatchRowCount": 0
}
}
],
"url": "http://example.com"
}
],
"next": "http://example.com",
"previous": "http://example.com"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | A list of training prediction jobs | TrainingPredictionsListResponse |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/projects/{projectId}/trainingPredictions/¶
Create training data predictions
Code samples¶
# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/projects/{projectId}/trainingPredictions/ \
-H "Content-Type: application/json" \
-H "Authorization: Bearer {access-token}" \
-d '{undefined}'
Body parameter¶
{
"dataSubset": "all",
"explanationAlgorithm": "string",
"maxExplanations": 1,
"modelId": "string"
}
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
projectId | path | string | true | Project ID to compute training predictions for |
body | body | CreateTrainingPrediction | false | none |
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
202 | Accepted | Submitted successfully. See Location header. | None |
422 | Unprocessable Entity | - Model/Timeseries/Blender does not support shap based prediction explanations | |
- Error message from StackedPredictionRequestValidationError | |||
- Could not create training predictions job. Request with same parameters already submitted. | None |
Response Headers¶
Status | Header | Type | Format | Description |
---|---|---|---|---|
202 | Location | string | URL for tracking async job status. |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/projects/{projectId}/trainingPredictions/{predictionId}/¶
Retrieve training predictions that have previously been computed
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/projects/{projectId}/trainingPredictions/{predictionId}/?offset=0&limit=0 \
-H "Accept: application/json" \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
offset | query | integer | true | This many results will be skipped |
limit | query | integer | true | At most this many results are returned |
projectId | path | string | true | Project ID to retrieve training predictions for |
predictionId | path | string | true | Prediction ID to retrieve training predictions for |
Accept | header | string | false | Requested MIME type for the returned data |
Enumerated Values¶
Parameter | Value |
---|---|
Accept | [application/json , text/csv ] |
Example responses¶
200 Response
{
"count": 0,
"data": [
{
"forecastDistance": 0,
"forecastPoint": "2019-08-24T14:15:22Z",
"partitionId": "string",
"prediction": 0,
"predictionExplanations": [
{
"feature": "string",
"featureValue": 0,
"label": "string",
"strength": 0
}
],
"predictionThreshold": 1,
"predictionValues": [
{
"label": "string",
"threshold": 1,
"value": 0
}
],
"rowId": 0,
"seriesId": "string",
"shapMetadata": {
"shapBaseValue": 0,
"shapRemainingTotal": 0,
"warnings": [
{
"maxNormalizedMismatch": 0,
"mismatchRowCount": 0
}
]
},
"timestamp": "2019-08-24T14:15:22Z"
}
],
"next": "http://example.com",
"previous": "http://example.com"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | Training predictions encoded either as JSON or CSV | string |
404 | Not Found | Job does not exist or is not completed | None |
Response Headers¶
Status | Header | Type | Format | Description |
---|---|---|---|---|
200 | Content-Type | string | MIME type of the returned data |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/scheduledJobs/¶
Get a list of scheduled batch prediction jobs a user can view
Code samples¶
# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/scheduledJobs/?offset=0&limit=20 \
-H "Accept: application/json" \
-H "Authorization: Bearer {access-token}"
Parameters
Name | In | Type | Required | Description |
---|---|---|---|---|
offset | query | integer | true | The number of scheduled jobs to skip. Defaults to 0. |
limit | query | integer | true | The number of scheduled jobs (max 100) to return. Defaults to 20 |
orderBy | query | string | false | The order to sort the scheduled jobs. Defaults to order by last successful run timestamp in descending order. |
search | query | string | false | Case insensitive search against scheduled jobs name or type name. |
deploymentId | query | string | false | Filter by the prediction integration deployment ID. Ignored for non prediction integration type ID. |
typeId | query | string | false | filter by scheduled job type ID. |
queryByUser | query | string | false | Which user field to filter with. |
filterEnabled | query | string | false | Filter jobs using the enabled field. If true , only enabled jobs are returned, otherwise if false , only disabled jobs are returned. The default returns both enabled and disabled jobs. |
Enumerated Values¶
Parameter | Value |
---|---|
typeId | datasetRefresh |
queryByUser | [createdBy , updatedBy ] |
filterEnabled | [false , False , true , True ] |
Example responses¶
200 Response
{
"count": 0,
"data": [
{
"createdBy": "string",
"deploymentId": "string",
"enabled": true,
"id": "string",
"name": "string",
"schedule": {
"dayOfMonth": [
"*"
],
"dayOfWeek": [
"*"
],
"hour": [
"*"
],
"minute": [
"*"
],
"month": [
"*"
]
},
"scheduledJobId": "string",
"status": {
"lastFailedRun": "2019-08-24T14:15:22Z",
"lastSuccessfulRun": "2019-08-24T14:15:22Z",
"nextRunTime": "2019-08-24T14:15:22Z",
"queuePosition": 0,
"running": true
},
"typeId": "string",
"updatedAt": "2019-08-24T14:15:22Z"
}
],
"next": "http://example.com",
"previous": "http://example.com",
"totalCount": 0,
"updatedAt": "2019-08-24T14:15:22Z",
"updatedBy": "string"
}
Responses¶
Status | Meaning | Description | Schema |
---|---|---|---|
200 | OK | A list of scheduled batch prediction jobs | ScheduledJobsListResponse |
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
Schemas¶
ActualValueColumnInfo
{
"missingCount": 0,
"name": "string"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
missingCount | integer | true | Count of the missing values in the column. | |
name | string | true | Name of the column. |
AzureDataStreamer
{
"credentialId": "string",
"format": "csv",
"type": "azure",
"url": "string"
}
Stream CSV data chunks from Azure
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
credentialId | any | false | Either the populated value of the field or [redacted] due to permission settings |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string¦null | false | Use the specified credential to access the url |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
format | string | false | Type of input file format | |
type | string | true | Type name for this intake type | |
url | string(url) | true | URL for the CSV file |
Enumerated Values¶
Property | Value |
---|---|
anonymous | [redacted] |
format | [csv , parquet ] |
type | azure |
AzureIntake
{
"credentialId": "string",
"format": "csv",
"type": "azure",
"url": "string"
}
Stream CSV data chunks from Azure
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
credentialId | string¦null | false | Use the specified credential to access the url | |
format | string | false | Type of input file format | |
type | string | true | Type name for this intake type | |
url | string(url) | true | URL for the CSV file |
Enumerated Values¶
Property | Value |
---|---|
format | [csv , parquet ] |
type | azure |
AzureOutput
{
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
}
Save CSV data chunks to Azure Blob Storage
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
credentialId | string¦null | false | Use the specified credential to access the url | |
format | string | false | Type of output file format | |
partitionColumns | [string] | false | maxItems: 100 |
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/"). |
type | string | true | Type name for this output type | |
url | string(url) | true | URL for the file or directory |
Enumerated Values¶
Property | Value |
---|---|
format | [csv , parquet ] |
type | azure |
AzureOutputAdaptor
{
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
}
Save CSV data chunks to Azure Blob Storage
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
credentialId | any | false | Either the populated value of the field or [redacted] due to permission settings |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string¦null | false | Use the specified credential to access the url |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
format | string | false | Type of output file format | |
partitionColumns | [string] | false | maxItems: 100 |
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/"). |
type | string | true | Type name for this output type | |
url | string(url) | true | URL for the file or directory |
Enumerated Values¶
Property | Value |
---|---|
anonymous | [redacted] |
format | [csv , parquet ] |
type | azure |
AzureServicePrincipalCredentials
{
"azureTenantId": "string",
"clientId": "string",
"clientSecret": "string",
"configId": "string",
"credentialType": "azure_service_principal"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
azureTenantId | string | false | Tenant ID of the Azure AD service principal. | |
clientId | string | false | Client ID of the Azure AD service principal. | |
clientSecret | string | false | Client Secret of the Azure AD service principal. | |
configId | string | false | ID of secure configurations of credentials shared by admin. | |
credentialType | string | true | The type of these credentials, 'azure_service_principal' here. |
Enumerated Values¶
Property | Value |
---|---|
credentialType | azure_service_principal |
BasicCredentials
{
"credentialType": "basic",
"password": "string",
"user": "string"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
credentialType | string | true | The type of these credentials, 'basic' here. | |
password | string | true | The password for database authentication. The password is encrypted at rest and never saved / stored. | |
user | string | true | The username for database authentication. |
Enumerated Values¶
Property | Value |
---|---|
credentialType | basic |
BatchJobCSVSettings
{
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
}
The CSV settings used for this job
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
delimiter | any | true | CSV fields are delimited by this character. Use the string "tab" to denote TSV (TAB separated values). |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | maxLength: 1 minLength: 1 minLength: 1 |
none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
encoding | string | true | The encoding to be used for intake and output. For example (but not limited to): "shift_jis", "latin_1" or "mskanji". | |
quotechar | string | true | maxLength: 1 minLength: 1 minLength: 1 |
Fields containing the delimiter or newlines must be quoted using this character. |
Enumerated Values¶
Property | Value |
---|---|
anonymous | tab |
BatchJobCreatedBy
{
"fullName": "string",
"userId": "string",
"username": "string"
}
Who created this job
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
fullName | string¦null | true | The full name of the user who created this job (if defined by the user) | |
userId | string | true | The User ID of the user who created this job | |
username | string | true | The username (e-mail address) of the user who created this job |
BatchJobDefinitionResponse
{
"createdBy": "string",
"id": "string",
"name": "string"
}
The Batch Prediction Job Definition linking to this job, if any.
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
createdBy | string | true | The ID of creator of this job definition | |
id | string | true | The ID of the Batch Prediction job definition | |
name | string | true | A human-readable name for the definition, must be unique across organisations |
BatchJobLinks
{
"csvUpload": "string",
"download": "string",
"self": "string"
}
Links useful for this job
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
csvUpload | string(url) | false | The URL used to upload the dataset for this job. Only available for localFile intake. | |
download | string¦null | false | The URL used to download the results from this job. Only available for localFile outputs. Will be null if the download is not yet available. | |
self | string(url) | true | The URL used access this job. |
BatchJobListResponse
{
"count": 0,
"data": [
{
"batchMonitoringJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"batchPredictionJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"elapsedTimeSec": 0,
"failedRows": 0,
"hidden": "2019-08-24T14:15:22Z",
"id": "string",
"intakeDatasetDisplayName": "string",
"jobIntakeSize": 0,
"jobOutputSize": 0,
"jobSpec": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"maxNgramExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringAggregation": {
"retentionPolicy": "samples",
"retentionValue": 0
},
"monitoringBatchPrefix": "string",
"monitoringColumns": {
"actedUponColumn": "string",
"actualsTimestampColumn": "string",
"actualsValueColumn": "string",
"associationIdColumn": "string",
"customMetricId": "string",
"customMetricTimestampColumn": "string",
"customMetricTimestampFormat": "string",
"customMetricValueColumn": "string",
"monitoredStatusColumn": "string",
"predictionsColumns": [
{
"className": "string",
"columnName": "string"
}
],
"reportDrift": true,
"reportPredictions": true,
"uniqueRowIdentifierColumns": [
"string"
]
},
"monitoringOutputSettings": {
"monitoredStatusColumn": "string",
"uniqueRowIdentifierColumns": [
"string"
]
},
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"links": {
"csvUpload": "string",
"download": "string",
"self": "string"
},
"logs": [
"string"
],
"monitoringBatchId": "string",
"percentageCompleted": 100,
"queuePosition": 0,
"queued": true,
"resultsDeleted": true,
"scoredRows": 0,
"skippedRows": 0,
"source": "string",
"status": "INITIALIZING",
"statusDetails": "string"
}
],
"next": "http://example.com",
"previous": "http://example.com",
"totalCount": 0
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
count | integer | false | Number of items returned on this page. | |
data | [BatchJobResponse] | true | maxItems: 10000 |
An array of jobs |
next | string(uri)¦null | true | URL pointing to the next page (if null, there is no next page). | |
previous | string(uri)¦null | true | URL pointing to the previous page (if null, there is no previous page). | |
totalCount | integer | true | The total number of items across all pages. |
BatchJobPredictionInstance
{
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
}
Override the default prediction instance from the deployment when scoring this job.
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
apiKey | string | false | By default, prediction requests will use the API key of the user that created the job. This allows you to make requests on behalf of other users. | |
datarobotKey | string | false | If running a job against a prediction instance in the Managed AI Cloud, you must provide the organization level DataRobot-Key. | |
hostName | string | true | Override the default host name of the deployment with this. | |
sslEnabled | boolean | true | Use SSL (HTTPS) when communicating with the overriden prediction server. |
BatchJobRemapping
{
"inputName": "string",
"outputName": "string"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
inputName | string | true | Rename column with this name | |
outputName | string¦null | true | Rename column to this name (leave as null to remove from the output) |
BatchJobResponse
{
"batchMonitoringJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"batchPredictionJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"elapsedTimeSec": 0,
"failedRows": 0,
"hidden": "2019-08-24T14:15:22Z",
"id": "string",
"intakeDatasetDisplayName": "string",
"jobIntakeSize": 0,
"jobOutputSize": 0,
"jobSpec": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"maxNgramExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringAggregation": {
"retentionPolicy": "samples",
"retentionValue": 0
},
"monitoringBatchPrefix": "string",
"monitoringColumns": {
"actedUponColumn": "string",
"actualsTimestampColumn": "string",
"actualsValueColumn": "string",
"associationIdColumn": "string",
"customMetricId": "string",
"customMetricTimestampColumn": "string",
"customMetricTimestampFormat": "string",
"customMetricValueColumn": "string",
"monitoredStatusColumn": "string",
"predictionsColumns": [
{
"className": "string",
"columnName": "string"
}
],
"reportDrift": true,
"reportPredictions": true,
"uniqueRowIdentifierColumns": [
"string"
]
},
"monitoringOutputSettings": {
"monitoredStatusColumn": "string",
"uniqueRowIdentifierColumns": [
"string"
]
},
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"links": {
"csvUpload": "string",
"download": "string",
"self": "string"
},
"logs": [
"string"
],
"monitoringBatchId": "string",
"percentageCompleted": 100,
"queuePosition": 0,
"queued": true,
"resultsDeleted": true,
"scoredRows": 0,
"skippedRows": 0,
"source": "string",
"status": "INITIALIZING",
"statusDetails": "string"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
batchMonitoringJobDefinition | BatchJobDefinitionResponse | false | The Batch Prediction Job Definition linking to this job, if any. | |
batchPredictionJobDefinition | BatchJobDefinitionResponse | false | The Batch Prediction Job Definition linking to this job, if any. | |
created | string(date-time) | true | When was this job created | |
createdBy | BatchJobCreatedBy | true | Who created this job | |
elapsedTimeSec | integer | true | minimum: 0 |
Number of seconds the job has been processing for |
failedRows | integer | true | minimum: 0 |
Number of rows that have failed scoring |
hidden | string(date-time) | false | When was this job was hidden last, blank if visible | |
id | string | true | The ID of the Batch job | |
intakeDatasetDisplayName | string¦null | false | If applicable (e.g. for AI catalog), will contain the dataset name used for the intake dataset. | |
jobIntakeSize | integer¦null | true | minimum: 0 |
Number of bytes in the intake dataset for this job |
jobOutputSize | integer¦null | true | minimum: 0 |
Number of bytes in the output dataset for this job |
jobSpec | BatchJobSpecResponse | true | The job configuration used to create this job | |
links | BatchJobLinks | true | Links useful for this job | |
logs | [string] | true | The job log. | |
monitoringBatchId | string¦null | true | Id of the monitoring batch created by this job. Only present if the job runs on a deployment with batch monitoring enabled. | |
percentageCompleted | number | true | maximum: 100 minimum: 0 |
Indicates job progress which is based on number of already processed rows in dataset |
queuePosition | integer¦null | false | minimum: 0 |
To ensure a dedicated prediction instance is not overloaded, only one job will be run against it at a time. This is the number of jobs that are awaiting processing before this job start running. May not be available in all environments. |
queued | boolean | true | The job has been put on the queue for execution. | |
resultsDeleted | boolean | false | Indicates if the job was subject to garbage collection and had its artifacts deleted (output files, if any, and scoring data on local storage) | |
scoredRows | integer | true | minimum: 0 |
Number of rows that have been used in prediction computation |
skippedRows | integer | true | minimum: 0 |
Number of rows that have been skipped during scoring. May contain non-zero value only in time-series predictions case if provided dataset contains more than required historical rows. |
source | string | false | Source from which batch job was started | |
status | string | true | The current job status | |
statusDetails | string | true | Explanation for current status |
Enumerated Values¶
Property | Value |
---|---|
status | [INITIALIZING , RUNNING , COMPLETED , ABORTED , FAILED ] |
BatchJobSpecResponse
{
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"maxNgramExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringAggregation": {
"retentionPolicy": "samples",
"retentionValue": 0
},
"monitoringBatchPrefix": "string",
"monitoringColumns": {
"actedUponColumn": "string",
"actualsTimestampColumn": "string",
"actualsValueColumn": "string",
"associationIdColumn": "string",
"customMetricId": "string",
"customMetricTimestampColumn": "string",
"customMetricTimestampFormat": "string",
"customMetricValueColumn": "string",
"monitoredStatusColumn": "string",
"predictionsColumns": [
{
"className": "string",
"columnName": "string"
}
],
"reportDrift": true,
"reportPredictions": true,
"uniqueRowIdentifierColumns": [
"string"
]
},
"monitoringOutputSettings": {
"monitoredStatusColumn": "string",
"uniqueRowIdentifierColumns": [
"string"
]
},
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
}
The job configuration used to create this job
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
abortOnError | boolean | true | Should this job abort if too many errors are encountered | |
batchJobType | string | false | Batch job type. | |
chunkSize | any | false | Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes. |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | integer | false | maximum: 41943040 minimum: 20 |
none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
columnNamesRemapping | any | false | Remap (rename or remove columns from) the output from this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | object | false | Provide a dictionary with key/value pairs to remap (deprecated) |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | [BatchJobRemapping] | false | maxItems: 1000 |
Provide a list of items to remap |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
csvSettings | BatchJobCSVSettings | true | The CSV settings used for this job | |
deploymentId | string | false | ID of deployment which is used in job for processing predictions dataset | |
disableRowLevelErrorHandling | boolean | true | Skip row by row error handling | |
explanationAlgorithm | string | false | Which algorithm will be used to calculate prediction explanations | |
explanationClassNames | [string] | false | maxItems: 10 minItems: 1 |
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1 |
explanationNumTopClasses | integer | false | maximum: 10 minimum: 1 |
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1 |
includePredictionStatus | boolean | true | Include prediction status column in the output | |
includeProbabilities | boolean | true | Include probabilities for all classes | |
includeProbabilitiesClasses | [string] | true | maxItems: 100 |
Include only probabilities for these specific class names. |
intakeSettings | any | true | The response option configured for this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | AzureDataStreamer | false | Stream CSV data chunks from Azure |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | DataStageDataStreamer | false | Stream CSV data chunks from data stage storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | CatalogDataStreamer | false | Stream CSV data chunks from AI catalog dataset |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | GCPDataStreamer | false | Stream CSV data chunks from Google Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BigQueryDataStreamer | false | Stream CSV data chunks from Big Query using GCS |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | S3DataStreamer | false | Stream CSV data chunks from Amazon Cloud Storage S3 |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SnowflakeDataStreamer | false | Stream CSV data chunks from Snowflake |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SynapseDataStreamer | false | Stream CSV data chunks from Azure Synapse |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | DSSDataStreamer | false | Stream CSV data chunks from DSS dataset |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | FileSystemDataStreamer | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | HTTPDataStreamer | false | Stream CSV data chunks from HTTP |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | JDBCDataStreamer | false | Stream CSV data chunks from JDBC |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | LocalFileDataStreamer | false | Stream CSV data chunks from local file storage |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
maxExplanations | integer | true | maximum: 100 minimum: 0 |
Number of explanations requested. Will be ordered by strength. |
maxNgramExplanations | any | false | The maximum number of text ngram explanations to supply per row of the dataset. The default recommended maxNgramExplanations is all (no limit) |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | integer | false | minimum: 0 |
none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
modelId | string | false | ID of leaderboard model which is used in job for processing predictions dataset | |
modelPackageId | string | false | ID of model package from registry is used in job for processing predictions dataset | |
monitoringAggregation | MonitoringAggregation | false | Defines the aggregation policy for monitoring jobs. | |
monitoringBatchPrefix | string¦null | false | Name of the batch to create with this job | |
monitoringColumns | MonitoringColumnsMapping | false | Column names mapping for monitoring | |
monitoringOutputSettings | MonitoringOutputSettings | false | Output settings for monitoring jobs | |
numConcurrent | integer | false | minimum: 1 |
Number of simultaneous requests to run against the prediction instance |
outputSettings | any | false | The response option configured for this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | AzureOutputAdaptor | false | Save CSV data chunks to Azure Blob Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | GCPOutputAdaptor | false | Save CSV data chunks to Google Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BigQueryOutputAdaptor | false | Save CSV data chunks to Google BigQuery in bulk |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | S3OutputAdaptor | false | Saves CSV data chunks to Amazon Cloud Storage S3 |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SnowflakeOutputAdaptor | false | Save CSV data chunks to Snowflake in bulk |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SynapseOutputAdaptor | false | Save CSV data chunks to Azure Synapse in bulk |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | FileSystemOutputAdaptor | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | HttpOutputAdaptor | false | Save CSV data chunks to HTTP data endpoint |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | JdbcOutputAdaptor | false | Save CSV data chunks via JDBC |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | LocalFileOutputAdaptor | false | Save CSV data chunks to local file storage |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
passthroughColumns | [string] | false | maxItems: 100 |
Pass through columns from the original dataset |
passthroughColumnsSet | string | false | Pass through all columns from the original dataset | |
pinnedModelId | string | false | Specify a model ID used for scoring | |
predictionInstance | BatchJobPredictionInstance | false | Override the default prediction instance from the deployment when scoring this job. | |
predictionWarningEnabled | boolean¦null | false | Enable prediction warnings. | |
redactedFields | [string] | true | A list of qualified field names from intake- and/or outputSettings that was redacted due to permissions and sharing settings. For example: intakeSettings.dataStoreId | |
skipDriftTracking | boolean | true | Skip drift tracking for this job. | |
thresholdHigh | number | false | Compute explanations for predictions above this threshold | |
thresholdLow | number | false | Compute explanations for predictions below this threshold | |
timeseriesSettings | any | false | Time Series settings included of this job is a Time Series job. |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchJobTimeSeriesSettingsForecast | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchJobTimeSeriesSettingsHistorical | false | none |
Enumerated Values¶
Property | Value |
---|---|
batchJobType | [monitoring , prediction ] |
anonymous | [auto , fixed , dynamic ] |
explanationAlgorithm | [shap , xemp ] |
anonymous | all |
passthroughColumnsSet | all |
BatchJobTimeSeriesSettingsForecast
{
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
forecastPoint | string(date-time) | false | Used for forecast predictions in order to override the inferred forecast point from the dataset. | |
relaxKnownInAdvanceFeaturesCheck | boolean | false | If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed. | |
type | string | true | Forecast mode makes predictions using forecastPoint or rows in the dataset without target. |
Enumerated Values¶
Property | Value |
---|---|
type | forecast |
BatchJobTimeSeriesSettingsHistorical
{
"predictionsEndDate": "2019-08-24T14:15:22Z",
"predictionsStartDate": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "historical"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
predictionsEndDate | string(date-time) | false | Used for historical predictions in order to override date to which predictions should be calculated. By default value will be inferred automatically from the dataset. | |
predictionsStartDate | string(date-time) | false | Used for historical predictions in order to override date from which predictions should be calculated. By default value will be inferred automatically from the dataset. | |
relaxKnownInAdvanceFeaturesCheck | boolean | false | If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed. | |
type | string | true | Historical mode enables bulk predictions which calculates predictions for all possible forecast points and forecast distances in the dataset within the predictionsStartDate/predictionsEndDate range. |
Enumerated Values¶
Property | Value |
---|---|
type | historical |
BatchPredictionCreatedBy
{
"fullName": "string",
"userId": "string",
"username": "string"
}
Who created this job
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
fullName | string¦null | true | The full name of the user who created this job (if defined by the user) | |
userId | string | true | The User ID of the user who created this job | |
username | string | true | The username (e-mail address) of the user who created this job |
BatchPredictionJobCSVSettings
{
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
}
The CSV settings used for this job
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
delimiter | any | true | CSV fields are delimited by this character. Use the string "tab" to denote TSV (TAB separated values). |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | maxLength: 1 minLength: 1 minLength: 1 |
none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
encoding | string | true | The encoding to be used for intake and output. For example (but not limited to): "shift_jis", "latin_1" or "mskanji". | |
quotechar | string | true | maxLength: 1 minLength: 1 minLength: 1 |
Fields containing the delimiter or newlines must be quoted using this character. |
Enumerated Values¶
Property | Value |
---|---|
anonymous | tab |
BatchPredictionJobCreate
{
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringBatchPrefix": "string",
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionThreshold": 1,
"predictionWarningEnabled": true,
"secondaryDatasetsConfigId": "string",
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
abortOnError | boolean | true | Should this job abort if too many errors are encountered | |
batchJobType | string | false | Batch job type. | |
chunkSize | any | false | Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes. |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | integer | false | maximum: 41943040 minimum: 20 |
none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
columnNamesRemapping | any | false | Remap (rename or remove columns from) the output from this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | object | false | Provide a dictionary with key/value pairs to remap (deprecated) |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | [BatchPredictionJobRemapping] | false | maxItems: 1000 |
Provide a list of items to remap |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
csvSettings | BatchPredictionJobCSVSettings | true | The CSV settings used for this job | |
deploymentId | string | false | ID of deployment which is used in job for processing predictions dataset | |
disableRowLevelErrorHandling | boolean | true | Skip row by row error handling | |
explanationAlgorithm | string | false | Which algorithm will be used to calculate prediction explanations | |
explanationClassNames | [string] | false | maxItems: 10 minItems: 1 |
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1 |
explanationNumTopClasses | integer | false | maximum: 10 minimum: 1 |
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1 |
includePredictionStatus | boolean | true | Include prediction status column in the output | |
includeProbabilities | boolean | true | Include probabilities for all classes | |
includeProbabilitiesClasses | [string] | true | maxItems: 100 |
Include only probabilities for these specific class names. |
intakeSettings | any | true | The intake option configured for this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | AzureIntake | false | Stream CSV data chunks from Azure |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BigQueryIntake | false | Stream CSV data chunks from Big Query using GCS |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | DataStageIntake | false | Stream CSV data chunks from data stage storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | Catalog | false | Stream CSV data chunks from AI catalog dataset |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | DSS | false | Stream CSV data chunks from DSS dataset |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | FileSystemIntake | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | GCPIntake | false | Stream CSV data chunks from Google Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | HTTPIntake | false | Stream CSV data chunks from HTTP |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | JDBCIntake | false | Stream CSV data chunks from JDBC |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | LocalFileIntake | false | Stream CSV data chunks from local file storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | S3Intake | false | Stream CSV data chunks from Amazon Cloud Storage S3 |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SnowflakeIntake | false | Stream CSV data chunks from Snowflake |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SynapseIntake | false | Stream CSV data chunks from Azure Synapse |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
maxExplanations | integer | true | maximum: 100 minimum: 0 |
Number of explanations requested. Will be ordered by strength. |
modelId | string | false | ID of leaderboard model which is used in job for processing predictions dataset | |
modelPackageId | string | false | ID of model package from registry is used in job for processing predictions dataset | |
monitoringBatchPrefix | string¦null | false | Name of the batch to create with this job | |
numConcurrent | integer | false | minimum: 1 |
Number of simultaneous requests to run against the prediction instance |
outputSettings | any | false | The output option configured for this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | AzureOutput | false | Save CSV data chunks to Azure Blob Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BigQueryOutput | false | Save CSV data chunks to Google BigQuery in bulk |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | FileSystemOutput | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | GCPOutput | false | Save CSV data chunks to Google Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | HTTPOutput | false | Save CSV data chunks to HTTP data endpoint |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | JDBCOutput | false | Save CSV data chunks via JDBC |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | LocalFileOutput | false | Save CSV data chunks to local file storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | S3Output | false | Saves CSV data chunks to Amazon Cloud Storage S3 |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SnowflakeOutput | false | Save CSV data chunks to Snowflake in bulk |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SynapseOutput | false | Save CSV data chunks to Azure Synapse in bulk |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
passthroughColumns | [string] | false | maxItems: 100 |
Pass through columns from the original dataset |
passthroughColumnsSet | string | false | Pass through all columns from the original dataset | |
pinnedModelId | string | false | Specify a model ID used for scoring | |
predictionInstance | BatchPredictionJobPredictionInstance | false | Override the default prediction instance from the deployment when scoring this job. | |
predictionThreshold | number | false | maximum: 1 minimum: 0 |
Threshold is the point that sets the class boundary for a predicted value. The model classifies an observation below the threshold as FALSE, and an observation above the threshold as TRUE. In other words, DataRobot automatically assigns the positive class label to any prediction exceeding the threshold. This value can be set between 0.0 and 1.0. |
predictionWarningEnabled | boolean¦null | false | Enable prediction warnings. | |
secondaryDatasetsConfigId | string | false | Configuration id for secondary datasets to use when making a prediction. | |
skipDriftTracking | boolean | true | Skip drift tracking for this job. | |
thresholdHigh | number | false | Compute explanations for predictions above this threshold | |
thresholdLow | number | false | Compute explanations for predictions below this threshold | |
timeseriesSettings | any | false | Time Series settings included of this job is a Time Series job. |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchPredictionJobTimeSeriesSettingsForecast | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchPredictionJobTimeSeriesSettingsHistorical | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchPredictionJobTimeSeriesSettingsTraining | false | none |
Enumerated Values¶
Property | Value |
---|---|
batchJobType | [monitoring , prediction ] |
anonymous | [auto , fixed , dynamic ] |
explanationAlgorithm | [shap , xemp ] |
passthroughColumnsSet | all |
BatchPredictionJobDefinitionId
{
"jobDefinitionId": "string"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
jobDefinitionId | string | true | ID of the Batch Prediction job definition |
BatchPredictionJobDefinitionJobSpecResponse
{
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"maxNgramExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringAggregation": {
"retentionPolicy": "samples",
"retentionValue": 0
},
"monitoringBatchPrefix": "string",
"monitoringColumns": {
"actedUponColumn": "string",
"actualsTimestampColumn": "string",
"actualsValueColumn": "string",
"associationIdColumn": "string",
"customMetricId": "string",
"customMetricTimestampColumn": "string",
"customMetricTimestampFormat": "string",
"customMetricValueColumn": "string",
"monitoredStatusColumn": "string",
"predictionsColumns": [
{
"className": "string",
"columnName": "string"
}
],
"reportDrift": true,
"reportPredictions": true,
"uniqueRowIdentifierColumns": [
"string"
]
},
"monitoringOutputSettings": {
"monitoredStatusColumn": "string",
"uniqueRowIdentifierColumns": [
"string"
]
},
"numConcurrent": 0,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
}
The Batch Prediction Job specification to be put on the queue in intervals
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
abortOnError | boolean | true | Should this job abort if too many errors are encountered | |
batchJobType | string | false | Batch job type. | |
chunkSize | any | false | Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes. |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | integer | false | maximum: 41943040 minimum: 20 |
none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
columnNamesRemapping | any | false | Remap (rename or remove columns from) the output from this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | object | false | Provide a dictionary with key/value pairs to remap (deprecated) |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | [BatchJobRemapping] | false | maxItems: 1000 |
Provide a list of items to remap |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
csvSettings | BatchJobCSVSettings | true | The CSV settings used for this job | |
deploymentId | string | false | ID of deployment which is used in job for processing predictions dataset | |
disableRowLevelErrorHandling | boolean | true | Skip row by row error handling | |
explanationAlgorithm | string | false | Which algorithm will be used to calculate prediction explanations | |
explanationClassNames | [string] | false | maxItems: 10 minItems: 1 |
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1 |
explanationNumTopClasses | integer | false | maximum: 10 minimum: 1 |
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1 |
includePredictionStatus | boolean | true | Include prediction status column in the output | |
includeProbabilities | boolean | true | Include probabilities for all classes | |
includeProbabilitiesClasses | [string] | true | maxItems: 100 |
Include only probabilities for these specific class names. |
intakeSettings | any | true | The response option configured for this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | AzureDataStreamer | false | Stream CSV data chunks from Azure |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | DataStageDataStreamer | false | Stream CSV data chunks from data stage storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | CatalogDataStreamer | false | Stream CSV data chunks from AI catalog dataset |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | GCPDataStreamer | false | Stream CSV data chunks from Google Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BigQueryDataStreamer | false | Stream CSV data chunks from Big Query using GCS |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | S3DataStreamer | false | Stream CSV data chunks from Amazon Cloud Storage S3 |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SnowflakeDataStreamer | false | Stream CSV data chunks from Snowflake |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SynapseDataStreamer | false | Stream CSV data chunks from Azure Synapse |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | DSSDataStreamer | false | Stream CSV data chunks from DSS dataset |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | FileSystemDataStreamer | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | HTTPDataStreamer | false | Stream CSV data chunks from HTTP |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | JDBCDataStreamer | false | Stream CSV data chunks from JDBC |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | LocalFileDataStreamer | false | Stream CSV data chunks from local file storage |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
maxExplanations | integer | true | maximum: 100 minimum: 0 |
Number of explanations requested. Will be ordered by strength. |
maxNgramExplanations | any | false | The maximum number of text ngram explanations to supply per row of the dataset. The default recommended maxNgramExplanations is all (no limit) |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | integer | false | minimum: 0 |
none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
modelId | string | false | ID of leaderboard model which is used in job for processing predictions dataset | |
modelPackageId | string | false | ID of model package from registry is used in job for processing predictions dataset | |
monitoringAggregation | MonitoringAggregation | false | Defines the aggregation policy for monitoring jobs. | |
monitoringBatchPrefix | string¦null | false | Name of the batch to create with this job | |
monitoringColumns | MonitoringColumnsMapping | false | Column names mapping for monitoring | |
monitoringOutputSettings | MonitoringOutputSettings | false | Output settings for monitoring jobs | |
numConcurrent | integer | true | minimum: 0 |
Number of simultaneous requests to run against the prediction instance |
outputSettings | any | false | The response option configured for this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | AzureOutputAdaptor | false | Save CSV data chunks to Azure Blob Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | GCPOutputAdaptor | false | Save CSV data chunks to Google Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BigQueryOutputAdaptor | false | Save CSV data chunks to Google BigQuery in bulk |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | S3OutputAdaptor | false | Saves CSV data chunks to Amazon Cloud Storage S3 |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SnowflakeOutputAdaptor | false | Save CSV data chunks to Snowflake in bulk |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SynapseOutputAdaptor | false | Save CSV data chunks to Azure Synapse in bulk |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | FileSystemOutputAdaptor | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | HttpOutputAdaptor | false | Save CSV data chunks to HTTP data endpoint |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | JdbcOutputAdaptor | false | Save CSV data chunks via JDBC |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | LocalFileOutputAdaptor | false | Save CSV data chunks to local file storage |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
passthroughColumns | [string] | false | maxItems: 100 |
Pass through columns from the original dataset |
passthroughColumnsSet | string | false | Pass through all columns from the original dataset | |
pinnedModelId | string | false | Specify a model ID used for scoring | |
predictionInstance | BatchJobPredictionInstance | false | Override the default prediction instance from the deployment when scoring this job. | |
predictionWarningEnabled | boolean¦null | false | Enable prediction warnings. | |
redactedFields | [string] | true | A list of qualified field names from intake- and/or outputSettings that was redacted due to permissions and sharing settings. For example: intakeSettings.dataStoreId | |
skipDriftTracking | boolean | true | Skip drift tracking for this job. | |
thresholdHigh | number | false | Compute explanations for predictions above this threshold | |
thresholdLow | number | false | Compute explanations for predictions below this threshold | |
timeseriesSettings | any | false | Time Series settings included of this job is a Time Series job. |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchJobTimeSeriesSettingsForecast | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchPredictionJobTimeSeriesSettingsForecastWithPolicy | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchJobTimeSeriesSettingsHistorical | false | none |
Enumerated Values¶
Property | Value |
---|---|
batchJobType | [monitoring , prediction ] |
anonymous | [auto , fixed , dynamic ] |
explanationAlgorithm | [shap , xemp ] |
anonymous | all |
passthroughColumnsSet | all |
BatchPredictionJobDefinitionResponse
{
"createdBy": "string",
"id": "string",
"name": "string"
}
The Batch Prediction Job Definition linking to this job, if any.
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
createdBy | string | true | The ID of creator of this job definition | |
id | string | true | The ID of the Batch Prediction job definition | |
name | string | true | A human-readable name for the definition, must be unique across organisations |
BatchPredictionJobDefinitionsCreate
{
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"enabled": true,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringBatchPrefix": "string",
"name": "string",
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionThreshold": 1,
"predictionWarningEnabled": true,
"schedule": {
"dayOfMonth": [
"*"
],
"dayOfWeek": [
"*"
],
"hour": [
"*"
],
"minute": [
"*"
],
"month": [
"*"
]
},
"secondaryDatasetsConfigId": "string",
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
abortOnError | boolean | true | Should this job abort if too many errors are encountered | |
batchJobType | string | false | Batch job type. | |
chunkSize | any | false | Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes. |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | integer | false | maximum: 41943040 minimum: 20 |
none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
columnNamesRemapping | any | false | Remap (rename or remove columns from) the output from this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | object | false | Provide a dictionary with key/value pairs to remap (deprecated) |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | [BatchPredictionJobRemapping] | false | maxItems: 1000 |
Provide a list of items to remap |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
csvSettings | BatchPredictionJobCSVSettings | true | The CSV settings used for this job | |
deploymentId | string | true | ID of deployment which is used in job for processing predictions dataset | |
disableRowLevelErrorHandling | boolean | true | Skip row by row error handling | |
enabled | boolean | false | If this job definition is enabled as a scheduled job. Optional if no schedule is supplied. | |
explanationAlgorithm | string | false | Which algorithm will be used to calculate prediction explanations | |
explanationClassNames | [string] | false | maxItems: 10 minItems: 1 |
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1 |
explanationNumTopClasses | integer | false | maximum: 10 minimum: 1 |
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1 |
includePredictionStatus | boolean | true | Include prediction status column in the output | |
includeProbabilities | boolean | true | Include probabilities for all classes | |
includeProbabilitiesClasses | [string] | true | maxItems: 100 |
Include only probabilities for these specific class names. |
intakeSettings | any | true | The intake option configured for this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | AzureIntake | false | Stream CSV data chunks from Azure |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BigQueryIntake | false | Stream CSV data chunks from Big Query using GCS |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | DataStageIntake | false | Stream CSV data chunks from data stage storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | Catalog | false | Stream CSV data chunks from AI catalog dataset |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | DSS | false | Stream CSV data chunks from DSS dataset |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | FileSystemIntake | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | GCPIntake | false | Stream CSV data chunks from Google Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | HTTPIntake | false | Stream CSV data chunks from HTTP |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | JDBCIntake | false | Stream CSV data chunks from JDBC |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | LocalFileIntake | false | Stream CSV data chunks from local file storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | S3Intake | false | Stream CSV data chunks from Amazon Cloud Storage S3 |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SnowflakeIntake | false | Stream CSV data chunks from Snowflake |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SynapseIntake | false | Stream CSV data chunks from Azure Synapse |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
maxExplanations | integer | true | maximum: 100 minimum: 0 |
Number of explanations requested. Will be ordered by strength. |
modelId | string | false | ID of leaderboard model which is used in job for processing predictions dataset | |
modelPackageId | string | false | ID of model package from registry is used in job for processing predictions dataset | |
monitoringBatchPrefix | string¦null | false | Name of the batch to create with this job | |
name | string | false | maxLength: 100 minLength: 1 minLength: 1 |
A human-readable name for the definition, must be unique across organisations, if left out the backend will generate one for you. |
numConcurrent | integer | false | minimum: 1 |
Number of simultaneous requests to run against the prediction instance |
outputSettings | any | false | The output option configured for this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | AzureOutput | false | Save CSV data chunks to Azure Blob Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BigQueryOutput | false | Save CSV data chunks to Google BigQuery in bulk |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | FileSystemOutput | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | GCPOutput | false | Save CSV data chunks to Google Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | HTTPOutput | false | Save CSV data chunks to HTTP data endpoint |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | JDBCOutput | false | Save CSV data chunks via JDBC |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | LocalFileOutput | false | Save CSV data chunks to local file storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | S3Output | false | Saves CSV data chunks to Amazon Cloud Storage S3 |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SnowflakeOutput | false | Save CSV data chunks to Snowflake in bulk |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SynapseOutput | false | Save CSV data chunks to Azure Synapse in bulk |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
passthroughColumns | [string] | false | maxItems: 100 |
Pass through columns from the original dataset |
passthroughColumnsSet | string | false | Pass through all columns from the original dataset | |
pinnedModelId | string | false | Specify a model ID used for scoring | |
predictionInstance | BatchPredictionJobPredictionInstance | false | Override the default prediction instance from the deployment when scoring this job. | |
predictionThreshold | number | false | maximum: 1 minimum: 0 |
Threshold is the point that sets the class boundary for a predicted value. The model classifies an observation below the threshold as FALSE, and an observation above the threshold as TRUE. In other words, DataRobot automatically assigns the positive class label to any prediction exceeding the threshold. This value can be set between 0.0 and 1.0. |
predictionWarningEnabled | boolean¦null | false | Enable prediction warnings. | |
schedule | Schedule | false | The scheduling information defining how often and when to execute this job to the Job Scheduling service. Optional if enabled = False. | |
secondaryDatasetsConfigId | string | false | Configuration id for secondary datasets to use when making a prediction. | |
skipDriftTracking | boolean | true | Skip drift tracking for this job. | |
thresholdHigh | number | false | Compute explanations for predictions above this threshold | |
thresholdLow | number | false | Compute explanations for predictions below this threshold | |
timeseriesSettings | any | false | Time Series settings included of this job is a Time Series job. |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchJobTimeSeriesSettingsForecast | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchPredictionJobTimeSeriesSettingsForecastWithPolicy | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchJobTimeSeriesSettingsHistorical | false | none |
Enumerated Values¶
Property | Value |
---|---|
batchJobType | [monitoring , prediction ] |
anonymous | [auto , fixed , dynamic ] |
explanationAlgorithm | [shap , xemp ] |
passthroughColumnsSet | all |
BatchPredictionJobDefinitionsListResponse
{
"count": 0,
"data": [
{
"batchPredictionJob": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"maxNgramExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringAggregation": {
"retentionPolicy": "samples",
"retentionValue": 0
},
"monitoringBatchPrefix": "string",
"monitoringColumns": {
"actedUponColumn": "string",
"actualsTimestampColumn": "string",
"actualsValueColumn": "string",
"associationIdColumn": "string",
"customMetricId": "string",
"customMetricTimestampColumn": "string",
"customMetricTimestampFormat": "string",
"customMetricValueColumn": "string",
"monitoredStatusColumn": "string",
"predictionsColumns": [
{
"className": "string",
"columnName": "string"
}
],
"reportDrift": true,
"reportPredictions": true,
"uniqueRowIdentifierColumns": [
"string"
]
},
"monitoringOutputSettings": {
"monitoredStatusColumn": "string",
"uniqueRowIdentifierColumns": [
"string"
]
},
"numConcurrent": 0,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"enabled": false,
"id": "string",
"lastFailedRunTime": "2019-08-24T14:15:22Z",
"lastScheduledRunTime": "2019-08-24T14:15:22Z",
"lastStartedJobStatus": "INITIALIZING",
"lastStartedJobTime": "2019-08-24T14:15:22Z",
"lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
"name": "string",
"nextScheduledRunTime": "2019-08-24T14:15:22Z",
"schedule": {
"dayOfMonth": [
"*"
],
"dayOfWeek": [
"*"
],
"hour": [
"*"
],
"minute": [
"*"
],
"month": [
"*"
]
},
"updated": "2019-08-24T14:15:22Z",
"updatedBy": {
"fullName": "string",
"userId": "string",
"username": "string"
}
}
],
"next": "http://example.com",
"previous": "http://example.com",
"totalCount": 0
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
count | integer | false | Number of items returned on this page. | |
data | [BatchPredictionJobDefinitionsResponse] | true | An array of scheduled jobs | |
next | string(uri)¦null | true | URL pointing to the next page (if null, there is no next page). | |
previous | string(uri)¦null | true | URL pointing to the previous page (if null, there is no previous page). | |
totalCount | integer | true | The total number of items across all pages. |
BatchPredictionJobDefinitionsResponse
{
"batchPredictionJob": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"maxNgramExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringAggregation": {
"retentionPolicy": "samples",
"retentionValue": 0
},
"monitoringBatchPrefix": "string",
"monitoringColumns": {
"actedUponColumn": "string",
"actualsTimestampColumn": "string",
"actualsValueColumn": "string",
"associationIdColumn": "string",
"customMetricId": "string",
"customMetricTimestampColumn": "string",
"customMetricTimestampFormat": "string",
"customMetricValueColumn": "string",
"monitoredStatusColumn": "string",
"predictionsColumns": [
{
"className": "string",
"columnName": "string"
}
],
"reportDrift": true,
"reportPredictions": true,
"uniqueRowIdentifierColumns": [
"string"
]
},
"monitoringOutputSettings": {
"monitoredStatusColumn": "string",
"uniqueRowIdentifierColumns": [
"string"
]
},
"numConcurrent": 0,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"enabled": false,
"id": "string",
"lastFailedRunTime": "2019-08-24T14:15:22Z",
"lastScheduledRunTime": "2019-08-24T14:15:22Z",
"lastStartedJobStatus": "INITIALIZING",
"lastStartedJobTime": "2019-08-24T14:15:22Z",
"lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
"name": "string",
"nextScheduledRunTime": "2019-08-24T14:15:22Z",
"schedule": {
"dayOfMonth": [
"*"
],
"dayOfWeek": [
"*"
],
"hour": [
"*"
],
"minute": [
"*"
],
"month": [
"*"
]
},
"updated": "2019-08-24T14:15:22Z",
"updatedBy": {
"fullName": "string",
"userId": "string",
"username": "string"
}
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
batchPredictionJob | BatchPredictionJobDefinitionJobSpecResponse | true | The Batch Prediction Job specification to be put on the queue in intervals | |
created | string(date-time) | true | When was this job created | |
createdBy | BatchJobCreatedBy | true | Who created this job | |
enabled | boolean | true | If this job definition is enabled as a scheduled job. | |
id | string | true | The ID of the Batch job definition | |
lastFailedRunTime | string(date-time)¦null | false | Last time this job had a failed run | |
lastScheduledRunTime | string(date-time)¦null | false | Last time this job was scheduled to run (though not guaranteed it actually ran at that time) | |
lastStartedJobStatus | string¦null | true | The status of the latest job launched to the queue (if any). | |
lastStartedJobTime | string(date-time)¦null | true | The last time (if any) a job was launched. | |
lastSuccessfulRunTime | string(date-time)¦null | false | Last time this job had a successful run | |
name | string | true | A human-readable name for the definition, must be unique across organisations | |
nextScheduledRunTime | string(date-time)¦null | false | Next time this job is scheduled to run | |
schedule | Schedule | false | The scheduling information defining how often and when to execute this job to the Job Scheduling service. Optional if enabled = False. | |
updated | string(date-time) | true | When was this job last updated | |
updatedBy | BatchJobCreatedBy | true | Who created this job |
Enumerated Values¶
Property | Value |
---|---|
lastStartedJobStatus | [INITIALIZING , RUNNING , COMPLETED , ABORTED , FAILED ] |
BatchPredictionJobDefinitionsUpdate
{
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"enabled": true,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringBatchPrefix": "string",
"name": "string",
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionThreshold": 1,
"predictionWarningEnabled": true,
"schedule": {
"dayOfMonth": [
"*"
],
"dayOfWeek": [
"*"
],
"hour": [
"*"
],
"minute": [
"*"
],
"month": [
"*"
]
},
"secondaryDatasetsConfigId": "string",
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
abortOnError | boolean | false | Should this job abort if too many errors are encountered | |
batchJobType | string | false | Batch job type. | |
chunkSize | any | false | Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes. |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | integer | false | maximum: 41943040 minimum: 20 |
none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
columnNamesRemapping | any | false | Remap (rename or remove columns from) the output from this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | object | false | Provide a dictionary with key/value pairs to remap (deprecated) |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | [BatchPredictionJobRemapping] | false | maxItems: 1000 |
Provide a list of items to remap |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
csvSettings | BatchPredictionJobCSVSettings | false | The CSV settings used for this job | |
deploymentId | string | false | ID of deployment which is used in job for processing predictions dataset | |
disableRowLevelErrorHandling | boolean | false | Skip row by row error handling | |
enabled | boolean | false | If this job definition is enabled as a scheduled job. Optional if no schedule is supplied. | |
explanationAlgorithm | string | false | Which algorithm will be used to calculate prediction explanations | |
explanationClassNames | [string] | false | maxItems: 10 minItems: 1 |
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1 |
explanationNumTopClasses | integer | false | maximum: 10 minimum: 1 |
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1 |
includePredictionStatus | boolean | false | Include prediction status column in the output | |
includeProbabilities | boolean | false | Include probabilities for all classes | |
includeProbabilitiesClasses | [string] | false | maxItems: 100 |
Include only probabilities for these specific class names. |
intakeSettings | any | false | The intake option configured for this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | AzureIntake | false | Stream CSV data chunks from Azure |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BigQueryIntake | false | Stream CSV data chunks from Big Query using GCS |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | DataStageIntake | false | Stream CSV data chunks from data stage storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | Catalog | false | Stream CSV data chunks from AI catalog dataset |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | DSS | false | Stream CSV data chunks from DSS dataset |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | FileSystemIntake | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | GCPIntake | false | Stream CSV data chunks from Google Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | HTTPIntake | false | Stream CSV data chunks from HTTP |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | JDBCIntake | false | Stream CSV data chunks from JDBC |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | LocalFileIntake | false | Stream CSV data chunks from local file storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | S3Intake | false | Stream CSV data chunks from Amazon Cloud Storage S3 |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SnowflakeIntake | false | Stream CSV data chunks from Snowflake |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SynapseIntake | false | Stream CSV data chunks from Azure Synapse |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
maxExplanations | integer | false | maximum: 100 minimum: 0 |
Number of explanations requested. Will be ordered by strength. |
modelId | string | false | ID of leaderboard model which is used in job for processing predictions dataset | |
modelPackageId | string | false | ID of model package from registry is used in job for processing predictions dataset | |
monitoringBatchPrefix | string¦null | false | Name of the batch to create with this job | |
name | string | false | maxLength: 100 minLength: 1 minLength: 1 |
A human-readable name for the definition, must be unique across organisations, if left out the backend will generate one for you. |
numConcurrent | integer | false | minimum: 1 |
Number of simultaneous requests to run against the prediction instance |
outputSettings | any | false | The output option configured for this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | AzureOutput | false | Save CSV data chunks to Azure Blob Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BigQueryOutput | false | Save CSV data chunks to Google BigQuery in bulk |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | FileSystemOutput | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | GCPOutput | false | Save CSV data chunks to Google Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | HTTPOutput | false | Save CSV data chunks to HTTP data endpoint |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | JDBCOutput | false | Save CSV data chunks via JDBC |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | LocalFileOutput | false | Save CSV data chunks to local file storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | S3Output | false | Saves CSV data chunks to Amazon Cloud Storage S3 |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SnowflakeOutput | false | Save CSV data chunks to Snowflake in bulk |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SynapseOutput | false | Save CSV data chunks to Azure Synapse in bulk |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
passthroughColumns | [string] | false | maxItems: 100 |
Pass through columns from the original dataset |
passthroughColumnsSet | string | false | Pass through all columns from the original dataset | |
pinnedModelId | string | false | Specify a model ID used for scoring | |
predictionInstance | BatchPredictionJobPredictionInstance | false | Override the default prediction instance from the deployment when scoring this job. | |
predictionThreshold | number | false | maximum: 1 minimum: 0 |
Threshold is the point that sets the class boundary for a predicted value. The model classifies an observation below the threshold as FALSE, and an observation above the threshold as TRUE. In other words, DataRobot automatically assigns the positive class label to any prediction exceeding the threshold. This value can be set between 0.0 and 1.0. |
predictionWarningEnabled | boolean¦null | false | Enable prediction warnings. | |
schedule | Schedule | false | The scheduling information defining how often and when to execute this job to the Job Scheduling service. Optional if enabled = False. | |
secondaryDatasetsConfigId | string | false | Configuration id for secondary datasets to use when making a prediction. | |
skipDriftTracking | boolean | false | Skip drift tracking for this job. | |
thresholdHigh | number | false | Compute explanations for predictions above this threshold | |
thresholdLow | number | false | Compute explanations for predictions below this threshold | |
timeseriesSettings | any | false | Time Series settings included of this job is a Time Series job. |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchJobTimeSeriesSettingsForecast | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchPredictionJobTimeSeriesSettingsForecastWithPolicy | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchJobTimeSeriesSettingsHistorical | false | none |
Enumerated Values¶
Property | Value |
---|---|
batchJobType | [monitoring , prediction ] |
anonymous | [auto , fixed , dynamic ] |
explanationAlgorithm | [shap , xemp ] |
passthroughColumnsSet | all |
BatchPredictionJobId
{
"partNumber": 0,
"predictionJobId": "string"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
partNumber | integer | true | minimum: 0 |
The number of which csv part is being uploaded when using multipart upload |
predictionJobId | string | true | ID of the Batch Prediction job |
BatchPredictionJobLinks
{
"csvUpload": "string",
"download": "string",
"self": "string"
}
Links useful for this job
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
csvUpload | string(url) | false | The URL used to upload the dataset for this job. Only available for localFile intake. | |
download | string¦null | false | The URL used to download the results from this job. Only available for localFile outputs. Will be null if the download is not yet available. | |
self | string(url) | true | The URL used access this job. |
BatchPredictionJobListResponse
{
"count": 0,
"data": [
{
"batchPredictionJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"elapsedTimeSec": 0,
"failedRows": 0,
"hidden": "2019-08-24T14:15:22Z",
"id": "string",
"intakeDatasetDisplayName": "string",
"jobIntakeSize": 0,
"jobOutputSize": 0,
"jobSpec": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringBatchPrefix": "string",
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionThreshold": 1,
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"secondaryDatasetsConfigId": "string",
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"links": {
"csvUpload": "string",
"download": "string",
"self": "string"
},
"logs": [
"string"
],
"monitoringBatchId": "string",
"percentageCompleted": 100,
"queuePosition": 0,
"queued": true,
"resultsDeleted": true,
"scoredRows": 0,
"skippedRows": 0,
"source": "string",
"status": "INITIALIZING",
"statusDetails": "string"
}
],
"next": "http://example.com",
"previous": "http://example.com",
"totalCount": 0
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
count | integer | false | Number of items returned on this page. | |
data | [BatchPredictionJobResponse] | true | An array of jobs | |
next | string(uri)¦null | true | URL pointing to the next page (if null, there is no next page). | |
previous | string(uri)¦null | true | URL pointing to the previous page (if null, there is no previous page). | |
totalCount | integer | true | The total number of items across all pages. |
BatchPredictionJobPredictionInstance
{
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
}
Override the default prediction instance from the deployment when scoring this job.
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
apiKey | string | false | By default, prediction requests will use the API key of the user that created the job. This allows you to make requests on behalf of other users. | |
datarobotKey | string | false | If running a job against a prediction instance in the Managed AI Cloud, you must provide the organization level DataRobot-Key. | |
hostName | string | true | Override the default host name of the deployment with this. | |
sslEnabled | boolean | true | Use SSL (HTTPS) when communicating with the overriden prediction server. |
BatchPredictionJobRemapping
{
"inputName": "string",
"outputName": "string"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
inputName | string | true | Rename column with this name | |
outputName | string¦null | true | Rename column to this name (leave as null to remove from the output) |
BatchPredictionJobResponse
{
"batchPredictionJobDefinition": {
"createdBy": "string",
"id": "string",
"name": "string"
},
"created": "2019-08-24T14:15:22Z",
"createdBy": {
"fullName": "string",
"userId": "string",
"username": "string"
},
"elapsedTimeSec": 0,
"failedRows": 0,
"hidden": "2019-08-24T14:15:22Z",
"id": "string",
"intakeDatasetDisplayName": "string",
"jobIntakeSize": 0,
"jobOutputSize": 0,
"jobSpec": {
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringBatchPrefix": "string",
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionThreshold": 1,
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"secondaryDatasetsConfigId": "string",
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
},
"links": {
"csvUpload": "string",
"download": "string",
"self": "string"
},
"logs": [
"string"
],
"monitoringBatchId": "string",
"percentageCompleted": 100,
"queuePosition": 0,
"queued": true,
"resultsDeleted": true,
"scoredRows": 0,
"skippedRows": 0,
"source": "string",
"status": "INITIALIZING",
"statusDetails": "string"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
batchPredictionJobDefinition | BatchPredictionJobDefinitionResponse | false | The Batch Prediction Job Definition linking to this job, if any. | |
created | string(date-time) | true | When was this job created | |
createdBy | BatchPredictionCreatedBy | true | Who created this job | |
elapsedTimeSec | integer | true | minimum: 0 |
Number of seconds the job has been processing for |
failedRows | integer | true | minimum: 0 |
Number of rows that have failed scoring |
hidden | string(date-time) | false | When was this job was hidden last, blank if visible | |
id | string | true | The ID of the Batch Prediction job | |
intakeDatasetDisplayName | string¦null | false | If applicable (e.g. for AI catalog), will contain the dataset name used for the intake dataset. | |
jobIntakeSize | integer¦null | true | minimum: 0 |
Number of bytes in the intake dataset for this job |
jobOutputSize | integer¦null | true | minimum: 0 |
Number of bytes in the output dataset for this job |
jobSpec | BatchPredictionJobSpecResponse | true | The job configuration used to create this job | |
links | BatchPredictionJobLinks | true | Links useful for this job | |
logs | [string] | true | The job log. | |
monitoringBatchId | string¦null | true | Id of the monitoring batch created by this job. Only present if the job runs on a deployment with batch monitoring enabled. | |
percentageCompleted | number | true | maximum: 100 minimum: 0 |
Indicates job progress which is based on number of already processed rows in dataset |
queuePosition | integer¦null | false | minimum: 0 |
To ensure a dedicated prediction instance is not overloaded, only one job will be run against it at a time. This is the number of jobs that are awaiting processing before this job start running. May not be available in all environments. |
queued | boolean | true | The job has been put on the queue for execution. | |
resultsDeleted | boolean | false | Indicates if the job was subject to garbage collection and had its artifacts deleted (output files, if any, and scoring data on local storage) | |
scoredRows | integer | true | minimum: 0 |
Number of rows that have been used in prediction computation |
skippedRows | integer | true | minimum: 0 |
Number of rows that have been skipped during scoring. May contain non-zero value only in time-series predictions case if provided dataset contains more than required historical rows. |
source | string | false | Source from which batch job was started | |
status | string | true | The current job status | |
statusDetails | string | true | Explanation for current status |
Enumerated Values¶
Property | Value |
---|---|
status | [INITIALIZING , RUNNING , COMPLETED , ABORTED , FAILED ] |
BatchPredictionJobSpecResponse
{
"abortOnError": true,
"batchJobType": "monitoring",
"chunkSize": "auto",
"columnNamesRemapping": {},
"csvSettings": {
"delimiter": ",",
"encoding": "utf-8",
"quotechar": "\""
},
"deploymentId": "string",
"disableRowLevelErrorHandling": false,
"explanationAlgorithm": "shap",
"explanationClassNames": [
"string"
],
"explanationNumTopClasses": 1,
"includePredictionStatus": false,
"includeProbabilities": true,
"includeProbabilitiesClasses": [],
"intakeSettings": {
"type": "localFile"
},
"maxExplanations": 0,
"modelId": "string",
"modelPackageId": "string",
"monitoringBatchPrefix": "string",
"numConcurrent": 1,
"outputSettings": {
"credentialId": "string",
"format": "csv",
"partitionColumns": [
"string"
],
"type": "azure",
"url": "string"
},
"passthroughColumns": [
"string"
],
"passthroughColumnsSet": "all",
"pinnedModelId": "string",
"predictionInstance": {
"apiKey": "string",
"datarobotKey": "string",
"hostName": "string",
"sslEnabled": true
},
"predictionThreshold": 1,
"predictionWarningEnabled": true,
"redactedFields": [
"string"
],
"secondaryDatasetsConfigId": "string",
"skipDriftTracking": false,
"thresholdHigh": 0,
"thresholdLow": 0,
"timeseriesSettings": {
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
}
The job configuration used to create this job
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
abortOnError | boolean | true | Should this job abort if too many errors are encountered | |
batchJobType | string | false | Batch job type. | |
chunkSize | any | false | Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes. |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | integer | false | maximum: 41943040 minimum: 20 |
none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
columnNamesRemapping | any | false | Remap (rename or remove columns from) the output from this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | object | false | Provide a dictionary with key/value pairs to remap (deprecated) |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | [BatchPredictionJobRemapping] | false | maxItems: 1000 |
Provide a list of items to remap |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
csvSettings | BatchPredictionJobCSVSettings | true | The CSV settings used for this job | |
deploymentId | string | false | ID of deployment which is used in job for processing predictions dataset | |
disableRowLevelErrorHandling | boolean | true | Skip row by row error handling | |
explanationAlgorithm | string | false | Which algorithm will be used to calculate prediction explanations | |
explanationClassNames | [string] | false | maxItems: 10 minItems: 1 |
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1 |
explanationNumTopClasses | integer | false | maximum: 10 minimum: 1 |
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1 |
includePredictionStatus | boolean | true | Include prediction status column in the output | |
includeProbabilities | boolean | true | Include probabilities for all classes | |
includeProbabilitiesClasses | [string] | true | maxItems: 100 |
Include only probabilities for these specific class names. |
intakeSettings | any | true | The response option configured for this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | AzureDataStreamer | false | Stream CSV data chunks from Azure |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | DataStageDataStreamer | false | Stream CSV data chunks from data stage storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | CatalogDataStreamer | false | Stream CSV data chunks from AI catalog dataset |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | GCPDataStreamer | false | Stream CSV data chunks from Google Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BigQueryDataStreamer | false | Stream CSV data chunks from Big Query using GCS |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | S3DataStreamer | false | Stream CSV data chunks from Amazon Cloud Storage S3 |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SnowflakeDataStreamer | false | Stream CSV data chunks from Snowflake |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SynapseDataStreamer | false | Stream CSV data chunks from Azure Synapse |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | DSSDataStreamer | false | Stream CSV data chunks from DSS dataset |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | FileSystemDataStreamer | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | HTTPDataStreamer | false | Stream CSV data chunks from HTTP |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | JDBCDataStreamer | false | Stream CSV data chunks from JDBC |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | LocalFileDataStreamer | false | Stream CSV data chunks from local file storage |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
maxExplanations | integer | true | maximum: 100 minimum: 0 |
Number of explanations requested. Will be ordered by strength. |
modelId | string | false | ID of leaderboard model which is used in job for processing predictions dataset | |
modelPackageId | string | false | ID of model package from registry is used in job for processing predictions dataset | |
monitoringBatchPrefix | string¦null | false | Name of the batch to create with this job | |
numConcurrent | integer | false | minimum: 1 |
Number of simultaneous requests to run against the prediction instance |
outputSettings | any | false | The response option configured for this job |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | AzureOutputAdaptor | false | Save CSV data chunks to Azure Blob Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | GCPOutputAdaptor | false | Save CSV data chunks to Google Storage |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BigQueryOutputAdaptor | false | Save CSV data chunks to Google BigQuery in bulk |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | S3OutputAdaptor | false | Saves CSV data chunks to Amazon Cloud Storage S3 |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SnowflakeOutputAdaptor | false | Save CSV data chunks to Snowflake in bulk |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | SynapseOutputAdaptor | false | Save CSV data chunks to Azure Synapse in bulk |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | FileSystemOutputAdaptor | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | HttpOutputAdaptor | false | Save CSV data chunks to HTTP data endpoint |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | JdbcOutputAdaptor | false | Save CSV data chunks via JDBC |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | LocalFileOutputAdaptor | false | Save CSV data chunks to local file storage |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
passthroughColumns | [string] | false | maxItems: 100 |
Pass through columns from the original dataset |
passthroughColumnsSet | string | false | Pass through all columns from the original dataset | |
pinnedModelId | string | false | Specify a model ID used for scoring | |
predictionInstance | BatchPredictionJobPredictionInstance | false | Override the default prediction instance from the deployment when scoring this job. | |
predictionThreshold | number | false | maximum: 1 minimum: 0 |
Threshold is the point that sets the class boundary for a predicted value. The model classifies an observation below the threshold as FALSE, and an observation above the threshold as TRUE. In other words, DataRobot automatically assigns the positive class label to any prediction exceeding the threshold. This value can be set between 0.0 and 1.0. |
predictionWarningEnabled | boolean¦null | false | Enable prediction warnings. | |
redactedFields | [string] | true | A list of qualified field names from intake- and/or outputSettings that was redacted due to permissions and sharing settings. For example: intakeSettings.dataStoreId | |
secondaryDatasetsConfigId | string | false | Configuration id for secondary datasets to use when making a prediction. | |
skipDriftTracking | boolean | true | Skip drift tracking for this job. | |
thresholdHigh | number | false | Compute explanations for predictions above this threshold | |
thresholdLow | number | false | Compute explanations for predictions below this threshold | |
timeseriesSettings | any | false | Time Series settings included of this job is a Time Series job. |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchPredictionJobTimeSeriesSettingsForecast | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchPredictionJobTimeSeriesSettingsHistorical | false | none |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | BatchPredictionJobTimeSeriesSettingsTraining | false | none |
Enumerated Values¶
Property | Value |
---|---|
batchJobType | [monitoring , prediction ] |
anonymous | [auto , fixed , dynamic ] |
explanationAlgorithm | [shap , xemp ] |
passthroughColumnsSet | all |
BatchPredictionJobTimeSeriesSettingsForecast
{
"forecastPoint": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
forecastPoint | string(date-time) | false | Used for forecast predictions in order to override the inferred forecast point from the dataset. | |
relaxKnownInAdvanceFeaturesCheck | boolean | false | If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed. | |
type | string | true | Forecast mode makes predictions using forecastPoint or rows in the dataset without target. |
Enumerated Values¶
Property | Value |
---|---|
type | forecast |
BatchPredictionJobTimeSeriesSettingsForecastWithPolicy
{
"forecastPointPolicy": {
"configuration": {
"offset": "string"
},
"type": "jobRunTimeBased"
},
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "forecast"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
forecastPointPolicy | JobRunTimeBasedForecastPointPolicy | true | Forecast point policy | |
relaxKnownInAdvanceFeaturesCheck | boolean | false | If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed. | |
type | string | true | Forecast mode makes predictions using forecastPoint or rows in the dataset without target. |
Enumerated Values¶
Property | Value |
---|---|
type | forecast |
BatchPredictionJobTimeSeriesSettingsHistorical
{
"predictionsEndDate": "2019-08-24T14:15:22Z",
"predictionsStartDate": "2019-08-24T14:15:22Z",
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "historical"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
predictionsEndDate | string(date-time) | false | Used for historical predictions in order to override date to which predictions should be calculated. By default value will be inferred automatically from the dataset. | |
predictionsStartDate | string(date-time) | false | Used for historical predictions in order to override date from which predictions should be calculated. By default value will be inferred automatically from the dataset. | |
relaxKnownInAdvanceFeaturesCheck | boolean | false | If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed. | |
type | string | true | Historical mode enables bulk predictions which calculates predictions for all possible forecast points and forecast distances in the dataset within the predictionsStartDate/predictionsEndDate range. |
Enumerated Values¶
Property | Value |
---|---|
type | historical |
BatchPredictionJobTimeSeriesSettingsTraining
{
"relaxKnownInAdvanceFeaturesCheck": false,
"type": "training"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
relaxKnownInAdvanceFeaturesCheck | boolean | false | If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed. | |
type | string | true | Forecast mode used for making predictions on subsets of training data. |
Enumerated Values¶
Property | Value |
---|---|
type | training |
BatchPredictionJobUpdate
{
"aborted": "2019-08-24T14:15:22Z",
"completed": "2019-08-24T14:15:22Z",
"failedRows": 0,
"hidden": true,
"jobIntakeSize": 0,
"jobOutputSize": 0,
"logs": [
"string"
],
"scoredRows": 0,
"skippedRows": 0,
"started": "2019-08-24T14:15:22Z",
"status": "INITIALIZING"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
aborted | string(date-time)¦null | false | Time when job abortion happened | |
completed | string(date-time)¦null | false | Time when job completed scoring | |
failedRows | integer | false | Number of rows that have failed scoring | |
hidden | boolean | false | Hides or unhides the job from the job list | |
jobIntakeSize | integer¦null | false | Number of bytes in the intake dataset for this job | |
jobOutputSize | integer¦null | false | Number of bytes in the output dataset for this job | |
logs | [string] | false | The job log. | |
scoredRows | integer | false | Number of rows that have been used in prediction computation | |
skippedRows | integer | false | Number of rows that have been skipped during scoring. May contain non-zero value only in time-series predictions case if provided dataset contains more than required historical rows. | |
started | string(date-time)¦null | false | Time when job scoring begin | |
status | string | false | The current job status |
Enumerated Values¶
Property | Value |
---|---|
status | [INITIALIZING , RUNNING , COMPLETED , ABORTED , FAILED ] |
BigQueryDataStreamer
{
"bucket": "string",
"credentialId": "string",
"dataset": "string",
"table": "string",
"type": "bigquery"
}
Stream CSV data chunks from Big Query using GCS
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
bucket | string | true | The name of gcs bucket for data export | |
credentialId | any | true | Either the populated value of the field or [redacted] due to permission settings |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | The ID of the GCP credentials |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
dataset | string | true | The name of the specified big query dataset to read input data from | |
table | string | true | The name of the specified big query table to read input data from | |
type | string | true | Type name for this intake type |
Enumerated Values¶
Property | Value |
---|---|
anonymous | [redacted] |
type | bigquery |
BigQueryIntake
{
"bucket": "string",
"credentialId": "string",
"dataset": "string",
"table": "string",
"type": "bigquery"
}
Stream CSV data chunks from Big Query using GCS
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
bucket | string | true | The name of gcs bucket for data export | |
credentialId | string | true | The ID of the GCP credentials | |
dataset | string | true | The name of the specified big query dataset to read input data from | |
table | string | true | The name of the specified big query table to read input data from | |
type | string | true | Type name for this intake type |
Enumerated Values¶
Property | Value |
---|---|
type | bigquery |
BigQueryOutput
{
"bucket": "string",
"credentialId": "string",
"dataset": "string",
"table": "string",
"type": "bigquery"
}
Save CSV data chunks to Google BigQuery in bulk
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
bucket | string | true | The name of gcs bucket for data loading | |
credentialId | string | true | The ID of the GCP credentials | |
dataset | string | true | The name of the specified big query dataset to write data back | |
table | string | true | The name of the specified big query table to write data back | |
type | string | true | Type name for this output type |
Enumerated Values¶
Property | Value |
---|---|
type | bigquery |
BigQueryOutputAdaptor
{
"bucket": "string",
"credentialId": "string",
"dataset": "string",
"table": "string",
"type": "bigquery"
}
Save CSV data chunks to Google BigQuery in bulk
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
bucket | string | true | The name of gcs bucket for data loading | |
credentialId | any | true | Either the populated value of the field or [redacted] due to permission settings |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | The ID of the GCP credentials |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
dataset | string | true | The name of the specified big query dataset to write data back | |
table | string | true | The name of the specified big query table to write data back | |
type | string | true | Type name for this output type |
Enumerated Values¶
Property | Value |
---|---|
anonymous | [redacted] |
type | bigquery |
Catalog
{
"datasetId": "string",
"datasetVersionId": "string",
"type": "dataset"
}
Stream CSV data chunks from AI catalog dataset
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
datasetId | string | true | The ID of the AI catalog dataset | |
datasetVersionId | string | false | The ID of the AI catalog dataset version | |
type | string | true | Type name for this intake type |
Enumerated Values¶
Property | Value |
---|---|
type | dataset |
CatalogDataStreamer
{
"datasetId": "string",
"datasetVersionId": "string",
"type": "dataset"
}
Stream CSV data chunks from AI catalog dataset
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
datasetId | any | true | Either the populated value of the field or [redacted] due to permission settings |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | The ID of the AI catalog dataset |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
datasetVersionId | string | false | The ID of the AI catalog dataset version | |
type | string | true | Type name for this intake type |
Enumerated Values¶
Property | Value |
---|---|
anonymous | [redacted] |
type | dataset |
CreatePredictionDatasetResponse
{
"datasetId": "string"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
datasetId | string | true | The ID of the newly created prediction dataset. |
CreatePredictionFromDataset
{
"actualValueColumn": "string",
"datasetId": "string",
"explanationAlgorithm": "shap",
"forecastPoint": "2019-08-24T14:15:22Z",
"includeFdwCounts": false,
"includePredictionIntervals": true,
"maxExplanations": 1,
"modelId": "string",
"predictionIntervalsSize": 1,
"predictionThreshold": 1,
"predictionsEndDate": "2019-08-24T14:15:22Z",
"predictionsStartDate": "2019-08-24T14:15:22Z"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
actualValueColumn | string | false | For time series projects only. Actual value column name, valid for the prediction files if the project is unsupervised and the dataset is considered as bulk predictions dataset. This value is optional. | |
datasetId | string | true | The dataset to compute predictions for - must have previously been uploaded. | |
explanationAlgorithm | string | false | If set to shap , the response will include prediction explanations based on the SHAP explainer (SHapley Additive exPlanations). Defaults to null (no prediction explanations). |
|
forecastPoint | string(date-time) | false | For time series projects only. The time in the dataset relative to which predictions are generated. This value is optional. If not specified the default value is the value in the row with the latest specified timestamp. Specifying this value for a project that is not a time series project will result in an error. | |
includeFdwCounts | boolean | false | For time series projects with partial history only. Indicates if feature derivation window counts featureDerivationWindowCounts will be part of the response. |
|
includePredictionIntervals | boolean | false | Specifies whether prediction intervals should be calculated for this request. Defaults to True if predictionIntervalsSize is specified, otherwise defaults to False. |
|
maxExplanations | integer | false | maximum: 100 minimum: 1 |
Specifies the maximum number of explanation values that should be returned for each row, ordered by absolute value, greatest to least. In the case of 'shap': If not set, explanations are returned for all features. If the number of features is greater than the 'maxExplanations', the sum of remaining values will also be returned as 'shapRemainingTotal'. Defaults to null for datasets narrower than 100 columns, defaults to 100 for datasets wider than 100 columns. Cannot be set if 'explanationAlgorithm' is omitted. |
modelId | string | true | The model to make predictions on. | |
predictionIntervalsSize | integer | false | maximum: 100 minimum: 1 |
Represents the percentile to use for the size of the prediction intervals. Defaults to 80 if includePredictionIntervals is True. |
predictionThreshold | number | false | maximum: 1 minimum: 0 |
Threshold used for binary classification in predictions. Accepts values from 0.0 to 1.0. If not specified, model default prediction threshold will be used. |
predictionsEndDate | string(date-time) | false | The end date for bulk predictions, exclusive. Used for time series projects only. Note that this parameter is used for generating historical predictions using the training data, not for future predictions. If not specified, the dataset is not considered as a bulk predictions dataset. This parameter should be provided in conjunction with a predictionsStartDate , and cannot be provided with the forecastPoint parameter. |
|
predictionsStartDate | string(date-time) | false | The start date for bulk predictions. Used for time series projects only. Note that this parameter is used for generating historical predictions using the training data, not for future predictions. If not specified, the dataset is not considered as a bulk predictions dataset. This parameter should be provided in conjunction with a predictionsEndDate , and cannot be provided with the forecastPoint parameter. |
Enumerated Values¶
Property | Value |
---|---|
explanationAlgorithm | shap |
CreateTrainingPrediction
{
"dataSubset": "all",
"explanationAlgorithm": "string",
"maxExplanations": 1,
"modelId": "string"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
dataSubset | string | true | Subset of data predicted on: The value "all" returns predictions for all rows in the dataset including data used for training, validation, holdout and any rows discarded. This is not available for large datasets or projects created with Date/Time partitioning. The value "validationAndHoldout" returns predictions for the rows used to calculate the validation score and the holdout score. Not available for large projects or Date/Time projects for models trained into the validation set. The value "holdout" returns predictions for the rows used to calculate the holdout score. Not available for projects created without a holdout or for models trained into holdout for large datasets or created with Date/Time partitioning. The value "allBacktests" returns predictions for the rows used to calculate the backtesting scores for Date/Time projects. The value "validation" returns predictions for the rows used to calculate the validation score. | |
explanationAlgorithm | string | false | If set to "shap", the response will include prediction explanations based on the SHAP explainer (SHapley Additive exPlanations). Defaults to null (no prediction explanations) | |
maxExplanations | integer | false | maximum: 100 minimum: 1 |
Specifies the maximum number of explanation values that should be returned for each row, ordered by absolute value, greatest to least. In the case of "shap": If not set, explanations are returned for all features. If the number of features is greater than the "maxExplanations", the sum of remaining values will also be returned as "shapRemainingTotal". Defaults to null for datasets narrower than 100 columns, defaults to 100 for datasets wider than 100 columns. Cannot be set if "explanationAlgorithm" is omitted. |
modelId | string | true | The model to make predictions on |
Enumerated Values¶
Property | Value |
---|---|
dataSubset | [all , validationAndHoldout , holdout , allBacktests , validation , crossValidation ] |
CredentialId
{
"catalogVersionId": "string",
"credentialId": "string",
"url": "string"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
catalogVersionId | string | false | The ID of the latest version of the catalog entry. | |
credentialId | string | true | The ID of the set of credentials to use instead of user and password. Note that with this change, username and password will become optional. | |
url | string | false | The link to retrieve more detailed information about the entity that uses this catalog dataset. |
DSS
{
"datasetId": "string",
"partition": "holdout",
"projectId": "string",
"type": "dss"
}
Stream CSV data chunks from DSS dataset
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
datasetId | string | false | The ID of the dataset | |
partition | string | false | Partition used to predict | |
projectId | string | true | The ID of the project | |
type | string | true | Type name for this intake type |
Enumerated Values¶
Property | Value |
---|---|
partition | [holdout , validation , allBacktests , None ] |
type | dss |
DSSDataStreamer
{
"datasetId": "string",
"partition": "holdout",
"projectId": "string",
"type": "dss"
}
Stream CSV data chunks from DSS dataset
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
datasetId | any | false | Either the populated value of the field or [redacted] due to permission settings |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | The ID of the dataset |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
partition | string | false | Partition used to predict | |
projectId | string | true | The ID of the project | |
type | string | true | Type name for this intake type |
Enumerated Values¶
Property | Value |
---|---|
anonymous | [redacted] |
partition | [holdout , validation , allBacktests , None ] |
type | dss |
DataQualityWarningsRecord
{
"hasKiaMissingValuesInForecastWindow": true,
"insufficientRowsForEvaluatingModels": true,
"singleClassActualValueColumn": true
}
A Json object of available warnings about potential problems in this prediction dataset. Empty if no warnings.
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
hasKiaMissingValuesInForecastWindow | boolean | false | If true, known-in-advance features in this dataset have missing values in the forecast window. Absence of the known-in-advance values can negatively impact prediction quality. Only applies for time series projects. | |
insufficientRowsForEvaluatingModels | boolean | false | If true, the dataset has a target column present indicating it can be used to evaluate model performance but too few rows to be trustworthy in so doing. If false, either it has no target column at all or it has sufficient rows for model evaluation. Only applies for regression, binary classification, multiclass classification projects and time series unsupervised projects. | |
singleClassActualValueColumn | boolean | false | If true, actual value column has only one class and such insights as ROC curve can not be calculated. Only applies for binary classification projects or unsupervised projects. |
DataStageDataStreamer
{
"dataStageId": "string",
"type": "dataStage"
}
Stream CSV data chunks from data stage storage
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
dataStageId | string | true | The ID of the data stage | |
type | string | true | Type name for this intake type |
Enumerated Values¶
Property | Value |
---|---|
type | dataStage |
DataStageIntake
{
"dataStageId": "string",
"type": "dataStage"
}
Stream CSV data chunks from data stage storage
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
dataStageId | string | true | The ID of the data stage | |
type | string | true | Type name for this intake type |
Enumerated Values¶
Property | Value |
---|---|
type | dataStage |
DatabricksAccessTokenCredentials
{
"credentialType": "databricks_access_token_account",
"databricksAccessToken": "string"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
credentialType | string | true | The type of these credentials, 'databricks_access_token_account' here. | |
databricksAccessToken | string | true | minLength: 1 minLength: 1 |
Databricks personal access token. |
Enumerated Values¶
Property | Value |
---|---|
credentialType | databricks_access_token_account |
FileSystemDataStreamer
{
"path": "string",
"type": "filesystem"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
path | string | true | Path to data on host filesystem | |
type | string | true | Type name for this intake type |
Enumerated Values¶
Property | Value |
---|---|
type | filesystem |
FileSystemIntake
{
"path": "string",
"type": "filesystem"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
path | string | true | Path to data on host filesystem | |
type | string | true | Type name for this intake type |
Enumerated Values¶
Property | Value |
---|---|
type | filesystem |
FileSystemOutput
{
"path": "string",
"type": "filesystem"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
path | string | true | Path to results on host filesystem | |
type | string | true | Type name for this output type |
Enumerated Values¶
Property | Value |
---|---|
type | filesystem |
FileSystemOutputAdaptor
{
"path": "string",
"type": "filesystem"
}
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
path | string | true | Path to results on host filesystem | |
type | string | true | Type name for this output type |
Enumerated Values¶
Property | Value |
---|---|
type | filesystem |
GCPDataStreamer
{
"credentialId": "string",
"format": "csv",
"type": "gcp",
"url": "string"
}
Stream CSV data chunks from Google Storage
Properties¶
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
credentialId | any | false | Either the populated value of the field or [redacted] due to permission settings |
oneOf
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string¦null | false | Use the specified credential to access the url |
xor
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
» anonymous | string | false | none |
continued
Name | Type | Required | Restrictions | Description |
---|---|---|---|---|
format | string | false | Type of input file format | |
type | string | true | Type name for this intake type | |
url | string(url) | true | URL for the CSV file |
Enumerated Values¶
Property | Value |
---|---|
anonymous | [redacted] |
format | [csv , parquet ] |
type | gcp |