Skip to content

On-premise users: click in-app to access the full platform documentation for your version of DataRobot.

ML Ops

This page outlines the operations, endpoints, parameters, and example requests and responses for the ML Ops.

GET /api/v2/batchJobs/

Get a collection of batch jobs by statuses

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchJobs/?offset=0&limit=100 \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer true This many results will be skipped
limit query integer true At most this many results are returned
status query any false Includes only jobs that have the status value that matches this flag. Repeat the parameter for filtering on multiple statuses.
source query any false Includes only jobs that have the source value that matches this flag. Repeat the parameter for filtering on multiple statuses.Prefix values with a dash (-) to exclude those sources.
deploymentId query string false Includes only jobs for this particular deployment
modelId query string false ID of leaderboard model which is used in job for processing predictions dataset
jobId query string false Includes only job by specific id
orderBy query string false Sort order which will be applied to batch prediction list. Prefix the attribute name with a dash to sort in descending order, e.g. "-created".
allJobs query boolean false [DEPRECATED - replaced with RBAC permission model] - No effect
cutoffHours query integer false Only list jobs created at most this amount of hours ago.
startDateTime query string(date-time) false ISO-formatted datetime of the earliest time the job was added (inclusive). For example "2008-08-24T12:00:00Z". Will ignore cutoffHours if set.
endDateTime query string(date-time) false ISO-formatted datetime of the latest time the job was added (inclusive). For example "2008-08-24T12:00:00Z".
batchPredictionJobDefinitionId query string false Includes only jobs for this particular definition
hostname query any false Includes only jobs for this particular prediction instance hostname
batchJobType query any false Includes only jobs that have the batch job type that matches this flag. Repeat the parameter for filtering on multiple types.
intakeType query any false Includes only jobs for these particular intakes type
outputType query any false Includes only jobs for these particular outputs type

Enumerated Values

Parameter Value
orderBy [created, -created, status, -status]

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "batchMonitoringJobDefinition": {
        "createdBy": "string",
        "id": "string",
        "name": "string"
      },
      "batchPredictionJobDefinition": {
        "createdBy": "string",
        "id": "string",
        "name": "string"
      },
      "created": "2019-08-24T14:15:22Z",
      "createdBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      },
      "elapsedTimeSec": 0,
      "failedRows": 0,
      "hidden": "2019-08-24T14:15:22Z",
      "id": "string",
      "intakeDatasetDisplayName": "string",
      "jobIntakeSize": 0,
      "jobOutputSize": 0,
      "jobSpec": {
        "abortOnError": true,
        "batchJobType": "monitoring",
        "chunkSize": "auto",
        "columnNamesRemapping": {},
        "csvSettings": {
          "delimiter": ",",
          "encoding": "utf-8",
          "quotechar": "\""
        },
        "deploymentId": "string",
        "disableRowLevelErrorHandling": false,
        "explanationAlgorithm": "shap",
        "explanationClassNames": [
          "string"
        ],
        "explanationNumTopClasses": 1,
        "includePredictionStatus": false,
        "includeProbabilities": true,
        "includeProbabilitiesClasses": [],
        "intakeSettings": {
          "type": "localFile"
        },
        "maxExplanations": 0,
        "maxNgramExplanations": 0,
        "modelId": "string",
        "modelPackageId": "string",
        "monitoringAggregation": {
          "retentionPolicy": "samples",
          "retentionValue": 0
        },
        "monitoringBatchPrefix": "string",
        "monitoringColumns": {
          "actedUponColumn": "string",
          "actualsTimestampColumn": "string",
          "actualsValueColumn": "string",
          "associationIdColumn": "string",
          "customMetricId": "string",
          "customMetricTimestampColumn": "string",
          "customMetricTimestampFormat": "string",
          "customMetricValueColumn": "string",
          "monitoredStatusColumn": "string",
          "predictionsColumns": [
            {
              "className": "string",
              "columnName": "string"
            }
          ],
          "reportDrift": true,
          "reportPredictions": true,
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "monitoringOutputSettings": {
          "monitoredStatusColumn": "string",
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "numConcurrent": 1,
        "outputSettings": {
          "credentialId": "string",
          "format": "csv",
          "partitionColumns": [
            "string"
          ],
          "type": "azure",
          "url": "string"
        },
        "passthroughColumns": [
          "string"
        ],
        "passthroughColumnsSet": "all",
        "pinnedModelId": "string",
        "predictionInstance": {
          "apiKey": "string",
          "datarobotKey": "string",
          "hostName": "string",
          "sslEnabled": true
        },
        "predictionWarningEnabled": true,
        "redactedFields": [
          "string"
        ],
        "skipDriftTracking": false,
        "thresholdHigh": 0,
        "thresholdLow": 0,
        "timeseriesSettings": {
          "forecastPoint": "2019-08-24T14:15:22Z",
          "relaxKnownInAdvanceFeaturesCheck": false,
          "type": "forecast"
        }
      },
      "links": {
        "csvUpload": "string",
        "download": "string",
        "self": "string"
      },
      "logs": [
        "string"
      ],
      "monitoringBatchId": "string",
      "percentageCompleted": 100,
      "queuePosition": 0,
      "queued": true,
      "resultsDeleted": true,
      "scoredRows": 0,
      "skippedRows": 0,
      "source": "string",
      "status": "INITIALIZING",
      "statusDetails": "string"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Responses

Status Meaning Description Schema
200 OK A list of Batch job objects BatchJobListResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/batchJobs/fromJobDefinition/

Launches a one-time batch job based off of the previously supplied definition referring to the job definition ID and puts it on the queue.

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/batchJobs/fromJobDefinition/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{BatchPredictionJobDefinitionId}'

Body parameter

{
  "jobDefinitionId": "string"
}

Parameters

Name In Type Required Description
body body BatchPredictionJobDefinitionId false none

Example responses

202 Response

{
  "batchMonitoringJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "batchPredictionJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "elapsedTimeSec": 0,
  "failedRows": 0,
  "hidden": "2019-08-24T14:15:22Z",
  "id": "string",
  "intakeDatasetDisplayName": "string",
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "jobSpec": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 1,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "links": {
    "csvUpload": "string",
    "download": "string",
    "self": "string"
  },
  "logs": [
    "string"
  ],
  "monitoringBatchId": "string",
  "percentageCompleted": 100,
  "queuePosition": 0,
  "queued": true,
  "resultsDeleted": true,
  "scoredRows": 0,
  "skippedRows": 0,
  "source": "string",
  "status": "INITIALIZING",
  "statusDetails": "string"
}

Responses

Status Meaning Description Schema
202 Accepted Job details for the created Batch Prediction job BatchJobResponse
404 Not Found Job was deleted, never existed or you do not have access to it None
422 Unprocessable Entity Could not create a Batch job. Possible reasons: {} None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

DELETE /api/v2/batchJobs/{batchJobId}/

If the job is running, it will be aborted. Then it will be removed, meaning all underlying data will be deleted and the job is removed from the list of jobs.

Code samples

# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/batchJobs/{batchJobId}/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
batchJobId path string true ID of the Batch job
partNumber path integer true The number of which csv part is being uploaded when using multipart upload

Responses

Status Meaning Description Schema
202 Accepted Job cancelled None
404 Not Found Job does not exist or was not submitted to the queue. None
409 Conflict Job cannot be aborted for some reason. Possible reasons: job is already aborted or completed. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/batchJobs/{batchJobId}/

Retrieve a Batch job.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchJobs/{batchJobId}/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
batchJobId path string true ID of the Batch job
partNumber path integer true The number of which csv part is being uploaded when using multipart upload

Example responses

200 Response

{
  "batchMonitoringJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "batchPredictionJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "elapsedTimeSec": 0,
  "failedRows": 0,
  "hidden": "2019-08-24T14:15:22Z",
  "id": "string",
  "intakeDatasetDisplayName": "string",
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "jobSpec": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 1,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "links": {
    "csvUpload": "string",
    "download": "string",
    "self": "string"
  },
  "logs": [
    "string"
  ],
  "monitoringBatchId": "string",
  "percentageCompleted": 100,
  "queuePosition": 0,
  "queued": true,
  "resultsDeleted": true,
  "scoredRows": 0,
  "skippedRows": 0,
  "source": "string",
  "status": "INITIALIZING",
  "statusDetails": "string"
}

Responses

Status Meaning Description Schema
200 OK Job details for the requested Batch job BatchJobResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

PUT /api/v2/batchJobs/{batchJobId}/csvUpload/

Stream CSV data to the job. Only available for jobs thatuses the localFile intake option.

Code samples

# You can also use wget
curl -X PUT https://app.datarobot.com/api/v2/batchJobs/{batchJobId}/csvUpload/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
batchJobId path string true ID of the Batch job
partNumber path integer true The number of which csv part is being uploaded when using multipart upload

Responses

Status Meaning Description Schema
202 Accepted Job data was successfully submitted None
404 Not Found Job does not exist or does not require data None
409 Conflict Dataset upload has already begun None
415 Unsupported Media Type Not acceptable MIME type None
422 Unprocessable Entity Job was "ABORTED" due to too many errors in the data None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/batchJobs/{batchJobId}/download/

This is only valid for jobs scored using the "localFile" output option

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchJobs/{batchJobId}/download/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
batchJobId path string true ID of the Batch job
partNumber path integer true The number of which csv part is being uploaded when using multipart upload

Responses

Status Meaning Description Schema
200 OK Job was downloaded correctly None
404 Not Found Job does not exist or is not completed None
406 Not Acceptable Not acceptable MIME type None
422 Unprocessable Entity Job was "ABORTED" due to too many errors in the data None

Response Headers

Status Header Type Format Description
200 Content-Disposition string Contains an auto generated filename for this download ("attachment;filename=result-.csv").
200 Content-Type string MIME type of the returned data

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/batchMonitoring/

Submit the configuration for the job and it will be submitted to the queue

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/batchMonitoring/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{BatchMonitoringJobCreate}'

Body parameter

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Parameters

Name In Type Required Description
body body BatchMonitoringJobCreate false none

Example responses

202 Response

{
  "batchMonitoringJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "batchPredictionJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "elapsedTimeSec": 0,
  "failedRows": 0,
  "hidden": "2019-08-24T14:15:22Z",
  "id": "string",
  "intakeDatasetDisplayName": "string",
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "jobSpec": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 1,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "links": {
    "csvUpload": "string",
    "download": "string",
    "self": "string"
  },
  "logs": [
    "string"
  ],
  "monitoringBatchId": "string",
  "percentageCompleted": 100,
  "queuePosition": 0,
  "queued": true,
  "resultsDeleted": true,
  "scoredRows": 0,
  "skippedRows": 0,
  "source": "string",
  "status": "INITIALIZING",
  "statusDetails": "string"
}

Responses

Status Meaning Description Schema
202 Accepted Job details for the created Batch Monitoring job BatchJobResponse

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/batchMonitoringJobDefinitions/

List all Batch Monitoring jobs definitions available

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchMonitoringJobDefinitions/?offset=0&limit=100 \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer true This many results will be skipped
limit query integer true At most this many results are returned
searchName query string false A human-readable name for the definition, must be unique across organisations.
deploymentId query string false Includes only definitions for this particular deployment

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "batchMonitoringJob": {
        "abortOnError": true,
        "batchJobType": "monitoring",
        "chunkSize": "auto",
        "columnNamesRemapping": {},
        "csvSettings": {
          "delimiter": ",",
          "encoding": "utf-8",
          "quotechar": "\""
        },
        "deploymentId": "string",
        "disableRowLevelErrorHandling": false,
        "explanationAlgorithm": "shap",
        "explanationClassNames": [
          "string"
        ],
        "explanationNumTopClasses": 1,
        "includePredictionStatus": false,
        "includeProbabilities": true,
        "includeProbabilitiesClasses": [],
        "intakeSettings": {
          "type": "localFile"
        },
        "maxExplanations": 0,
        "maxNgramExplanations": 0,
        "modelId": "string",
        "modelPackageId": "string",
        "monitoringAggregation": {
          "retentionPolicy": "samples",
          "retentionValue": 0
        },
        "monitoringBatchPrefix": "string",
        "monitoringColumns": {
          "actedUponColumn": "string",
          "actualsTimestampColumn": "string",
          "actualsValueColumn": "string",
          "associationIdColumn": "string",
          "customMetricId": "string",
          "customMetricTimestampColumn": "string",
          "customMetricTimestampFormat": "string",
          "customMetricValueColumn": "string",
          "monitoredStatusColumn": "string",
          "predictionsColumns": [
            {
              "className": "string",
              "columnName": "string"
            }
          ],
          "reportDrift": true,
          "reportPredictions": true,
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "monitoringOutputSettings": {
          "monitoredStatusColumn": "string",
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "numConcurrent": 0,
        "outputSettings": {
          "credentialId": "string",
          "format": "csv",
          "partitionColumns": [
            "string"
          ],
          "type": "azure",
          "url": "string"
        },
        "passthroughColumns": [
          "string"
        ],
        "passthroughColumnsSet": "all",
        "pinnedModelId": "string",
        "predictionInstance": {
          "apiKey": "string",
          "datarobotKey": "string",
          "hostName": "string",
          "sslEnabled": true
        },
        "predictionWarningEnabled": true,
        "redactedFields": [
          "string"
        ],
        "skipDriftTracking": false,
        "thresholdHigh": 0,
        "thresholdLow": 0,
        "timeseriesSettings": {
          "forecastPoint": "2019-08-24T14:15:22Z",
          "relaxKnownInAdvanceFeaturesCheck": false,
          "type": "forecast"
        }
      },
      "created": "2019-08-24T14:15:22Z",
      "createdBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      },
      "enabled": false,
      "id": "string",
      "lastFailedRunTime": "2019-08-24T14:15:22Z",
      "lastScheduledRunTime": "2019-08-24T14:15:22Z",
      "lastStartedJobStatus": "INITIALIZING",
      "lastStartedJobTime": "2019-08-24T14:15:22Z",
      "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
      "name": "string",
      "nextScheduledRunTime": "2019-08-24T14:15:22Z",
      "schedule": {
        "dayOfMonth": [
          "*"
        ],
        "dayOfWeek": [
          "*"
        ],
        "hour": [
          "*"
        ],
        "minute": [
          "*"
        ],
        "month": [
          "*"
        ]
      },
      "updated": "2019-08-24T14:15:22Z",
      "updatedBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      }
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Responses

Status Meaning Description Schema
200 OK List of all available jobs BatchMonitoringJobDefinitionsListResponse
422 Unprocessable Entity Your input data or query arguments did not work together None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/batchMonitoringJobDefinitions/

Create a Batch Monitoring Job definition. A configuration for a Batch Monitoring job which can either be executed manually upon request or on scheduled intervals, if enabled. The API payload is the same as for /batchJobs along with optional enabled and schedule items.

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/batchMonitoringJobDefinitions/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{BatchMonitoringJobDefinitionsCreate}'

Body parameter

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "enabled": true,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "name": "string",
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Parameters

Name In Type Required Description
body body BatchMonitoringJobDefinitionsCreate false none

Example responses

202 Response

{
  "batchMonitoringJob": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 0,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "enabled": false,
  "id": "string",
  "lastFailedRunTime": "2019-08-24T14:15:22Z",
  "lastScheduledRunTime": "2019-08-24T14:15:22Z",
  "lastStartedJobStatus": "INITIALIZING",
  "lastStartedJobTime": "2019-08-24T14:15:22Z",
  "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
  "name": "string",
  "nextScheduledRunTime": "2019-08-24T14:15:22Z",
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "updated": "2019-08-24T14:15:22Z",
  "updatedBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  }
}

Responses

Status Meaning Description Schema
202 Accepted Job details for the created Batch Monitoring job definition BatchMonitoringJobDefinitionsResponse
403 Forbidden You are not authorized to create a job definition on this deployment due to your permissions role None
422 Unprocessable Entity You tried to create a job definition with incompatible or missing parameters to create a fully functioning job definition None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

DELETE /api/v2/batchMonitoringJobDefinitions/{jobDefinitionId}/

Delete a Batch Prediction job definition

Code samples

# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/batchMonitoringJobDefinitions/{jobDefinitionId}/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
jobDefinitionId path string true ID of the Batch Prediction job definition

Responses

Status Meaning Description Schema
204 No Content none None
403 Forbidden You are not authorized to delete this job definition due to your permissions role None
404 Not Found Job was deleted, never existed or you do not have access to it None
409 Conflict Job could not be deleted, as there are currently running jobs in the queue. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/batchMonitoringJobDefinitions/{jobDefinitionId}/

Retrieve a Batch Monitoring job definition

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/batchMonitoringJobDefinitions/{jobDefinitionId}/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
jobDefinitionId path string true ID of the Batch Prediction job definition

Example responses

200 Response

{
  "batchMonitoringJob": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 0,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "enabled": false,
  "id": "string",
  "lastFailedRunTime": "2019-08-24T14:15:22Z",
  "lastScheduledRunTime": "2019-08-24T14:15:22Z",
  "lastStartedJobStatus": "INITIALIZING",
  "lastStartedJobTime": "2019-08-24T14:15:22Z",
  "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
  "name": "string",
  "nextScheduledRunTime": "2019-08-24T14:15:22Z",
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "updated": "2019-08-24T14:15:22Z",
  "updatedBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  }
}

Responses

Status Meaning Description Schema
200 OK Job details for the requested Batch Monitoring job definition BatchMonitoringJobDefinitionsResponse
404 Not Found Job was deleted, never existed or you do not have access to it None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

PATCH /api/v2/batchMonitoringJobDefinitions/{jobDefinitionId}/

Update a Batch Monitoring job definition

Code samples

# You can also use wget
curl -X PATCH https://app.datarobot.com/api/v2/batchMonitoringJobDefinitions/{jobDefinitionId}/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "enabled": true,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "name": "string",
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Parameters

Name In Type Required Description
jobDefinitionId path string true ID of the Batch Prediction job definition
body body BatchMonitoringJobDefinitionsUpdate false none

Example responses

200 Response

{
  "batchMonitoringJob": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 0,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "enabled": false,
  "id": "string",
  "lastFailedRunTime": "2019-08-24T14:15:22Z",
  "lastScheduledRunTime": "2019-08-24T14:15:22Z",
  "lastStartedJobStatus": "INITIALIZING",
  "lastStartedJobTime": "2019-08-24T14:15:22Z",
  "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
  "name": "string",
  "nextScheduledRunTime": "2019-08-24T14:15:22Z",
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "updated": "2019-08-24T14:15:22Z",
  "updatedBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  }
}

Responses

Status Meaning Description Schema
200 OK Job details for the updated Batch Monitoring job definition BatchMonitoringJobDefinitionsResponse
403 Forbidden You are not authorized to alter the contents of this monitoring job due to your permissions role None
404 Not Found Job was deleted, never existed or you do not have access to it None
409 Conflict You chose a name of your monitoring job that was already existing within your organization None
422 Unprocessable Entity Could not update the monitoring job. Possible reasons: {} None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/guardConfigurations/

List resource tags.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/guardConfigurations/?entityId=string&entityType=customModel \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer false This many results will be skipped.
limit query integer false At most this many results are returned.
entityId query string true Filter guard configurations by the given entity ID.
entityType query string true Entity type of the given entity ID.

Enumerated Values

Parameter Value
entityType [customModel, customModelVersion, playground]

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "awsAccount": "string",
      "awsModel": "amazon-titan",
      "awsRegion": "string",
      "createdAt": "2019-08-24T14:15:22Z",
      "creatorId": "string",
      "creatorName": "string",
      "deploymentId": "string",
      "description": "string",
      "entityId": "string",
      "entityType": "customModel",
      "errorMessage": "string",
      "googleModel": "chat-bison",
      "googleRegion": "string",
      "googleServiceAccount": "string",
      "id": "string",
      "intervention": {
        "action": "block",
        "allowedActions": [
          "block"
        ],
        "conditionLogic": "any",
        "conditions": [
          {
            "comparand": true,
            "comparator": "greaterThan"
          }
        ],
        "message": "string",
        "sendNotification": false
      },
      "isValid": true,
      "llmType": "openAi",
      "modelInfo": {
        "classNames": [
          "string"
        ],
        "inputColumnName": "string",
        "modelId": "string",
        "modelName": "",
        "outputColumnName": "string",
        "replacementTextColumnName": "",
        "targetType": "Binary"
      },
      "name": "string",
      "nemoInfo": {
        "actions": "string",
        "blockedTerms": "string",
        "credentialId": "string",
        "llmPrompts": "string",
        "mainConfig": "string",
        "railsConfig": "string"
      },
      "ootbType": "token_count",
      "openaiApiBase": "string",
      "openaiApiKey": "string",
      "openaiCredential": "string",
      "openaiDeploymentId": "string",
      "stages": [
        "prompt"
      ],
      "type": "guardModel"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Responses

Status Meaning Description Schema
200 OK none GuardConfigurationListResponse
400 Bad Request Request invalid, refer to messages for detail. None
404 Not Found User permissions problem. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/guardConfigurations/

Create a guard configuration.

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/guardConfigurations/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{GuardConfigurationCreate}'

Body parameter

{
  "awsAccount": "string",
  "awsModel": "amazon-titan",
  "awsRegion": "string",
  "deploymentId": "string",
  "description": "string",
  "entityId": "string",
  "entityType": "customModel",
  "googleModel": "chat-bison",
  "googleRegion": "string",
  "googleServiceAccount": "string",
  "intervention": {
    "action": "block",
    "allowedActions": [
      "block"
    ],
    "conditionLogic": "any",
    "conditions": [
      {
        "comparand": true,
        "comparator": "greaterThan"
      }
    ],
    "message": "string",
    "sendNotification": false
  },
  "llmType": "openAi",
  "modelInfo": {
    "classNames": [
      "string"
    ],
    "inputColumnName": "string",
    "modelId": "string",
    "modelName": "",
    "outputColumnName": "string",
    "replacementTextColumnName": "",
    "targetType": "Binary"
  },
  "name": "string",
  "nemoInfo": {
    "actions": "string",
    "blockedTerms": "string",
    "credentialId": "string",
    "llmPrompts": "string",
    "mainConfig": "string",
    "railsConfig": "string"
  },
  "openaiApiBase": "string",
  "openaiApiKey": "string",
  "openaiCredential": "string",
  "openaiDeploymentId": "string",
  "stages": [
    "prompt"
  ],
  "templateId": "string"
}

Parameters

Name In Type Required Description
body body GuardConfigurationCreate false none

Example responses

201 Response

{
  "awsAccount": "string",
  "awsModel": "amazon-titan",
  "awsRegion": "string",
  "createdAt": "2019-08-24T14:15:22Z",
  "creatorId": "string",
  "creatorName": "string",
  "deploymentId": "string",
  "description": "string",
  "entityId": "string",
  "entityType": "customModel",
  "errorMessage": "string",
  "googleModel": "chat-bison",
  "googleRegion": "string",
  "googleServiceAccount": "string",
  "id": "string",
  "intervention": {
    "action": "block",
    "allowedActions": [
      "block"
    ],
    "conditionLogic": "any",
    "conditions": [
      {
        "comparand": true,
        "comparator": "greaterThan"
      }
    ],
    "message": "string",
    "sendNotification": false
  },
  "isValid": true,
  "llmType": "openAi",
  "modelInfo": {
    "classNames": [
      "string"
    ],
    "inputColumnName": "string",
    "modelId": "string",
    "modelName": "",
    "outputColumnName": "string",
    "replacementTextColumnName": "",
    "targetType": "Binary"
  },
  "name": "string",
  "nemoInfo": {
    "actions": "string",
    "blockedTerms": "string",
    "credentialId": "string",
    "llmPrompts": "string",
    "mainConfig": "string",
    "railsConfig": "string"
  },
  "ootbType": "token_count",
  "openaiApiBase": "string",
  "openaiApiKey": "string",
  "openaiCredential": "string",
  "openaiDeploymentId": "string",
  "stages": [
    "prompt"
  ],
  "type": "guardModel"
}

Responses

Status Meaning Description Schema
201 Created none GuardConfigurationRetrieveResponse
404 Not Found Either the resource does not exist or the user does not have permission to create the configuration. None
409 Conflict The proposed configuration name is already in use for the same entity. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/guardConfigurations/predictionEnvironmentsInUse/

AShow prediction environments in use for moderation.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/guardConfigurations/predictionEnvironmentsInUse/?customModelVersionId=string \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer false This many results will be skipped.
limit query integer false At most this many results are returned.
customModelVersionId query string true Show prediction environment information for this custom model version.

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "id": "string",
      "name": "string",
      "usedBy": [
        {
          "configurationId": "string",
          "deploymentId": "string",
          "name": "string"
        }
      ]
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Responses

Status Meaning Description Schema
200 OK none GuardConfigurationPredictionEnvironmentsInUseListResponse
404 Not Found Either the resource does not exist or the user does not have permission to view the configuration. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

POST /api/v2/guardConfigurations/toNewCustomModelVersion/

Apply moderation configuration to a new custom model version.

Code samples

# You can also use wget
curl -X POST https://app.datarobot.com/api/v2/guardConfigurations/toNewCustomModelVersion/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{GuardConfigurationToCustomModelVersion}'

Body parameter

{
  "customModelId": "string",
  "data": [
    {
      "awsAccount": "string",
      "awsModel": "amazon-titan",
      "awsRegion": "string",
      "deploymentId": "string",
      "description": "string",
      "errorMessage": "string",
      "googleModel": "chat-bison",
      "googleRegion": "string",
      "googleServiceAccount": "string",
      "intervention": {
        "action": "block",
        "allowedActions": [
          "block"
        ],
        "conditionLogic": "any",
        "conditions": [
          {
            "comparand": true,
            "comparator": "greaterThan"
          }
        ],
        "message": "string",
        "sendNotification": false
      },
      "isValid": true,
      "llmType": "openAi",
      "modelInfo": {
        "classNames": [
          "string"
        ],
        "inputColumnName": "string",
        "modelId": "string",
        "modelName": "",
        "outputColumnName": "string",
        "replacementTextColumnName": "",
        "targetType": "Binary"
      },
      "name": "string",
      "nemoInfo": {
        "actions": "string",
        "blockedTerms": "string",
        "credentialId": "string",
        "llmPrompts": "string",
        "mainConfig": "string",
        "railsConfig": "string"
      },
      "ootbType": "token_count",
      "openaiApiBase": "string",
      "openaiApiKey": "string",
      "openaiCredential": "string",
      "openaiDeploymentId": "string",
      "parameters": [
        "s"
      ],
      "stages": [
        "prompt"
      ],
      "type": "guardModel"
    }
  ],
  "overallConfig": {
    "timeoutAction": "block",
    "timeoutSec": 2
  }
}

Parameters

Name In Type Required Description
body body GuardConfigurationToCustomModelVersion false none

Example responses

200 Response

{
  "customModelVersionId": "string"
}

Responses

Status Meaning Description Schema
200 OK none GuardConfigurationToCustomModelResponse
404 Not Found Either the resource does not exist or the user does not have permission to create the configuration. None
409 Conflict The destination custom model version is frozen. Create a new version to save configuration. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

DELETE /api/v2/guardConfigurations/{configId}/

Delete a guard config.

Code samples

# You can also use wget
curl -X DELETE https://app.datarobot.com/api/v2/guardConfigurations/{configId}/ \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
configId path string true ID of the configuration.

Responses

Status Meaning Description Schema
204 No Content none None
404 Not Found Either the config does not exist or the user does not have permission to delete it. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/guardConfigurations/{configId}/

Retrieve info about a guard configuration.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/guardConfigurations/{configId}/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
configId path string true ID of the configuration.

Example responses

200 Response

{
  "awsAccount": "string",
  "awsModel": "amazon-titan",
  "awsRegion": "string",
  "createdAt": "2019-08-24T14:15:22Z",
  "creatorId": "string",
  "creatorName": "string",
  "deploymentId": "string",
  "description": "string",
  "entityId": "string",
  "entityType": "customModel",
  "errorMessage": "string",
  "googleModel": "chat-bison",
  "googleRegion": "string",
  "googleServiceAccount": "string",
  "id": "string",
  "intervention": {
    "action": "block",
    "allowedActions": [
      "block"
    ],
    "conditionLogic": "any",
    "conditions": [
      {
        "comparand": true,
        "comparator": "greaterThan"
      }
    ],
    "message": "string",
    "sendNotification": false
  },
  "isValid": true,
  "llmType": "openAi",
  "modelInfo": {
    "classNames": [
      "string"
    ],
    "inputColumnName": "string",
    "modelId": "string",
    "modelName": "",
    "outputColumnName": "string",
    "replacementTextColumnName": "",
    "targetType": "Binary"
  },
  "name": "string",
  "nemoInfo": {
    "actions": "string",
    "blockedTerms": "string",
    "credentialId": "string",
    "llmPrompts": "string",
    "mainConfig": "string",
    "railsConfig": "string"
  },
  "ootbType": "token_count",
  "openaiApiBase": "string",
  "openaiApiKey": "string",
  "openaiCredential": "string",
  "openaiDeploymentId": "string",
  "stages": [
    "prompt"
  ],
  "type": "guardModel"
}

Responses

Status Meaning Description Schema
200 OK none GuardConfigurationRetrieveResponse
404 Not Found Either the config does not exist or the user does not have permission to view it. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

PATCH /api/v2/guardConfigurations/{configId}/

Update a guard config.

Code samples

# You can also use wget
curl -X PATCH https://app.datarobot.com/api/v2/guardConfigurations/{configId}/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "awsAccount": "string",
  "awsModel": "amazon-titan",
  "awsRegion": "string",
  "deploymentId": "string",
  "description": "string",
  "googleModel": "chat-bison",
  "googleRegion": "string",
  "googleServiceAccount": "string",
  "intervention": {
    "action": "block",
    "allowedActions": [
      "block"
    ],
    "conditionLogic": "any",
    "conditions": [
      {
        "comparand": true,
        "comparator": "greaterThan"
      }
    ],
    "message": "string",
    "sendNotification": false
  },
  "llmType": "openAi",
  "modelInfo": {
    "classNames": [
      "string"
    ],
    "inputColumnName": "string",
    "modelId": "string",
    "modelName": "",
    "outputColumnName": "string",
    "replacementTextColumnName": "",
    "targetType": "Binary"
  },
  "name": "string",
  "nemoInfo": {
    "actions": "string",
    "blockedTerms": "string",
    "credentialId": "string",
    "llmPrompts": "string",
    "mainConfig": "string",
    "railsConfig": "string"
  },
  "openaiApiBase": "string",
  "openaiApiKey": "string",
  "openaiCredential": "string",
  "openaiDeploymentId": "string"
}

Parameters

Name In Type Required Description
configId path string true ID of the configuration.
body body GuardConfigurationUpdate false none

Example responses

200 Response

{
  "awsAccount": "string",
  "awsModel": "amazon-titan",
  "awsRegion": "string",
  "createdAt": "2019-08-24T14:15:22Z",
  "creatorId": "string",
  "creatorName": "string",
  "deploymentId": "string",
  "description": "string",
  "entityId": "string",
  "entityType": "customModel",
  "errorMessage": "string",
  "googleModel": "chat-bison",
  "googleRegion": "string",
  "googleServiceAccount": "string",
  "id": "string",
  "intervention": {
    "action": "block",
    "allowedActions": [
      "block"
    ],
    "conditionLogic": "any",
    "conditions": [
      {
        "comparand": true,
        "comparator": "greaterThan"
      }
    ],
    "message": "string",
    "sendNotification": false
  },
  "isValid": true,
  "llmType": "openAi",
  "modelInfo": {
    "classNames": [
      "string"
    ],
    "inputColumnName": "string",
    "modelId": "string",
    "modelName": "",
    "outputColumnName": "string",
    "replacementTextColumnName": "",
    "targetType": "Binary"
  },
  "name": "string",
  "nemoInfo": {
    "actions": "string",
    "blockedTerms": "string",
    "credentialId": "string",
    "llmPrompts": "string",
    "mainConfig": "string",
    "railsConfig": "string"
  },
  "ootbType": "token_count",
  "openaiApiBase": "string",
  "openaiApiKey": "string",
  "openaiCredential": "string",
  "openaiDeploymentId": "string",
  "stages": [
    "prompt"
  ],
  "type": "guardModel"
}

Responses

Status Meaning Description Schema
200 OK none GuardConfigurationRetrieveResponse
404 Not Found Either the resource does not exist or the user does not have permission to create the configuration. None
409 Conflict The proposed configuration name is already in use for the same entity. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/guardTemplates/

List guard templates.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/guardTemplates/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer false This many results will be skipped.
limit query integer false At most this many results are returned.

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "allowedStages": [
        "prompt"
      ],
      "awsModel": "amazon-titan",
      "awsRegion": "string",
      "createdAt": "2019-08-24T14:15:22Z",
      "creatorId": "string",
      "creatorName": "string",
      "description": "string",
      "errorMessage": "string",
      "googleModel": "chat-bison",
      "googleRegion": "string",
      "id": "string",
      "intervention": {
        "action": "block",
        "allowedActions": [
          "block"
        ],
        "conditionLogic": "any",
        "conditions": [
          {
            "comparand": true,
            "comparator": "greaterThan"
          }
        ],
        "modifyMessage": "string",
        "sendNotification": true
      },
      "isValid": true,
      "llmType": "openAi",
      "modelInfo": {
        "classNames": [
          "string"
        ],
        "inputColumnName": "string",
        "modelId": "string",
        "modelName": "",
        "outputColumnName": "string",
        "replacementTextColumnName": "",
        "targetType": "Binary"
      },
      "name": "string",
      "nemoInfo": {
        "actions": "",
        "blockedTerms": "string",
        "credentialId": "string",
        "llmPrompts": "",
        "mainConfig": "string",
        "railsConfig": "string"
      },
      "ootbType": "token_count",
      "openaiApiBase": "string",
      "openaiApiKey": "string",
      "openaiDeploymentId": "string",
      "orgId": "string",
      "productionOnly": true,
      "type": "guardModel"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Responses

Status Meaning Description Schema
200 OK none GuardTemplateListResponse
400 Bad Request Request invalid, refer to messages for detail. None
404 Not Found Missing feature flag. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/guardTemplates/{templateId}/

Retrieve info about a guard template.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/guardTemplates/{templateId}/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
templateId path string true ID of the template.

Example responses

200 Response

{
  "allowedStages": [
    "prompt"
  ],
  "awsModel": "amazon-titan",
  "awsRegion": "string",
  "createdAt": "2019-08-24T14:15:22Z",
  "creatorId": "string",
  "creatorName": "string",
  "description": "string",
  "errorMessage": "string",
  "googleModel": "chat-bison",
  "googleRegion": "string",
  "id": "string",
  "intervention": {
    "action": "block",
    "allowedActions": [
      "block"
    ],
    "conditionLogic": "any",
    "conditions": [
      {
        "comparand": true,
        "comparator": "greaterThan"
      }
    ],
    "modifyMessage": "string",
    "sendNotification": true
  },
  "isValid": true,
  "llmType": "openAi",
  "modelInfo": {
    "classNames": [
      "string"
    ],
    "inputColumnName": "string",
    "modelId": "string",
    "modelName": "",
    "outputColumnName": "string",
    "replacementTextColumnName": "",
    "targetType": "Binary"
  },
  "name": "string",
  "nemoInfo": {
    "actions": "",
    "blockedTerms": "string",
    "credentialId": "string",
    "llmPrompts": "",
    "mainConfig": "string",
    "railsConfig": "string"
  },
  "ootbType": "token_count",
  "openaiApiBase": "string",
  "openaiApiKey": "string",
  "openaiDeploymentId": "string",
  "orgId": "string",
  "productionOnly": true,
  "type": "guardModel"
}

Responses

Status Meaning Description Schema
200 OK none GuardTemplateRetrieveResponse
404 Not Found Either the template does not exist or the required feature flag is missing. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/mlops/portablePredictionServerImage/

Fetches the latest Portable Prediction Server (PPS) Docker image. The resulting image can be docker-loaded. Since the image can be quite large (14GB+) consider querying its metadata to check the image size in advance and content hash to verify the downloaded image afterwards. In some environments it can HTTP redirect to some other service (like S3 or GCP) using pre-signed URLs.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/mlops/portablePredictionServerImage/ \
  -H "Authorization: Bearer {access-token}"

Responses

Status Meaning Description Schema
200 OK Download the latest available Portable Prediction Server (PPS) Docker image None
302 Found Redirect to another service for more efficient content download using pre-signed URL None
404 Not Found No PPS images found in the system None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/mlops/portablePredictionServerImage/metadata/

Fetches currently active PPS Docker image metadata

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/mlops/portablePredictionServerImage/metadata/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Example responses

200 Response

{
  "baseImageId": "string",
  "created": "2019-08-24T14:15:22Z",
  "datarobotRuntimeImageTag": "string",
  "dockerImageId": "string",
  "filename": "string",
  "hash": "string",
  "hashAlgorithm": "SHA256",
  "imageSize": 0,
  "shortDockerImageId": "string"
}

Responses

Status Meaning Description Schema
200 OK Available Portable Prediction Server (PPS) image metadata PPSImageMetadataResponse
404 Not Found No PPS images found in the system None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/moderationSupportedLlms/

List supported LLMs for moderation.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/moderationSupportedLlms/ \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
offset query integer false This many results will be skipped.
limit query integer false At most this many results are returned.

Example responses

200 Response

{
  "count": 0,
  "data": [
    {
      "description": "string",
      "id": "string",
      "llmType": "string",
      "model": "string",
      "name": "string",
      "provider": "string",
      "vendor": "string"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Responses

Status Meaning Description Schema
200 OK none SupportedLlmListResponse
400 Bad Request Request invalid, refer to messages for detail. None
404 Not Found Missing feature flag. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

GET /api/v2/overallModerationConfiguration/

Get overall moderation configuration for an entity.

Code samples

# You can also use wget
curl -X GET https://app.datarobot.com/api/v2/overallModerationConfiguration/?entityId=string&entityType=customModel \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}"

Parameters

Name In Type Required Description
entityId query string true Retrieve overall moderation configuration for the given entity ID.
entityType query string true Entity type of the given entity ID.

Enumerated Values

Parameter Value
entityType [customModel, customModelVersion, playground]

Example responses

200 Response

{
  "entityId": "string",
  "entityType": "customModel",
  "timeoutAction": "block",
  "timeoutSec": 2,
  "updatedAt": "2019-08-24T14:15:22Z",
  "updaterId": "string"
}

Responses

Status Meaning Description Schema
200 OK none OverallModerationConfigurationResponse
404 Not Found User permissions problem. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

PATCH /api/v2/overallModerationConfiguration/

Update overall moderation configuration for an entity.

Code samples

# You can also use wget
curl -X PATCH https://app.datarobot.com/api/v2/overallModerationConfiguration/ \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer {access-token}" \
  -d '{undefined}'

Body parameter

{
  "entityId": "string",
  "entityType": "customModel",
  "timeoutAction": "block",
  "timeoutSec": 0
}

Parameters

Name In Type Required Description
entityId path string true Retrieve overall moderation configuration for the given entity ID.
entityType path string true Entity type of the given entity ID.
body body OverallModerationConfigurationUpdate false none

Enumerated Values

Parameter Value
entityType [customModel, customModelVersion, playground]

Example responses

200 Response

{
  "entityId": "string",
  "entityType": "customModel",
  "timeoutAction": "block",
  "timeoutSec": 2,
  "updatedAt": "2019-08-24T14:15:22Z",
  "updaterId": "string"
}

Responses

Status Meaning Description Schema
200 OK none OverallModerationConfigurationResponse
404 Not Found Either the resource does not exist or the user does not have permission to create the configuration. None

To perform this operation, you must be authenticated by means of one of the following methods:

BearerAuth

Schemas

AzureDataStreamer

{
  "credentialId": "string",
  "format": "csv",
  "type": "azure",
  "url": "string"
}

Stream CSV data chunks from Azure

Properties

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false Use the specified credential to access the url

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
format string false Type of input file format
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
anonymous [redacted]
format [csv, parquet]
type azure

AzureIntake

{
  "credentialId": "string",
  "format": "csv",
  "type": "azure",
  "url": "string"
}

Stream CSV data chunks from Azure

Properties

Name Type Required Restrictions Description
credentialId string¦null false Use the specified credential to access the url
format string false Type of input file format
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
format [csv, parquet]
type azure

AzureOutput

{
  "credentialId": "string",
  "format": "csv",
  "partitionColumns": [
    "string"
  ],
  "type": "azure",
  "url": "string"
}

Save CSV data chunks to Azure Blob Storage

Properties

Name Type Required Restrictions Description
credentialId string¦null false Use the specified credential to access the url
format string false Type of output file format
partitionColumns [string] false maxItems: 100
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/").
type string true Type name for this output type
url string(url) true URL for the file or directory

Enumerated Values

Property Value
format [csv, parquet]
type azure

AzureOutputAdaptor

{
  "credentialId": "string",
  "format": "csv",
  "partitionColumns": [
    "string"
  ],
  "type": "azure",
  "url": "string"
}

Save CSV data chunks to Azure Blob Storage

Properties

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false Use the specified credential to access the url

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
format string false Type of output file format
partitionColumns [string] false maxItems: 100
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/").
type string true Type name for this output type
url string(url) true URL for the file or directory

Enumerated Values

Property Value
anonymous [redacted]
format [csv, parquet]
type azure

BatchJobCSVSettings

{
  "delimiter": ",",
  "encoding": "utf-8",
  "quotechar": "\""
}

The CSV settings used for this job

Properties

Name Type Required Restrictions Description
delimiter any true CSV fields are delimited by this character. Use the string "tab" to denote TSV (TAB separated values).

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous string false maxLength: 1
minLength: 1
minLength: 1
none

continued

Name Type Required Restrictions Description
encoding string true The encoding to be used for intake and output. For example (but not limited to): "shift_jis", "latin_1" or "mskanji".
quotechar string true maxLength: 1
minLength: 1
minLength: 1
Fields containing the delimiter or newlines must be quoted using this character.

Enumerated Values

Property Value
anonymous tab

BatchJobCreatedBy

{
  "fullName": "string",
  "userId": "string",
  "username": "string"
}

Who created this job

Properties

Name Type Required Restrictions Description
fullName string¦null true The full name of the user who created this job (if defined by the user)
userId string true The User ID of the user who created this job
username string true The username (e-mail address) of the user who created this job

BatchJobDefinitionResponse

{
  "createdBy": "string",
  "id": "string",
  "name": "string"
}

The Batch Prediction Job Definition linking to this job, if any.

Properties

Name Type Required Restrictions Description
createdBy string true The ID of creator of this job definition
id string true The ID of the Batch Prediction job definition
name string true A human-readable name for the definition, must be unique across organisations

BatchJobDefinitionsSpecResponse

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "maxNgramExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "numConcurrent": 0,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionWarningEnabled": true,
  "redactedFields": [
    "string"
  ],
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

The Batch Monitoring Job specification to be put on the queue in intervals

Properties

Name Type Required Restrictions Description
abortOnError boolean true Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchJobCSVSettings true The CSV settings used for this job
deploymentId string false ID of deployment which is used in job for processing predictions dataset
disableRowLevelErrorHandling boolean true Skip row by row error handling
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean true Include prediction status column in the output
includeProbabilities boolean true Include probabilities for all classes
includeProbabilitiesClasses [string] true maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any true The response option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureDataStreamer false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous DataStageDataStreamer false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous CatalogDataStreamer false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous GCPDataStreamer false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryDataStreamer false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous S3DataStreamer false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeDataStreamer false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseDataStreamer false Stream CSV data chunks from Azure Synapse

xor

Name Type Required Restrictions Description
» anonymous DSSDataStreamer false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemDataStreamer false none

xor

Name Type Required Restrictions Description
» anonymous HTTPDataStreamer false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCDataStreamer false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileDataStreamer false Stream CSV data chunks from local file storage

continued

Name Type Required Restrictions Description
maxExplanations integer true maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
maxNgramExplanations any false The maximum number of text ngram explanations to supply per row of the dataset. The default recommended maxNgramExplanations is all (no limit)

oneOf

Name Type Required Restrictions Description
» anonymous integer false minimum: 0
none

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringAggregation MonitoringAggregation false Defines the aggregation policy for monitoring jobs.
monitoringBatchPrefix string¦null false Name of the batch to create with this job
monitoringColumns MonitoringColumnsMapping false Column names mapping for monitoring
monitoringOutputSettings MonitoringOutputSettings false Output settings for monitoring jobs
numConcurrent integer false minimum: 0
Number of simultaneous requests to run against the prediction instance
outputSettings any false The response option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutputAdaptor false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous GCPOutputAdaptor false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutputAdaptor false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous S3OutputAdaptor false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutputAdaptor false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutputAdaptor false Save CSV data chunks to Azure Synapse in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutputAdaptor false none

xor

Name Type Required Restrictions Description
» anonymous HttpOutputAdaptor false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JdbcOutputAdaptor false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutputAdaptor false Save CSV data chunks to local file storage

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionWarningEnabled boolean¦null false Enable prediction warnings.
redactedFields [string] true A list of qualified field names from intake- and/or outputSettings that was redacted due to permissions and sharing settings. For example: intakeSettings.dataStoreId
skipDriftTracking boolean true Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsHistorical false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
anonymous all
passthroughColumnsSet all

{
  "csvUpload": "string",
  "download": "string",
  "self": "string"
}

Links useful for this job

Properties

Name Type Required Restrictions Description
csvUpload string(url) false The URL used to upload the dataset for this job. Only available for localFile intake.
download string¦null false The URL used to download the results from this job. Only available for localFile outputs. Will be null if the download is not yet available.
self string(url) true The URL used access this job.

BatchJobListResponse

{
  "count": 0,
  "data": [
    {
      "batchMonitoringJobDefinition": {
        "createdBy": "string",
        "id": "string",
        "name": "string"
      },
      "batchPredictionJobDefinition": {
        "createdBy": "string",
        "id": "string",
        "name": "string"
      },
      "created": "2019-08-24T14:15:22Z",
      "createdBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      },
      "elapsedTimeSec": 0,
      "failedRows": 0,
      "hidden": "2019-08-24T14:15:22Z",
      "id": "string",
      "intakeDatasetDisplayName": "string",
      "jobIntakeSize": 0,
      "jobOutputSize": 0,
      "jobSpec": {
        "abortOnError": true,
        "batchJobType": "monitoring",
        "chunkSize": "auto",
        "columnNamesRemapping": {},
        "csvSettings": {
          "delimiter": ",",
          "encoding": "utf-8",
          "quotechar": "\""
        },
        "deploymentId": "string",
        "disableRowLevelErrorHandling": false,
        "explanationAlgorithm": "shap",
        "explanationClassNames": [
          "string"
        ],
        "explanationNumTopClasses": 1,
        "includePredictionStatus": false,
        "includeProbabilities": true,
        "includeProbabilitiesClasses": [],
        "intakeSettings": {
          "type": "localFile"
        },
        "maxExplanations": 0,
        "maxNgramExplanations": 0,
        "modelId": "string",
        "modelPackageId": "string",
        "monitoringAggregation": {
          "retentionPolicy": "samples",
          "retentionValue": 0
        },
        "monitoringBatchPrefix": "string",
        "monitoringColumns": {
          "actedUponColumn": "string",
          "actualsTimestampColumn": "string",
          "actualsValueColumn": "string",
          "associationIdColumn": "string",
          "customMetricId": "string",
          "customMetricTimestampColumn": "string",
          "customMetricTimestampFormat": "string",
          "customMetricValueColumn": "string",
          "monitoredStatusColumn": "string",
          "predictionsColumns": [
            {
              "className": "string",
              "columnName": "string"
            }
          ],
          "reportDrift": true,
          "reportPredictions": true,
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "monitoringOutputSettings": {
          "monitoredStatusColumn": "string",
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "numConcurrent": 1,
        "outputSettings": {
          "credentialId": "string",
          "format": "csv",
          "partitionColumns": [
            "string"
          ],
          "type": "azure",
          "url": "string"
        },
        "passthroughColumns": [
          "string"
        ],
        "passthroughColumnsSet": "all",
        "pinnedModelId": "string",
        "predictionInstance": {
          "apiKey": "string",
          "datarobotKey": "string",
          "hostName": "string",
          "sslEnabled": true
        },
        "predictionWarningEnabled": true,
        "redactedFields": [
          "string"
        ],
        "skipDriftTracking": false,
        "thresholdHigh": 0,
        "thresholdLow": 0,
        "timeseriesSettings": {
          "forecastPoint": "2019-08-24T14:15:22Z",
          "relaxKnownInAdvanceFeaturesCheck": false,
          "type": "forecast"
        }
      },
      "links": {
        "csvUpload": "string",
        "download": "string",
        "self": "string"
      },
      "logs": [
        "string"
      ],
      "monitoringBatchId": "string",
      "percentageCompleted": 100,
      "queuePosition": 0,
      "queued": true,
      "resultsDeleted": true,
      "scoredRows": 0,
      "skippedRows": 0,
      "source": "string",
      "status": "INITIALIZING",
      "statusDetails": "string"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Properties

Name Type Required Restrictions Description
count integer false Number of items returned on this page.
data [BatchJobResponse] true maxItems: 10000
An array of jobs
next string(uri)¦null true URL pointing to the next page (if null, there is no next page).
previous string(uri)¦null true URL pointing to the previous page (if null, there is no previous page).
totalCount integer true The total number of items across all pages.

BatchJobPredictionInstance

{
  "apiKey": "string",
  "datarobotKey": "string",
  "hostName": "string",
  "sslEnabled": true
}

Override the default prediction instance from the deployment when scoring this job.

Properties

Name Type Required Restrictions Description
apiKey string false By default, prediction requests will use the API key of the user that created the job. This allows you to make requests on behalf of other users.
datarobotKey string false If running a job against a prediction instance in the Managed AI Cloud, you must provide the organization level DataRobot-Key.
hostName string true Override the default host name of the deployment with this.
sslEnabled boolean true Use SSL (HTTPS) when communicating with the overriden prediction server.

BatchJobRemapping

{
  "inputName": "string",
  "outputName": "string"
}

Properties

Name Type Required Restrictions Description
inputName string true Rename column with this name
outputName string¦null true Rename column to this name (leave as null to remove from the output)

BatchJobResponse

{
  "batchMonitoringJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "batchPredictionJobDefinition": {
    "createdBy": "string",
    "id": "string",
    "name": "string"
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "elapsedTimeSec": 0,
  "failedRows": 0,
  "hidden": "2019-08-24T14:15:22Z",
  "id": "string",
  "intakeDatasetDisplayName": "string",
  "jobIntakeSize": 0,
  "jobOutputSize": 0,
  "jobSpec": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 1,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "links": {
    "csvUpload": "string",
    "download": "string",
    "self": "string"
  },
  "logs": [
    "string"
  ],
  "monitoringBatchId": "string",
  "percentageCompleted": 100,
  "queuePosition": 0,
  "queued": true,
  "resultsDeleted": true,
  "scoredRows": 0,
  "skippedRows": 0,
  "source": "string",
  "status": "INITIALIZING",
  "statusDetails": "string"
}

Properties

Name Type Required Restrictions Description
batchMonitoringJobDefinition BatchJobDefinitionResponse false The Batch Prediction Job Definition linking to this job, if any.
batchPredictionJobDefinition BatchJobDefinitionResponse false The Batch Prediction Job Definition linking to this job, if any.
created string(date-time) true When was this job created
createdBy BatchJobCreatedBy true Who created this job
elapsedTimeSec integer true minimum: 0
Number of seconds the job has been processing for
failedRows integer true minimum: 0
Number of rows that have failed scoring
hidden string(date-time) false When was this job was hidden last, blank if visible
id string true The ID of the Batch job
intakeDatasetDisplayName string¦null false If applicable (e.g. for AI catalog), will contain the dataset name used for the intake dataset.
jobIntakeSize integer¦null true minimum: 0
Number of bytes in the intake dataset for this job
jobOutputSize integer¦null true minimum: 0
Number of bytes in the output dataset for this job
jobSpec BatchJobSpecResponse true The job configuration used to create this job
links BatchJobLinks true Links useful for this job
logs [string] true The job log.
monitoringBatchId string¦null true Id of the monitoring batch created by this job. Only present if the job runs on a deployment with batch monitoring enabled.
percentageCompleted number true maximum: 100
minimum: 0
Indicates job progress which is based on number of already processed rows in dataset
queuePosition integer¦null false minimum: 0
To ensure a dedicated prediction instance is not overloaded, only one job will be run against it at a time. This is the number of jobs that are awaiting processing before this job start running. May not be available in all environments.
queued boolean true The job has been put on the queue for execution.
resultsDeleted boolean false Indicates if the job was subject to garbage collection and had its artifacts deleted (output files, if any, and scoring data on local storage)
scoredRows integer true minimum: 0
Number of rows that have been used in prediction computation
skippedRows integer true minimum: 0
Number of rows that have been skipped during scoring. May contain non-zero value only in time-series predictions case if provided dataset contains more than required historical rows.
source string false Source from which batch job was started
status string true The current job status
statusDetails string true Explanation for current status

Enumerated Values

Property Value
status [INITIALIZING, RUNNING, COMPLETED, ABORTED, FAILED]

BatchJobSpecResponse

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "maxNgramExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionWarningEnabled": true,
  "redactedFields": [
    "string"
  ],
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

The job configuration used to create this job

Properties

Name Type Required Restrictions Description
abortOnError boolean true Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchJobCSVSettings true The CSV settings used for this job
deploymentId string false ID of deployment which is used in job for processing predictions dataset
disableRowLevelErrorHandling boolean true Skip row by row error handling
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean true Include prediction status column in the output
includeProbabilities boolean true Include probabilities for all classes
includeProbabilitiesClasses [string] true maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any true The response option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureDataStreamer false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous DataStageDataStreamer false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous CatalogDataStreamer false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous GCPDataStreamer false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryDataStreamer false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous S3DataStreamer false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeDataStreamer false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseDataStreamer false Stream CSV data chunks from Azure Synapse

xor

Name Type Required Restrictions Description
» anonymous DSSDataStreamer false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemDataStreamer false none

xor

Name Type Required Restrictions Description
» anonymous HTTPDataStreamer false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCDataStreamer false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileDataStreamer false Stream CSV data chunks from local file storage

continued

Name Type Required Restrictions Description
maxExplanations integer true maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
maxNgramExplanations any false The maximum number of text ngram explanations to supply per row of the dataset. The default recommended maxNgramExplanations is all (no limit)

oneOf

Name Type Required Restrictions Description
» anonymous integer false minimum: 0
none

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringAggregation MonitoringAggregation false Defines the aggregation policy for monitoring jobs.
monitoringBatchPrefix string¦null false Name of the batch to create with this job
monitoringColumns MonitoringColumnsMapping false Column names mapping for monitoring
monitoringOutputSettings MonitoringOutputSettings false Output settings for monitoring jobs
numConcurrent integer false minimum: 1
Number of simultaneous requests to run against the prediction instance
outputSettings any false The response option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutputAdaptor false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous GCPOutputAdaptor false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutputAdaptor false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous S3OutputAdaptor false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutputAdaptor false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutputAdaptor false Save CSV data chunks to Azure Synapse in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutputAdaptor false none

xor

Name Type Required Restrictions Description
» anonymous HttpOutputAdaptor false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JdbcOutputAdaptor false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutputAdaptor false Save CSV data chunks to local file storage

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionWarningEnabled boolean¦null false Enable prediction warnings.
redactedFields [string] true A list of qualified field names from intake- and/or outputSettings that was redacted due to permissions and sharing settings. For example: intakeSettings.dataStoreId
skipDriftTracking boolean true Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchJobTimeSeriesSettingsHistorical false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
anonymous all
passthroughColumnsSet all

BatchJobTimeSeriesSettingsForecast

{
  "forecastPoint": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "forecast"
}

Properties

Name Type Required Restrictions Description
forecastPoint string(date-time) false Used for forecast predictions in order to override the inferred forecast point from the dataset.
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Forecast mode makes predictions using forecastPoint or rows in the dataset without target.

Enumerated Values

Property Value
type forecast

BatchJobTimeSeriesSettingsHistorical

{
  "predictionsEndDate": "2019-08-24T14:15:22Z",
  "predictionsStartDate": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "historical"
}

Properties

Name Type Required Restrictions Description
predictionsEndDate string(date-time) false Used for historical predictions in order to override date to which predictions should be calculated. By default value will be inferred automatically from the dataset.
predictionsStartDate string(date-time) false Used for historical predictions in order to override date from which predictions should be calculated. By default value will be inferred automatically from the dataset.
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Historical mode enables bulk predictions which calculates predictions for all possible forecast points and forecast distances in the dataset within the predictionsStartDate/predictionsEndDate range.

Enumerated Values

Property Value
type historical

BatchMonitoringJobCreate

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Properties

Name Type Required Restrictions Description
abortOnError boolean true Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchPredictionJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchPredictionJobCSVSettings true The CSV settings used for this job
deploymentId string false ID of deployment which is used in job for processing predictions dataset
disableRowLevelErrorHandling boolean true Skip row by row error handling
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean true Include prediction status column in the output
includeProbabilities boolean true Include probabilities for all classes
includeProbabilitiesClasses [string] true maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any true The intake option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureIntake false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous BigQueryIntake false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous DataStageIntake false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous Catalog false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous DSS false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemIntake false none

xor

Name Type Required Restrictions Description
» anonymous GCPIntake false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPIntake false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCIntake false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileIntake false Stream CSV data chunks from local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Intake false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeIntake false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseIntake false Stream CSV data chunks from Azure Synapse

continued

Name Type Required Restrictions Description
maxExplanations integer true maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringAggregation MonitoringAggregation false Defines the aggregation policy for monitoring jobs.
monitoringBatchPrefix string¦null false Name of the batch to create with this job
monitoringColumns MonitoringColumnsMapping false Column names mapping for monitoring
monitoringOutputSettings MonitoringOutputSettings false Output settings for monitoring jobs
numConcurrent integer false minimum: 1
Number of simultaneous requests to run against the prediction instance
outputSettings any false The output option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutput false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutput false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutput false none

xor

Name Type Required Restrictions Description
» anonymous GCPOutput false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPOutput false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JDBCOutput false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutput false Save CSV data chunks to local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Output false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutput false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutput false Save CSV data chunks to Azure Synapse in bulk

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchPredictionJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionThreshold number false maximum: 1
minimum: 0
Threshold is the point that sets the class boundary for a predicted value. The model classifies an observation below the threshold as FALSE, and an observation above the threshold as TRUE. In other words, DataRobot automatically assigns the positive class label to any prediction exceeding the threshold. This value can be set between 0.0 and 1.0.
predictionWarningEnabled boolean¦null false Enable prediction warnings.
secondaryDatasetsConfigId string false Configuration id for secondary datasets to use when making a prediction.
skipDriftTracking boolean true Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsHistorical false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsTraining false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
passthroughColumnsSet all

BatchMonitoringJobDefinitionsCreate

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "enabled": true,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "name": "string",
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Properties

Name Type Required Restrictions Description
abortOnError boolean true Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchPredictionJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchPredictionJobCSVSettings true The CSV settings used for this job
deploymentId string true ID of deployment that the monitoring jobs is associated with.
disableRowLevelErrorHandling boolean true Skip row by row error handling
enabled boolean false If this job definition is enabled as a scheduled job. Optional if no schedule is supplied.
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean true Include prediction status column in the output
includeProbabilities boolean true Include probabilities for all classes
includeProbabilitiesClasses [string] true maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any true The intake option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureIntake false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous BigQueryIntake false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous DataStageIntake false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous Catalog false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous DSS false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemIntake false none

xor

Name Type Required Restrictions Description
» anonymous GCPIntake false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPIntake false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCIntake false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileIntake false Stream CSV data chunks from local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Intake false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeIntake false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseIntake false Stream CSV data chunks from Azure Synapse

continued

Name Type Required Restrictions Description
maxExplanations integer true maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringAggregation MonitoringAggregation false Defines the aggregation policy for monitoring jobs.
monitoringBatchPrefix string¦null false Name of the batch to create with this job
monitoringColumns MonitoringColumnsMapping false Column names mapping for monitoring
monitoringOutputSettings MonitoringOutputSettings false Output settings for monitoring jobs
name string false maxLength: 100
minLength: 1
minLength: 1
A human-readable name for the definition, must be unique across organisations, if left out the backend will generate one for you.
numConcurrent integer false minimum: 1
Number of simultaneous requests to run against the prediction instance
outputSettings any false The output option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutput false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutput false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutput false none

xor

Name Type Required Restrictions Description
» anonymous GCPOutput false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPOutput false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JDBCOutput false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutput false Save CSV data chunks to local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Output false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutput false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutput false Save CSV data chunks to Azure Synapse in bulk

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchPredictionJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionThreshold number false maximum: 1
minimum: 0
Threshold is the point that sets the class boundary for a predicted value. The model classifies an observation below the threshold as FALSE, and an observation above the threshold as TRUE. In other words, DataRobot automatically assigns the positive class label to any prediction exceeding the threshold. This value can be set between 0.0 and 1.0.
predictionWarningEnabled boolean¦null false Enable prediction warnings.
schedule Schedule false The scheduling information defining how often and when to execute this job to the Job Scheduling service. Optional if enabled = False.
secondaryDatasetsConfigId string false Configuration id for secondary datasets to use when making a prediction.
skipDriftTracking boolean true Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsHistorical false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsTraining false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
passthroughColumnsSet all

BatchMonitoringJobDefinitionsListResponse

{
  "count": 0,
  "data": [
    {
      "batchMonitoringJob": {
        "abortOnError": true,
        "batchJobType": "monitoring",
        "chunkSize": "auto",
        "columnNamesRemapping": {},
        "csvSettings": {
          "delimiter": ",",
          "encoding": "utf-8",
          "quotechar": "\""
        },
        "deploymentId": "string",
        "disableRowLevelErrorHandling": false,
        "explanationAlgorithm": "shap",
        "explanationClassNames": [
          "string"
        ],
        "explanationNumTopClasses": 1,
        "includePredictionStatus": false,
        "includeProbabilities": true,
        "includeProbabilitiesClasses": [],
        "intakeSettings": {
          "type": "localFile"
        },
        "maxExplanations": 0,
        "maxNgramExplanations": 0,
        "modelId": "string",
        "modelPackageId": "string",
        "monitoringAggregation": {
          "retentionPolicy": "samples",
          "retentionValue": 0
        },
        "monitoringBatchPrefix": "string",
        "monitoringColumns": {
          "actedUponColumn": "string",
          "actualsTimestampColumn": "string",
          "actualsValueColumn": "string",
          "associationIdColumn": "string",
          "customMetricId": "string",
          "customMetricTimestampColumn": "string",
          "customMetricTimestampFormat": "string",
          "customMetricValueColumn": "string",
          "monitoredStatusColumn": "string",
          "predictionsColumns": [
            {
              "className": "string",
              "columnName": "string"
            }
          ],
          "reportDrift": true,
          "reportPredictions": true,
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "monitoringOutputSettings": {
          "monitoredStatusColumn": "string",
          "uniqueRowIdentifierColumns": [
            "string"
          ]
        },
        "numConcurrent": 0,
        "outputSettings": {
          "credentialId": "string",
          "format": "csv",
          "partitionColumns": [
            "string"
          ],
          "type": "azure",
          "url": "string"
        },
        "passthroughColumns": [
          "string"
        ],
        "passthroughColumnsSet": "all",
        "pinnedModelId": "string",
        "predictionInstance": {
          "apiKey": "string",
          "datarobotKey": "string",
          "hostName": "string",
          "sslEnabled": true
        },
        "predictionWarningEnabled": true,
        "redactedFields": [
          "string"
        ],
        "skipDriftTracking": false,
        "thresholdHigh": 0,
        "thresholdLow": 0,
        "timeseriesSettings": {
          "forecastPoint": "2019-08-24T14:15:22Z",
          "relaxKnownInAdvanceFeaturesCheck": false,
          "type": "forecast"
        }
      },
      "created": "2019-08-24T14:15:22Z",
      "createdBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      },
      "enabled": false,
      "id": "string",
      "lastFailedRunTime": "2019-08-24T14:15:22Z",
      "lastScheduledRunTime": "2019-08-24T14:15:22Z",
      "lastStartedJobStatus": "INITIALIZING",
      "lastStartedJobTime": "2019-08-24T14:15:22Z",
      "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
      "name": "string",
      "nextScheduledRunTime": "2019-08-24T14:15:22Z",
      "schedule": {
        "dayOfMonth": [
          "*"
        ],
        "dayOfWeek": [
          "*"
        ],
        "hour": [
          "*"
        ],
        "minute": [
          "*"
        ],
        "month": [
          "*"
        ]
      },
      "updated": "2019-08-24T14:15:22Z",
      "updatedBy": {
        "fullName": "string",
        "userId": "string",
        "username": "string"
      }
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Properties

Name Type Required Restrictions Description
count integer false Number of items returned on this page.
data [BatchMonitoringJobDefinitionsResponse] true maxItems: 10000
An array of scheduled jobs
next string(uri)¦null true URL pointing to the next page (if null, there is no next page).
previous string(uri)¦null true URL pointing to the previous page (if null, there is no previous page).
totalCount integer true The total number of items across all pages.

BatchMonitoringJobDefinitionsResponse

{
  "batchMonitoringJob": {
    "abortOnError": true,
    "batchJobType": "monitoring",
    "chunkSize": "auto",
    "columnNamesRemapping": {},
    "csvSettings": {
      "delimiter": ",",
      "encoding": "utf-8",
      "quotechar": "\""
    },
    "deploymentId": "string",
    "disableRowLevelErrorHandling": false,
    "explanationAlgorithm": "shap",
    "explanationClassNames": [
      "string"
    ],
    "explanationNumTopClasses": 1,
    "includePredictionStatus": false,
    "includeProbabilities": true,
    "includeProbabilitiesClasses": [],
    "intakeSettings": {
      "type": "localFile"
    },
    "maxExplanations": 0,
    "maxNgramExplanations": 0,
    "modelId": "string",
    "modelPackageId": "string",
    "monitoringAggregation": {
      "retentionPolicy": "samples",
      "retentionValue": 0
    },
    "monitoringBatchPrefix": "string",
    "monitoringColumns": {
      "actedUponColumn": "string",
      "actualsTimestampColumn": "string",
      "actualsValueColumn": "string",
      "associationIdColumn": "string",
      "customMetricId": "string",
      "customMetricTimestampColumn": "string",
      "customMetricTimestampFormat": "string",
      "customMetricValueColumn": "string",
      "monitoredStatusColumn": "string",
      "predictionsColumns": [
        {
          "className": "string",
          "columnName": "string"
        }
      ],
      "reportDrift": true,
      "reportPredictions": true,
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "monitoringOutputSettings": {
      "monitoredStatusColumn": "string",
      "uniqueRowIdentifierColumns": [
        "string"
      ]
    },
    "numConcurrent": 0,
    "outputSettings": {
      "credentialId": "string",
      "format": "csv",
      "partitionColumns": [
        "string"
      ],
      "type": "azure",
      "url": "string"
    },
    "passthroughColumns": [
      "string"
    ],
    "passthroughColumnsSet": "all",
    "pinnedModelId": "string",
    "predictionInstance": {
      "apiKey": "string",
      "datarobotKey": "string",
      "hostName": "string",
      "sslEnabled": true
    },
    "predictionWarningEnabled": true,
    "redactedFields": [
      "string"
    ],
    "skipDriftTracking": false,
    "thresholdHigh": 0,
    "thresholdLow": 0,
    "timeseriesSettings": {
      "forecastPoint": "2019-08-24T14:15:22Z",
      "relaxKnownInAdvanceFeaturesCheck": false,
      "type": "forecast"
    }
  },
  "created": "2019-08-24T14:15:22Z",
  "createdBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  },
  "enabled": false,
  "id": "string",
  "lastFailedRunTime": "2019-08-24T14:15:22Z",
  "lastScheduledRunTime": "2019-08-24T14:15:22Z",
  "lastStartedJobStatus": "INITIALIZING",
  "lastStartedJobTime": "2019-08-24T14:15:22Z",
  "lastSuccessfulRunTime": "2019-08-24T14:15:22Z",
  "name": "string",
  "nextScheduledRunTime": "2019-08-24T14:15:22Z",
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "updated": "2019-08-24T14:15:22Z",
  "updatedBy": {
    "fullName": "string",
    "userId": "string",
    "username": "string"
  }
}

Properties

Name Type Required Restrictions Description
batchMonitoringJob BatchJobDefinitionsSpecResponse true The Batch Monitoring Job specification to be put on the queue in intervals
created string(date-time) true When was this job created
createdBy BatchJobCreatedBy true Who created this job
enabled boolean true If this job definition is enabled as a scheduled job.
id string true The ID of the Batch job definition
lastFailedRunTime string(date-time)¦null false Last time this job had a failed run
lastScheduledRunTime string(date-time)¦null false Last time this job was scheduled to run (though not guaranteed it actually ran at that time)
lastStartedJobStatus string¦null true The status of the latest job launched to the queue (if any).
lastStartedJobTime string(date-time)¦null true The last time (if any) a job was launched.
lastSuccessfulRunTime string(date-time)¦null false Last time this job had a successful run
name string true A human-readable name for the definition, must be unique across organisations
nextScheduledRunTime string(date-time)¦null false Next time this job is scheduled to run
schedule Schedule false The scheduling information defining how often and when to execute this job to the Job Scheduling service. Optional if enabled = False.
updated string(date-time) true When was this job last updated
updatedBy BatchJobCreatedBy true Who created this job

Enumerated Values

Property Value
lastStartedJobStatus [INITIALIZING, RUNNING, COMPLETED, ABORTED, FAILED]

BatchMonitoringJobDefinitionsUpdate

{
  "abortOnError": true,
  "batchJobType": "monitoring",
  "chunkSize": "auto",
  "columnNamesRemapping": {},
  "csvSettings": {
    "delimiter": ",",
    "encoding": "utf-8",
    "quotechar": "\""
  },
  "deploymentId": "string",
  "disableRowLevelErrorHandling": false,
  "enabled": true,
  "explanationAlgorithm": "shap",
  "explanationClassNames": [
    "string"
  ],
  "explanationNumTopClasses": 1,
  "includePredictionStatus": false,
  "includeProbabilities": true,
  "includeProbabilitiesClasses": [],
  "intakeSettings": {
    "type": "localFile"
  },
  "maxExplanations": 0,
  "modelId": "string",
  "modelPackageId": "string",
  "monitoringAggregation": {
    "retentionPolicy": "samples",
    "retentionValue": 0
  },
  "monitoringBatchPrefix": "string",
  "monitoringColumns": {
    "actedUponColumn": "string",
    "actualsTimestampColumn": "string",
    "actualsValueColumn": "string",
    "associationIdColumn": "string",
    "customMetricId": "string",
    "customMetricTimestampColumn": "string",
    "customMetricTimestampFormat": "string",
    "customMetricValueColumn": "string",
    "monitoredStatusColumn": "string",
    "predictionsColumns": [
      {
        "className": "string",
        "columnName": "string"
      }
    ],
    "reportDrift": true,
    "reportPredictions": true,
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "monitoringOutputSettings": {
    "monitoredStatusColumn": "string",
    "uniqueRowIdentifierColumns": [
      "string"
    ]
  },
  "name": "string",
  "numConcurrent": 1,
  "outputSettings": {
    "credentialId": "string",
    "format": "csv",
    "partitionColumns": [
      "string"
    ],
    "type": "azure",
    "url": "string"
  },
  "passthroughColumns": [
    "string"
  ],
  "passthroughColumnsSet": "all",
  "pinnedModelId": "string",
  "predictionInstance": {
    "apiKey": "string",
    "datarobotKey": "string",
    "hostName": "string",
    "sslEnabled": true
  },
  "predictionThreshold": 1,
  "predictionWarningEnabled": true,
  "schedule": {
    "dayOfMonth": [
      "*"
    ],
    "dayOfWeek": [
      "*"
    ],
    "hour": [
      "*"
    ],
    "minute": [
      "*"
    ],
    "month": [
      "*"
    ]
  },
  "secondaryDatasetsConfigId": "string",
  "skipDriftTracking": false,
  "thresholdHigh": 0,
  "thresholdLow": 0,
  "timeseriesSettings": {
    "forecastPoint": "2019-08-24T14:15:22Z",
    "relaxKnownInAdvanceFeaturesCheck": false,
    "type": "forecast"
  }
}

Properties

Name Type Required Restrictions Description
abortOnError boolean false Should this job abort if too many errors are encountered
batchJobType string false Batch job type.
chunkSize any false Which strategy should be used to determine the chunk size. Can be either a named strategy or a fixed size in bytes.

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous integer false maximum: 41943040
minimum: 20
none

continued

Name Type Required Restrictions Description
columnNamesRemapping any false Remap (rename or remove columns from) the output from this job

oneOf

Name Type Required Restrictions Description
» anonymous object false Provide a dictionary with key/value pairs to remap (deprecated)

xor

Name Type Required Restrictions Description
» anonymous [BatchPredictionJobRemapping] false maxItems: 1000
Provide a list of items to remap

continued

Name Type Required Restrictions Description
csvSettings BatchPredictionJobCSVSettings false The CSV settings used for this job
deploymentId string false ID of deployment which is used in job for processing predictions dataset
disableRowLevelErrorHandling boolean false Skip row by row error handling
enabled boolean false If this job definition is enabled as a scheduled job. Optional if no schedule is supplied.
explanationAlgorithm string false Which algorithm will be used to calculate prediction explanations
explanationClassNames [string] false maxItems: 10
minItems: 1
List of class names that will be explained for each row for multiclass. Mutually exclusive with explanationNumTopClasses. If neither specified - we assume explanationNumTopClasses=1
explanationNumTopClasses integer false maximum: 10
minimum: 1
Number of top predicted classes for each row that will be explained for multiclass. Mutually exclusive with explanationClassNames. If neither specified - we assume explanationNumTopClasses=1
includePredictionStatus boolean false Include prediction status column in the output
includeProbabilities boolean false Include probabilities for all classes
includeProbabilitiesClasses [string] false maxItems: 100
Include only probabilities for these specific class names.
intakeSettings any false The intake option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureIntake false Stream CSV data chunks from Azure

xor

Name Type Required Restrictions Description
» anonymous BigQueryIntake false Stream CSV data chunks from Big Query using GCS

xor

Name Type Required Restrictions Description
» anonymous DataStageIntake false Stream CSV data chunks from data stage storage

xor

Name Type Required Restrictions Description
» anonymous Catalog false Stream CSV data chunks from AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous DSS false Stream CSV data chunks from DSS dataset

xor

Name Type Required Restrictions Description
» anonymous FileSystemIntake false none

xor

Name Type Required Restrictions Description
» anonymous GCPIntake false Stream CSV data chunks from Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPIntake false Stream CSV data chunks from HTTP

xor

Name Type Required Restrictions Description
» anonymous JDBCIntake false Stream CSV data chunks from JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileIntake false Stream CSV data chunks from local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Intake false Stream CSV data chunks from Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeIntake false Stream CSV data chunks from Snowflake

xor

Name Type Required Restrictions Description
» anonymous SynapseIntake false Stream CSV data chunks from Azure Synapse

continued

Name Type Required Restrictions Description
maxExplanations integer false maximum: 100
minimum: 0
Number of explanations requested. Will be ordered by strength.
modelId string false ID of leaderboard model which is used in job for processing predictions dataset
modelPackageId string false ID of model package from registry is used in job for processing predictions dataset
monitoringAggregation MonitoringAggregation false Defines the aggregation policy for monitoring jobs.
monitoringBatchPrefix string¦null false Name of the batch to create with this job
monitoringColumns MonitoringColumnsMapping false Column names mapping for monitoring
monitoringOutputSettings MonitoringOutputSettings false Output settings for monitoring jobs
name string false maxLength: 100
minLength: 1
minLength: 1
A human-readable name for the definition, must be unique across organisations, if left out the backend will generate one for you.
numConcurrent integer false minimum: 1
Number of simultaneous requests to run against the prediction instance
outputSettings any false The output option configured for this job

oneOf

Name Type Required Restrictions Description
» anonymous AzureOutput false Save CSV data chunks to Azure Blob Storage

xor

Name Type Required Restrictions Description
» anonymous BigQueryOutput false Save CSV data chunks to Google BigQuery in bulk

xor

Name Type Required Restrictions Description
» anonymous FileSystemOutput false none

xor

Name Type Required Restrictions Description
» anonymous GCPOutput false Save CSV data chunks to Google Storage

xor

Name Type Required Restrictions Description
» anonymous HTTPOutput false Save CSV data chunks to HTTP data endpoint

xor

Name Type Required Restrictions Description
» anonymous JDBCOutput false Save CSV data chunks via JDBC

xor

Name Type Required Restrictions Description
» anonymous LocalFileOutput false Save CSV data chunks to local file storage

xor

Name Type Required Restrictions Description
» anonymous S3Output false Saves CSV data chunks to Amazon Cloud Storage S3

xor

Name Type Required Restrictions Description
» anonymous SnowflakeOutput false Save CSV data chunks to Snowflake in bulk

xor

Name Type Required Restrictions Description
» anonymous SynapseOutput false Save CSV data chunks to Azure Synapse in bulk

continued

Name Type Required Restrictions Description
passthroughColumns [string] false maxItems: 100
Pass through columns from the original dataset
passthroughColumnsSet string false Pass through all columns from the original dataset
pinnedModelId string false Specify a model ID used for scoring
predictionInstance BatchPredictionJobPredictionInstance false Override the default prediction instance from the deployment when scoring this job.
predictionThreshold number false maximum: 1
minimum: 0
Threshold is the point that sets the class boundary for a predicted value. The model classifies an observation below the threshold as FALSE, and an observation above the threshold as TRUE. In other words, DataRobot automatically assigns the positive class label to any prediction exceeding the threshold. This value can be set between 0.0 and 1.0.
predictionWarningEnabled boolean¦null false Enable prediction warnings.
schedule Schedule false The scheduling information defining how often and when to execute this job to the Job Scheduling service. Optional if enabled = False.
secondaryDatasetsConfigId string false Configuration id for secondary datasets to use when making a prediction.
skipDriftTracking boolean false Skip drift tracking for this job.
thresholdHigh number false Compute explanations for predictions above this threshold
thresholdLow number false Compute explanations for predictions below this threshold
timeseriesSettings any false Time Series settings included of this job is a Time Series job.

oneOf

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsForecast false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsHistorical false none

xor

Name Type Required Restrictions Description
» anonymous BatchPredictionJobTimeSeriesSettingsTraining false none

Enumerated Values

Property Value
batchJobType [monitoring, prediction]
anonymous [auto, fixed, dynamic]
explanationAlgorithm [shap, xemp]
passthroughColumnsSet all

BatchPredictionJobCSVSettings

{
  "delimiter": ",",
  "encoding": "utf-8",
  "quotechar": "\""
}

The CSV settings used for this job

Properties

Name Type Required Restrictions Description
delimiter any true CSV fields are delimited by this character. Use the string "tab" to denote TSV (TAB separated values).

oneOf

Name Type Required Restrictions Description
» anonymous string false none

xor

Name Type Required Restrictions Description
» anonymous string false maxLength: 1
minLength: 1
minLength: 1
none

continued

Name Type Required Restrictions Description
encoding string true The encoding to be used for intake and output. For example (but not limited to): "shift_jis", "latin_1" or "mskanji".
quotechar string true maxLength: 1
minLength: 1
minLength: 1
Fields containing the delimiter or newlines must be quoted using this character.

Enumerated Values

Property Value
anonymous tab

BatchPredictionJobDefinitionId

{
  "jobDefinitionId": "string"
}

Properties

Name Type Required Restrictions Description
jobDefinitionId string true ID of the Batch Prediction job definition

BatchPredictionJobPredictionInstance

{
  "apiKey": "string",
  "datarobotKey": "string",
  "hostName": "string",
  "sslEnabled": true
}

Override the default prediction instance from the deployment when scoring this job.

Properties

Name Type Required Restrictions Description
apiKey string false By default, prediction requests will use the API key of the user that created the job. This allows you to make requests on behalf of other users.
datarobotKey string false If running a job against a prediction instance in the Managed AI Cloud, you must provide the organization level DataRobot-Key.
hostName string true Override the default host name of the deployment with this.
sslEnabled boolean true Use SSL (HTTPS) when communicating with the overriden prediction server.

BatchPredictionJobRemapping

{
  "inputName": "string",
  "outputName": "string"
}

Properties

Name Type Required Restrictions Description
inputName string true Rename column with this name
outputName string¦null true Rename column to this name (leave as null to remove from the output)

BatchPredictionJobTimeSeriesSettingsForecast

{
  "forecastPoint": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "forecast"
}

Properties

Name Type Required Restrictions Description
forecastPoint string(date-time) false Used for forecast predictions in order to override the inferred forecast point from the dataset.
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Forecast mode makes predictions using forecastPoint or rows in the dataset without target.

Enumerated Values

Property Value
type forecast

BatchPredictionJobTimeSeriesSettingsHistorical

{
  "predictionsEndDate": "2019-08-24T14:15:22Z",
  "predictionsStartDate": "2019-08-24T14:15:22Z",
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "historical"
}

Properties

Name Type Required Restrictions Description
predictionsEndDate string(date-time) false Used for historical predictions in order to override date to which predictions should be calculated. By default value will be inferred automatically from the dataset.
predictionsStartDate string(date-time) false Used for historical predictions in order to override date from which predictions should be calculated. By default value will be inferred automatically from the dataset.
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Historical mode enables bulk predictions which calculates predictions for all possible forecast points and forecast distances in the dataset within the predictionsStartDate/predictionsEndDate range.

Enumerated Values

Property Value
type historical

BatchPredictionJobTimeSeriesSettingsTraining

{
  "relaxKnownInAdvanceFeaturesCheck": false,
  "type": "training"
}

Properties

Name Type Required Restrictions Description
relaxKnownInAdvanceFeaturesCheck boolean false If activated, missing values in the known in advance features are allowed in the forecast window at prediction time. If omitted or false, missing values are not allowed.
type string true Forecast mode used for making predictions on subsets of training data.

Enumerated Values

Property Value
type training

BigQueryDataStreamer

{
  "bucket": "string",
  "credentialId": "string",
  "dataset": "string",
  "table": "string",
  "type": "bigquery"
}

Stream CSV data chunks from Big Query using GCS

Properties

Name Type Required Restrictions Description
bucket string true The name of gcs bucket for data export
credentialId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false The ID of the GCP credentials

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataset string true The name of the specified big query dataset to read input data from
table string true The name of the specified big query table to read input data from
type string true Type name for this intake type

Enumerated Values

Property Value
anonymous [redacted]
type bigquery

BigQueryIntake

{
  "bucket": "string",
  "credentialId": "string",
  "dataset": "string",
  "table": "string",
  "type": "bigquery"
}

Stream CSV data chunks from Big Query using GCS

Properties

Name Type Required Restrictions Description
bucket string true The name of gcs bucket for data export
credentialId string true The ID of the GCP credentials
dataset string true The name of the specified big query dataset to read input data from
table string true The name of the specified big query table to read input data from
type string true Type name for this intake type

Enumerated Values

Property Value
type bigquery

BigQueryOutput

{
  "bucket": "string",
  "credentialId": "string",
  "dataset": "string",
  "table": "string",
  "type": "bigquery"
}

Save CSV data chunks to Google BigQuery in bulk

Properties

Name Type Required Restrictions Description
bucket string true The name of gcs bucket for data loading
credentialId string true The ID of the GCP credentials
dataset string true The name of the specified big query dataset to write data back
table string true The name of the specified big query table to write data back
type string true Type name for this output type

Enumerated Values

Property Value
type bigquery

BigQueryOutputAdaptor

{
  "bucket": "string",
  "credentialId": "string",
  "dataset": "string",
  "table": "string",
  "type": "bigquery"
}

Save CSV data chunks to Google BigQuery in bulk

Properties

Name Type Required Restrictions Description
bucket string true The name of gcs bucket for data loading
credentialId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false The ID of the GCP credentials

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataset string true The name of the specified big query dataset to write data back
table string true The name of the specified big query table to write data back
type string true Type name for this output type

Enumerated Values

Property Value
anonymous [redacted]
type bigquery

Catalog

{
  "datasetId": "string",
  "datasetVersionId": "string",
  "type": "dataset"
}

Stream CSV data chunks from AI catalog dataset

Properties

Name Type Required Restrictions Description
datasetId string true The ID of the AI catalog dataset
datasetVersionId string false The ID of the AI catalog dataset version
type string true Type name for this intake type

Enumerated Values

Property Value
type dataset

CatalogDataStreamer

{
  "datasetId": "string",
  "datasetVersionId": "string",
  "type": "dataset"
}

Stream CSV data chunks from AI catalog dataset

Properties

Name Type Required Restrictions Description
datasetId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false The ID of the AI catalog dataset

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
datasetVersionId string false The ID of the AI catalog dataset version
type string true Type name for this intake type

Enumerated Values

Property Value
anonymous [redacted]
type dataset

DSS

{
  "datasetId": "string",
  "partition": "holdout",
  "projectId": "string",
  "type": "dss"
}

Stream CSV data chunks from DSS dataset

Properties

Name Type Required Restrictions Description
datasetId string false The ID of the dataset
partition string false Partition used to predict
projectId string true The ID of the project
type string true Type name for this intake type

Enumerated Values

Property Value
partition [holdout, validation, allBacktests, None]
type dss

DSSDataStreamer

{
  "datasetId": "string",
  "partition": "holdout",
  "projectId": "string",
  "type": "dss"
}

Stream CSV data chunks from DSS dataset

Properties

Name Type Required Restrictions Description
datasetId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false The ID of the dataset

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
partition string false Partition used to predict
projectId string true The ID of the project
type string true Type name for this intake type

Enumerated Values

Property Value
anonymous [redacted]
partition [holdout, validation, allBacktests, None]
type dss

DataStageDataStreamer

{
  "dataStageId": "string",
  "type": "dataStage"
}

Stream CSV data chunks from data stage storage

Properties

Name Type Required Restrictions Description
dataStageId string true The ID of the data stage
type string true Type name for this intake type

Enumerated Values

Property Value
type dataStage

DataStageIntake

{
  "dataStageId": "string",
  "type": "dataStage"
}

Stream CSV data chunks from data stage storage

Properties

Name Type Required Restrictions Description
dataStageId string true The ID of the data stage
type string true Type name for this intake type

Enumerated Values

Property Value
type dataStage

DeploymentAndGuardResponse

{
  "configurationId": "string",
  "deploymentId": "string",
  "name": "string"
}

Properties

Name Type Required Restrictions Description
configurationId string true ID of guard configuration.
deploymentId string true ID of guard model deployment.
name string true Name of guard configuration.

FileSystemDataStreamer

{
  "path": "string",
  "type": "filesystem"
}

Properties

Name Type Required Restrictions Description
path string true Path to data on host filesystem
type string true Type name for this intake type

Enumerated Values

Property Value
type filesystem

FileSystemIntake

{
  "path": "string",
  "type": "filesystem"
}

Properties

Name Type Required Restrictions Description
path string true Path to data on host filesystem
type string true Type name for this intake type

Enumerated Values

Property Value
type filesystem

FileSystemOutput

{
  "path": "string",
  "type": "filesystem"
}

Properties

Name Type Required Restrictions Description
path string true Path to results on host filesystem
type string true Type name for this output type

Enumerated Values

Property Value
type filesystem

FileSystemOutputAdaptor

{
  "path": "string",
  "type": "filesystem"
}

Properties

Name Type Required Restrictions Description
path string true Path to results on host filesystem
type string true Type name for this output type

Enumerated Values

Property Value
type filesystem

GCPDataStreamer

{
  "credentialId": "string",
  "format": "csv",
  "type": "gcp",
  "url": "string"
}

Stream CSV data chunks from Google Storage

Properties

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false Use the specified credential to access the url

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
format string false Type of input file format
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
anonymous [redacted]
format [csv, parquet]
type gcp

GCPIntake

{
  "credentialId": "string",
  "format": "csv",
  "type": "gcp",
  "url": "string"
}

Stream CSV data chunks from Google Storage

Properties

Name Type Required Restrictions Description
credentialId string¦null false Use the specified credential to access the url
format string false Type of input file format
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
format [csv, parquet]
type gcp

GCPOutput

{
  "credentialId": "string",
  "format": "csv",
  "partitionColumns": [
    "string"
  ],
  "type": "gcp",
  "url": "string"
}

Save CSV data chunks to Google Storage

Properties

Name Type Required Restrictions Description
credentialId string¦null false Use the specified credential to access the url
format string false Type of input file format
partitionColumns [string] false maxItems: 100
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/").
type string true Type name for this output type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
format [csv, parquet]
type gcp

GCPOutputAdaptor

{
  "credentialId": "string",
  "format": "csv",
  "partitionColumns": [
    "string"
  ],
  "type": "gcp",
  "url": "string"
}

Save CSV data chunks to Google Storage

Properties

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false Use the specified credential to access the url

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
format string false Type of input file format
partitionColumns [string] false maxItems: 100
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/").
type string true Type name for this output type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
anonymous [redacted]
format [csv, parquet]
type gcp

GuardConditionResponse

{
  "comparand": true,
  "comparator": "greaterThan"
}

Condition to trigger intervention

Properties

Name Type Required Restrictions Description
comparand any true Condition comparand (basis of comparison)

anyOf

Name Type Required Restrictions Description
» anonymous boolean false none

or

Name Type Required Restrictions Description
» anonymous number false none

or

Name Type Required Restrictions Description
» anonymous string false none

or

Name Type Required Restrictions Description
» anonymous [string] false maxItems: 10
none

continued

Name Type Required Restrictions Description
comparator string true Condition comparator (operator)

Enumerated Values

Property Value
comparator [greaterThan, lessThan, equals, notEquals, is, isNot, matches, doesNotMatch, contains, doesNotContain]

GuardConfigurationConditionResponse

{
  "comparand": true,
  "comparator": "greaterThan"
}

Condition to trigger intervention

Properties

Name Type Required Restrictions Description
comparand any true Condition comparand (basis of comparison

anyOf

Name Type Required Restrictions Description
» anonymous boolean false none

or

Name Type Required Restrictions Description
» anonymous number false none

or

Name Type Required Restrictions Description
» anonymous string false none

or

Name Type Required Restrictions Description
» anonymous [string] false maxItems: 10
none

continued

Name Type Required Restrictions Description
comparator string true Condition comparator (operator)

Enumerated Values

Property Value
comparator [greaterThan, lessThan, equals, notEquals, is, isNot, matches, doesNotMatch, contains, doesNotContain]

GuardConfigurationCreate

{
  "awsAccount": "string",
  "awsModel": "amazon-titan",
  "awsRegion": "string",
  "deploymentId": "string",
  "description": "string",
  "entityId": "string",
  "entityType": "customModel",
  "googleModel": "chat-bison",
  "googleRegion": "string",
  "googleServiceAccount": "string",
  "intervention": {
    "action": "block",
    "allowedActions": [
      "block"
    ],
    "conditionLogic": "any",
    "conditions": [
      {
        "comparand": true,
        "comparator": "greaterThan"
      }
    ],
    "message": "string",
    "sendNotification": false
  },
  "llmType": "openAi",
  "modelInfo": {
    "classNames": [
      "string"
    ],
    "inputColumnName": "string",
    "modelId": "string",
    "modelName": "",
    "outputColumnName": "string",
    "replacementTextColumnName": "",
    "targetType": "Binary"
  },
  "name": "string",
  "nemoInfo": {
    "actions": "string",
    "blockedTerms": "string",
    "credentialId": "string",
    "llmPrompts": "string",
    "mainConfig": "string",
    "railsConfig": "string"
  },
  "openaiApiBase": "string",
  "openaiApiKey": "string",
  "openaiCredential": "string",
  "openaiDeploymentId": "string",
  "stages": [
    "prompt"
  ],
  "templateId": "string"
}

Properties

Name Type Required Restrictions Description
awsAccount string¦null false ID of user credential containing an AWS account.
awsModel string¦null false AWS model.
awsRegion string¦null false maxLength: 255
AWS model region.
deploymentId string¦null false ID of deployed model, for model guards.
description string false maxLength: 4096
Guard configuration description
entityId string true ID of custom model or playground for this guard.
entityType string true Type of associated entity.
googleModel string¦null false Google model.
googleRegion string¦null false maxLength: 255
Google model region.
googleServiceAccount string¦null false ID of user credential containing a Google service account.
intervention GuardConfigurationInterventionResponse false Intervention configuration for the guard.
llmType string¦null false Type of LLM used by this guard
modelInfo GuardConfigurationPayloadModelInfo false Configuration info for guards using deployed models.
name string true maxLength: 255
Guard configuration name
nemoInfo GuardConfigurationNemoInfoResponse false Configuration info for NeMo guards.
openaiApiBase string¦null false maxLength: 255
Azure OpenAI API Base URL
openaiApiKey string¦null false maxLength: 255
Deprecated; use openai_credential instead
openaiCredential string¦null false ID of user credential containing an OpenAI token.
openaiDeploymentId string¦null false maxLength: 255
Open API Deployment ID
stages [string] true maxItems: 16
The stages where the guard can run.
templateId string true ID of template this guard is based on.

Enumerated Values

Property Value
awsModel [amazon-titan, anthropic-claude-2, anthropic-claude-3-haiku, anthropic-claude-3-sonnet, anthropic-claude-3-opus]
entityType [customModel, customModelVersion, playground]
googleModel [chat-bison, google-gemini-1.5-flash, google-gemini-1.5-pro]
llmType [openAi, azureOpenAi, google, amazon, datarobot]

GuardConfigurationFullPost

{
  "awsAccount": "string",
  "awsModel": "amazon-titan",
  "awsRegion": "string",
  "deploymentId": "string",
  "description": "string",
  "errorMessage": "string",
  "googleModel": "chat-bison",
  "googleRegion": "string",
  "googleServiceAccount": "string",
  "intervention": {
    "action": "block",
    "allowedActions": [
      "block"
    ],
    "conditionLogic": "any",
    "conditions": [
      {
        "comparand": true,
        "comparator": "greaterThan"
      }
    ],
    "message": "string",
    "sendNotification": false
  },
  "isValid": true,
  "llmType": "openAi",
  "modelInfo": {
    "classNames": [
      "string"
    ],
    "inputColumnName": "string",
    "modelId": "string",
    "modelName": "",
    "outputColumnName": "string",
    "replacementTextColumnName": "",
    "targetType": "Binary"
  },
  "name": "string",
  "nemoInfo": {
    "actions": "string",
    "blockedTerms": "string",
    "credentialId": "string",
    "llmPrompts": "string",
    "mainConfig": "string",
    "railsConfig": "string"
  },
  "ootbType": "token_count",
  "openaiApiBase": "string",
  "openaiApiKey": "string",
  "openaiCredential": "string",
  "openaiDeploymentId": "string",
  "parameters": [
    "s"
  ],
  "stages": [
    "prompt"
  ],
  "type": "guardModel"
}

Complete guard configuration to push

Properties

Name Type Required Restrictions Description
awsAccount string¦null false ID of user credential containing an AWS account.
awsModel string¦null false AWS model.
awsRegion string¦null false maxLength: 255
AWS model region.
deploymentId string¦null false ID of deployed model, for model guards.
description string true maxLength: 4096
Guard configuration description
errorMessage string¦null false Error message if the guard configuration is invalid.
googleModel string¦null false Google model.
googleRegion string¦null false maxLength: 255
Google model region.
googleServiceAccount string¦null false ID of user credential containing a Google service account.
intervention GuardConfigurationInterventionResponse false Intervention configuration for the guard.
isValid boolean false Whether the guard is valid or not.
llmType string¦null false Type of LLM used by this guard
modelInfo GuardModelInfoResponse false Configuration info for guards using deployed models.
name string true maxLength: 255
Guard configuration name
nemoInfo GuardConfigurationNemoInfoResponse false Configuration info for NeMo guards.
ootbType string¦null false Guard template "Out of the Box" metric type
openaiApiBase string¦null false maxLength: 255
Azure OpenAI API Base URL
openaiApiKey string¦null false maxLength: 255
Deprecated; use openai_credential instead
openaiCredential string¦null false ID of user credential containing an OpenAI token.
openaiDeploymentId string¦null false maxLength: 255
OpenAI Deployment ID
parameters [string] false maxItems: 1
Parameter list, not used, deprecated.
stages [string] true maxItems: 16
The stages where the guard is configured to run.
type string true Guard configuration type

Enumerated Values

Property Value
awsModel [amazon-titan, anthropic-claude-2, anthropic-claude-3-haiku, anthropic-claude-3-sonnet, anthropic-claude-3-opus]
googleModel [chat-bison, google-gemini-1.5-flash, google-gemini-1.5-pro]
llmType [openAi, azureOpenAi, google, amazon, datarobot]
ootbType [token_count, faithfulness, rouge_1]
type [guardModel, nemo, ootb, pii, userModel]

GuardConfigurationInterventionResponse

{
  "action": "block",
  "allowedActions": [
    "block"
  ],
  "conditionLogic": "any",
  "conditions": [
    {
      "comparand": true,
      "comparator": "greaterThan"
    }
  ],
  "message": "string",
  "sendNotification": false
}

Intervention configuration for the guard.

Properties

Name Type Required Restrictions Description
action string true Action to take if conditions are met
allowedActions [string] false maxItems: 10
The actions this guard is allowed to take.
conditionLogic string false Action to take if conditions are met
conditions [GuardConfigurationConditionResponse] true maxItems: 1
List of conditions to trigger intervention
message string true maxLength: 4096
Message to use if prompt or response is blocked
sendNotification boolean false Create a notification event if intervention is triggered

Enumerated Values

Property Value
action [block, report, replace]
conditionLogic any

GuardConfigurationListResponse

{
  "count": 0,
  "data": [
    {
      "awsAccount": "string",
      "awsModel": "amazon-titan",
      "awsRegion": "string",
      "createdAt": "2019-08-24T14:15:22Z",
      "creatorId": "string",
      "creatorName": "string",
      "deploymentId": "string",
      "description": "string",
      "entityId": "string",
      "entityType": "customModel",
      "errorMessage": "string",
      "googleModel": "chat-bison",
      "googleRegion": "string",
      "googleServiceAccount": "string",
      "id": "string",
      "intervention": {
        "action": "block",
        "allowedActions": [
          "block"
        ],
        "conditionLogic": "any",
        "conditions": [
          {
            "comparand": true,
            "comparator": "greaterThan"
          }
        ],
        "message": "string",
        "sendNotification": false
      },
      "isValid": true,
      "llmType": "openAi",
      "modelInfo": {
        "classNames": [
          "string"
        ],
        "inputColumnName": "string",
        "modelId": "string",
        "modelName": "",
        "outputColumnName": "string",
        "replacementTextColumnName": "",
        "targetType": "Binary"
      },
      "name": "string",
      "nemoInfo": {
        "actions": "string",
        "blockedTerms": "string",
        "credentialId": "string",
        "llmPrompts": "string",
        "mainConfig": "string",
        "railsConfig": "string"
      },
      "ootbType": "token_count",
      "openaiApiBase": "string",
      "openaiApiKey": "string",
      "openaiCredential": "string",
      "openaiDeploymentId": "string",
      "stages": [
        "prompt"
      ],
      "type": "guardModel"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Properties

Name Type Required Restrictions Description
count integer false Number of items returned on this page.
data [GuardConfigurationRetrieveResponse] true maxItems: 200
list of guard configurations.
next string(uri)¦null true URL pointing to the next page (if null, there is no next page).
previous string(uri)¦null true URL pointing to the previous page (if null, there is no previous page).
totalCount integer true The total number of items across all pages.

GuardConfigurationNemoInfoResponse

{
  "actions": "string",
  "blockedTerms": "string",
  "credentialId": "string",
  "llmPrompts": "string",
  "mainConfig": "string",
  "railsConfig": "string"
}

Configuration info for NeMo guards.

Properties

Name Type Required Restrictions Description
actions string false maxLength: 4096
NeMo guardrails actions file
blockedTerms string true maxLength: 4096
NeMo guardrails blocked terms list
credentialId string¦null false NeMo guardrails credential ID (deprecated; use "openai_credential")
llmPrompts string false maxLength: 4096
NeMo guardrails prompts
mainConfig string true maxLength: 4096
Overall NeMo configuration YAML
railsConfig string true maxLength: 4096
NeMo guardrails configuration Colang

GuardConfigurationPayloadModelInfo

{
  "classNames": [
    "string"
  ],
  "inputColumnName": "string",
  "modelId": "string",
  "modelName": "",
  "outputColumnName": "string",
  "replacementTextColumnName": "",
  "targetType": "Binary"
}

Configuration info for guards using deployed models.

Properties

Name Type Required Restrictions Description
classNames [string] false maxItems: 100
List of class names for multiclass models
inputColumnName string true maxLength: 255
Input column name
modelId string¦null false ID of registered model, for model guards.
modelName string false maxLength: 255
ID of registered model, for .model guards
outputColumnName string true maxLength: 255
Output column name
replacementTextColumnName string false maxLength: 255
Name of the output column with replacement text. Required only if intervention.action is replace.
targetType string¦null false Target type

Enumerated Values

Property Value
targetType [Binary, Regression, Multiclass, TextGeneration]

GuardConfigurationPredictionEnvironmentsInUseListResponse

{
  "count": 0,
  "data": [
    {
      "id": "string",
      "name": "string",
      "usedBy": [
        {
          "configurationId": "string",
          "deploymentId": "string",
          "name": "string"
        }
      ]
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Properties

Name Type Required Restrictions Description
count integer false Number of items returned on this page.
data [PredictionEnvironmentInUseResponse] true maxItems: 32
list of prediction environments in use for this custom model version.
next string(uri)¦null true URL pointing to the next page (if null, there is no next page).
previous string(uri)¦null true URL pointing to the previous page (if null, there is no previous page).
totalCount integer true The total number of items across all pages.

GuardConfigurationRetrieveResponse

{
  "awsAccount": "string",
  "awsModel": "amazon-titan",
  "awsRegion": "string",
  "createdAt": "2019-08-24T14:15:22Z",
  "creatorId": "string",
  "creatorName": "string",
  "deploymentId": "string",
  "description": "string",
  "entityId": "string",
  "entityType": "customModel",
  "errorMessage": "string",
  "googleModel": "chat-bison",
  "googleRegion": "string",
  "googleServiceAccount": "string",
  "id": "string",
  "intervention": {
    "action": "block",
    "allowedActions": [
      "block"
    ],
    "conditionLogic": "any",
    "conditions": [
      {
        "comparand": true,
        "comparator": "greaterThan"
      }
    ],
    "message": "string",
    "sendNotification": false
  },
  "isValid": true,
  "llmType": "openAi",
  "modelInfo": {
    "classNames": [
      "string"
    ],
    "inputColumnName": "string",
    "modelId": "string",
    "modelName": "",
    "outputColumnName": "string",
    "replacementTextColumnName": "",
    "targetType": "Binary"
  },
  "name": "string",
  "nemoInfo": {
    "actions": "string",
    "blockedTerms": "string",
    "credentialId": "string",
    "llmPrompts": "string",
    "mainConfig": "string",
    "railsConfig": "string"
  },
  "ootbType": "token_count",
  "openaiApiBase": "string",
  "openaiApiKey": "string",
  "openaiCredential": "string",
  "openaiDeploymentId": "string",
  "stages": [
    "prompt"
  ],
  "type": "guardModel"
}

Properties

Name Type Required Restrictions Description
awsAccount string¦null false ID of user credential containing an AWS account.
awsModel string¦null false AWS model.
awsRegion string¦null false maxLength: 255
AWS model region.
createdAt string(date-time) true When the configuration was created.
creatorId string¦null false ID of the user who created the Guard configuration.
creatorName string false maxLength: 1000
Name of the user who created the Guard configuration.
deploymentId string¦null false ID of deployed model, for model guards.
description string true maxLength: 4096
Guard configuration description
entityId string¦null true ID of custom model or playground for this guard.
entityType string true Type of associated entity.
errorMessage string¦null false Error message if the guard configuration is invalid.
googleModel string¦null false Google model.
googleRegion string¦null false maxLength: 255
Google model region.
googleServiceAccount string¦null false ID of user credential containing a Google service account.
id string true Guard configuration object ID
intervention GuardConfigurationInterventionResponse false Intervention configuration for the guard.
isValid boolean false Whether the guard is valid or not.
llmType string¦null false Type of LLM used by this guard
modelInfo GuardModelInfoResponse false Configuration info for guards using deployed models.
name string true maxLength: 255
Guard configuration name
nemoInfo GuardConfigurationNemoInfoResponse false Configuration info for NeMo guards.
ootbType string¦null false Guard template "Out of the Box" metric type
openaiApiBase string¦null false maxLength: 255
Azure OpenAI API Base URL
openaiApiKey string¦null false maxLength: 255
Deprecated; use openai_credential instead
openaiCredential string¦null false ID of user credential containing an OpenAI token.
openaiDeploymentId string¦null false maxLength: 255
Open API Deployment ID
stages [string] true maxItems: 16
The stages where the guard is configured to run.
type string true Guard configuration type

Enumerated Values

Property Value
awsModel [amazon-titan, anthropic-claude-2, anthropic-claude-3-haiku, anthropic-claude-3-sonnet, anthropic-claude-3-opus]
entityType [customModel, customModelVersion, playground]
googleModel [chat-bison, google-gemini-1.5-flash, google-gemini-1.5-pro]
llmType [openAi, azureOpenAi, google, amazon, datarobot]
ootbType [token_count, faithfulness, rouge_1]
type [guardModel, nemo, ootb, pii, userModel]

GuardConfigurationToCustomModelResponse

{
  "customModelVersionId": "string"
}

Properties

Name Type Required Restrictions Description
customModelVersionId string true ID of the new custom model version created.

GuardConfigurationToCustomModelVersion

{
  "customModelId": "string",
  "data": [
    {
      "awsAccount": "string",
      "awsModel": "amazon-titan",
      "awsRegion": "string",
      "deploymentId": "string",
      "description": "string",
      "errorMessage": "string",
      "googleModel": "chat-bison",
      "googleRegion": "string",
      "googleServiceAccount": "string",
      "intervention": {
        "action": "block",
        "allowedActions": [
          "block"
        ],
        "conditionLogic": "any",
        "conditions": [
          {
            "comparand": true,
            "comparator": "greaterThan"
          }
        ],
        "message": "string",
        "sendNotification": false
      },
      "isValid": true,
      "llmType": "openAi",
      "modelInfo": {
        "classNames": [
          "string"
        ],
        "inputColumnName": "string",
        "modelId": "string",
        "modelName": "",
        "outputColumnName": "string",
        "replacementTextColumnName": "",
        "targetType": "Binary"
      },
      "name": "string",
      "nemoInfo": {
        "actions": "string",
        "blockedTerms": "string",
        "credentialId": "string",
        "llmPrompts": "string",
        "mainConfig": "string",
        "railsConfig": "string"
      },
      "ootbType": "token_count",
      "openaiApiBase": "string",
      "openaiApiKey": "string",
      "openaiCredential": "string",
      "openaiDeploymentId": "string",
      "parameters": [
        "s"
      ],
      "stages": [
        "prompt"
      ],
      "type": "guardModel"
    }
  ],
  "overallConfig": {
    "timeoutAction": "block",
    "timeoutSec": 2
  }
}

Properties

Name Type Required Restrictions Description
customModelId string true ID the custom model the user is working with.
data [GuardConfigurationFullPost] true maxItems: 200
List of complete guard configurations to push
overallConfig OverallConfigUpdate false Overall moderation configuration to push (not specific to one guard)

GuardConfigurationUpdate

{
  "awsAccount": "string",
  "awsModel": "amazon-titan",
  "awsRegion": "string",
  "deploymentId": "string",
  "description": "string",
  "googleModel": "chat-bison",
  "googleRegion": "string",
  "googleServiceAccount": "string",
  "intervention": {
    "action": "block",
    "allowedActions": [
      "block"
    ],
    "conditionLogic": "any",
    "conditions": [
      {
        "comparand": true,
        "comparator": "greaterThan"
      }
    ],
    "message": "string",
    "sendNotification": false
  },
  "llmType": "openAi",
  "modelInfo": {
    "classNames": [
      "string"
    ],
    "inputColumnName": "string",
    "modelId": "string",
    "modelName": "",
    "outputColumnName": "string",
    "replacementTextColumnName": "",
    "targetType": "Binary"
  },
  "name": "string",
  "nemoInfo": {
    "actions": "string",
    "blockedTerms": "string",
    "credentialId": "string",
    "llmPrompts": "string",
    "mainConfig": "string",
    "railsConfig": "string"
  },
  "openaiApiBase": "string",
  "openaiApiKey": "string",
  "openaiCredential": "string",
  "openaiDeploymentId": "string"
}

Properties

Name Type Required Restrictions Description
awsAccount string¦null false ID of user credential containing an AWS account.
awsModel string¦null false AWS model.
awsRegion string¦null false maxLength: 255
AWS model region.
deploymentId string¦null false ID of deployed model, for model guards.
description string false maxLength: 4096
Guard configuration description
googleModel string¦null false Google model.
googleRegion string¦null false maxLength: 255
Google model region.
googleServiceAccount string¦null false ID of user credential containing a Google service account.
intervention GuardConfigurationInterventionResponse false Intervention configuration for the guard.
llmType string¦null false Type of LLM used by this guard
modelInfo GuardConfigurationPayloadModelInfo false Configuration info for guards using deployed models.
name string false maxLength: 255
Guard configuration name
nemoInfo GuardConfigurationNemoInfoResponse false Configuration info for NeMo guards.
openaiApiBase string¦null false maxLength: 255
Azure OpenAI API Base URL
openaiApiKey string¦null false maxLength: 255
Deprecated; use openai_credential instead
openaiCredential string¦null false ID of user credential containing an OpenAI token.
openaiDeploymentId string¦null false maxLength: 255
Open API Deployment ID

Enumerated Values

Property Value
awsModel [amazon-titan, anthropic-claude-2, anthropic-claude-3-haiku, anthropic-claude-3-sonnet, anthropic-claude-3-opus]
googleModel [chat-bison, google-gemini-1.5-flash, google-gemini-1.5-pro]
llmType [openAi, azureOpenAi, google, amazon, datarobot]

GuardInterventionResponse

{
  "action": "block",
  "allowedActions": [
    "block"
  ],
  "conditionLogic": "any",
  "conditions": [
    {
      "comparand": true,
      "comparator": "greaterThan"
    }
  ],
  "modifyMessage": "string",
  "sendNotification": true
}

Intervention configuration for the guard.

Properties

Name Type Required Restrictions Description
action string true Action to take if conditions are met
allowedActions [string] false maxItems: 10
The actions this guard is allowed to take.
conditionLogic string false Action to take if conditions are met
conditions [GuardConditionResponse] true maxItems: 1
List of conditions to trigger intervention
modifyMessage string true maxLength: 4096
Message to use if prompt or response is blocked
sendNotification boolean false Create a notification event if intervention is triggered

Enumerated Values

Property Value
action [block, report, replace]
conditionLogic any

GuardModelInfoResponse

{
  "classNames": [
    "string"
  ],
  "inputColumnName": "string",
  "modelId": "string",
  "modelName": "",
  "outputColumnName": "string",
  "replacementTextColumnName": "",
  "targetType": "Binary"
}

Configuration info for guards using deployed models.

Properties

Name Type Required Restrictions Description
classNames [string] false maxItems: 100
List of class names for multiclass models
inputColumnName string true maxLength: 255
Input column name
modelId string¦null false ID of registered model, for model guards.
modelName string false maxLength: 255
ID of registered model, for .model guards
outputColumnName string true maxLength: 255
Output column name
replacementTextColumnName string false maxLength: 255
Name of the output column with replacement text. Required only if intervention.action is replace.
targetType string true Target type

Enumerated Values

Property Value
targetType [Binary, Regression, Multiclass, TextGeneration]

GuardNemoInfoResponse

{
  "actions": "",
  "blockedTerms": "string",
  "credentialId": "string",
  "llmPrompts": "",
  "mainConfig": "string",
  "railsConfig": "string"
}

Configuration info for NeMo guards.

Properties

Name Type Required Restrictions Description
actions string false maxLength: 4096
NeMo guardrails actions
blockedTerms string true maxLength: 4096
NeMo guardrails blocked terms list
credentialId string¦null false NeMo guardrails credential ID
llmPrompts string false maxLength: 4096
NeMo guardrails prompts
mainConfig string true maxLength: 4096
Overall NeMo configuration YAML
railsConfig string true maxLength: 4096
NeMo guardrails configuration Colang

GuardTemplateListResponse

{
  "count": 0,
  "data": [
    {
      "allowedStages": [
        "prompt"
      ],
      "awsModel": "amazon-titan",
      "awsRegion": "string",
      "createdAt": "2019-08-24T14:15:22Z",
      "creatorId": "string",
      "creatorName": "string",
      "description": "string",
      "errorMessage": "string",
      "googleModel": "chat-bison",
      "googleRegion": "string",
      "id": "string",
      "intervention": {
        "action": "block",
        "allowedActions": [
          "block"
        ],
        "conditionLogic": "any",
        "conditions": [
          {
            "comparand": true,
            "comparator": "greaterThan"
          }
        ],
        "modifyMessage": "string",
        "sendNotification": true
      },
      "isValid": true,
      "llmType": "openAi",
      "modelInfo": {
        "classNames": [
          "string"
        ],
        "inputColumnName": "string",
        "modelId": "string",
        "modelName": "",
        "outputColumnName": "string",
        "replacementTextColumnName": "",
        "targetType": "Binary"
      },
      "name": "string",
      "nemoInfo": {
        "actions": "",
        "blockedTerms": "string",
        "credentialId": "string",
        "llmPrompts": "",
        "mainConfig": "string",
        "railsConfig": "string"
      },
      "ootbType": "token_count",
      "openaiApiBase": "string",
      "openaiApiKey": "string",
      "openaiDeploymentId": "string",
      "orgId": "string",
      "productionOnly": true,
      "type": "guardModel"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Properties

Name Type Required Restrictions Description
count integer false Number of items returned on this page.
data [GuardTemplateRetrieveResponse] true maxItems: 200
list of guard templates.
next string(uri)¦null true URL pointing to the next page (if null, there is no next page).
previous string(uri)¦null true URL pointing to the previous page (if null, there is no previous page).
totalCount integer true The total number of items across all pages.

GuardTemplateRetrieveResponse

{
  "allowedStages": [
    "prompt"
  ],
  "awsModel": "amazon-titan",
  "awsRegion": "string",
  "createdAt": "2019-08-24T14:15:22Z",
  "creatorId": "string",
  "creatorName": "string",
  "description": "string",
  "errorMessage": "string",
  "googleModel": "chat-bison",
  "googleRegion": "string",
  "id": "string",
  "intervention": {
    "action": "block",
    "allowedActions": [
      "block"
    ],
    "conditionLogic": "any",
    "conditions": [
      {
        "comparand": true,
        "comparator": "greaterThan"
      }
    ],
    "modifyMessage": "string",
    "sendNotification": true
  },
  "isValid": true,
  "llmType": "openAi",
  "modelInfo": {
    "classNames": [
      "string"
    ],
    "inputColumnName": "string",
    "modelId": "string",
    "modelName": "",
    "outputColumnName": "string",
    "replacementTextColumnName": "",
    "targetType": "Binary"
  },
  "name": "string",
  "nemoInfo": {
    "actions": "",
    "blockedTerms": "string",
    "credentialId": "string",
    "llmPrompts": "",
    "mainConfig": "string",
    "railsConfig": "string"
  },
  "ootbType": "token_count",
  "openaiApiBase": "string",
  "openaiApiKey": "string",
  "openaiDeploymentId": "string",
  "orgId": "string",
  "productionOnly": true,
  "type": "guardModel"
}

Properties

Name Type Required Restrictions Description
allowedStages [string] true maxItems: 16
The stages where the guard can run.
awsModel string¦null false AWS model.
awsRegion string¦null false maxLength: 255
AWS model region.
createdAt string(date-time) true When the template was created.
creatorId string¦null false ID of the user who created the Guard template.
creatorName string false maxLength: 1000
ID of the user who created the Guard template.
description string true maxLength: 4096
Guard template description
errorMessage string¦null false Error message if the guard configuration is invalid.
googleModel string¦null false Google model.
googleRegion string¦null false maxLength: 255
Google model region.
id string true Guard template object ID
intervention GuardInterventionResponse false Intervention configuration for the guard.
isValid boolean false True if the guard is fully configured and valid.
llmType string¦null false Type of LLM used by this guard
modelInfo GuardModelInfoResponse false Configuration info for guards using deployed models.
name string true maxLength: 255
Guard template name
nemoInfo GuardNemoInfoResponse false Configuration info for NeMo guards.
ootbType string¦null false Guard template "Out of the Box" metric type
openaiApiBase string¦null false maxLength: 255
Azure OpenAI API Base URL
openaiApiKey string¦null false maxLength: 255
Deprecated; use openai_credential instead
openaiDeploymentId string¦null false maxLength: 255
Open API Deployment ID
orgId string¦null false Organization ID of the user who created the Guard template.
productionOnly boolean¦null false Whether the guard is for production only, or if it can be used in production and playground.
type string true Guard template type

Enumerated Values

Property Value
awsModel [amazon-titan, anthropic-claude-2, anthropic-claude-3-haiku, anthropic-claude-3-sonnet, anthropic-claude-3-opus]
googleModel [chat-bison, google-gemini-1.5-flash, google-gemini-1.5-pro]
llmType [openAi, azureOpenAi, google, amazon, datarobot]
ootbType [token_count, faithfulness, rouge_1]
type [guardModel, nemo, ootb, pii, userModel]

HTTPDataStreamer

{
  "type": "http",
  "url": "string"
}

Stream CSV data chunks from HTTP

Properties

Name Type Required Restrictions Description
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
type http

HTTPIntake

{
  "type": "http",
  "url": "string"
}

Stream CSV data chunks from HTTP

Properties

Name Type Required Restrictions Description
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
type http

HTTPOutput

{
  "headers": {},
  "method": "POST",
  "type": "http",
  "url": "string"
}

Save CSV data chunks to HTTP data endpoint

Properties

Name Type Required Restrictions Description
headers object false Extra headers to send with the request
method string true Method to use when saving the CSV file
type string true Type name for this output type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
method [POST, PUT]
type http

HttpOutputAdaptor

{
  "headers": {},
  "method": "POST",
  "type": "http",
  "url": "string"
}

Save CSV data chunks to HTTP data endpoint

Properties

Name Type Required Restrictions Description
headers object false Extra headers to send with the request
method string true Method to use when saving the CSV file
type string true Type name for this output type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
method [POST, PUT]
type http

JDBCDataStreamer

{
  "catalog": "string",
  "credentialId": "string",
  "dataStoreId": "string",
  "fetchSize": 1,
  "query": "string",
  "schema": "string",
  "table": "string",
  "type": "jdbc"
}

Stream CSV data chunks from JDBC

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to read input data from.
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false The ID of the credential holding information about a user with read access to the JDBC data source.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataStoreId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false ID of the data store to connect to

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
fetchSize integer false maximum: 1000000
minimum: 1
A user specified fetch size. Changing it can be used to balance throughput and memory usage. Deprecated and ignored since v2.21.
query string false A self-supplied SELECT statement of the dataset you wish to score. Helpful for supplying a more fine-grained selection of data not achievable through specification of "table" and/or "schema" parameters exclusively.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
schema string false The name of the specified database schema to read input data from.
table string false The name of the specified database table to read input data from.
type string true Type name for this intake type

Enumerated Values

Property Value
anonymous [redacted]
anonymous [redacted]
type jdbc

JDBCIntake

{
  "catalog": "string",
  "credentialId": "string",
  "dataStoreId": "string",
  "fetchSize": 1,
  "query": "string",
  "schema": "string",
  "table": "string",
  "type": "jdbc"
}

Stream CSV data chunks from JDBC

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to read input data from.
credentialId string¦null false The ID of the credential holding information about a user with read access to the JDBC data source.
dataStoreId string true ID of the data store to connect to
fetchSize integer false maximum: 1000000
minimum: 1
A user specified fetch size. Changing it can be used to balance throughput and memory usage. Deprecated and ignored since v2.21.
query string false A self-supplied SELECT statement of the dataset you wish to score. Helpful for supplying a more fine-grained selection of data not achievable through specification of "table" and/or "schema" parameters exclusively.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
schema string false The name of the specified database schema to read input data from.
table string false The name of the specified database table to read input data from.
type string true Type name for this intake type

Enumerated Values

Property Value
type jdbc

JDBCOutput

{
  "catalog": "string",
  "commitInterval": 600,
  "createTableIfNotExists": false,
  "credentialId": "string",
  "dataStoreId": "string",
  "schema": "string",
  "statementType": "createTable",
  "table": "string",
  "type": "jdbc",
  "updateColumns": [
    "string"
  ],
  "whereColumns": [
    "string"
  ]
}

Save CSV data chunks via JDBC

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to write output data to.
commitInterval integer false maximum: 86400
minimum: 0
Defines a time interval in seconds between each commit is done to the JDBC source. If set to 0, the batch prediction operation will write the entire job before committing.
createTableIfNotExists boolean false Attempt to create the table first if no existing one is detected, before writing data with the strategy defined in the statementType parameter.
credentialId string¦null false The ID of the credential holding information about a user with write access to the JDBC data source.
dataStoreId string true ID of the data store to connect to
schema string false The name of the specified database schema to write the results to.
statementType string true The statement type to use when writing the results. Deprecation Warning: Use of create_table is now discouraged. Use one of the other possibilities along with the parameter createTableIfNotExists set to true.
table string true The name of the specified database table to write the results to.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
type string true Type name for this intake type
updateColumns [string] false maxItems: 100
The column names to be updated if statementType is set to either update or upsert.
whereColumns [string] false maxItems: 100
The column names to be used in the where clause if statementType is set to update or upsert.

Enumerated Values

Property Value
statementType [createTable, create_table, insert, insertUpdate, insert_update, update]
type jdbc

JdbcOutputAdaptor

{
  "catalog": "string",
  "commitInterval": 600,
  "createTableIfNotExists": false,
  "credentialId": "string",
  "dataStoreId": "string",
  "schema": "string",
  "statementType": "createTable",
  "table": "string",
  "type": "jdbc",
  "updateColumns": [
    "string"
  ],
  "whereColumns": [
    "string"
  ]
}

Save CSV data chunks via JDBC

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to write output data to.
commitInterval integer false maximum: 86400
minimum: 0
Defines a time interval in seconds between each commit is done to the JDBC source. If set to 0, the batch prediction operation will write the entire job before committing.
createTableIfNotExists boolean false Attempt to create the table first if no existing one is detected, before writing data with the strategy defined in the statementType parameter.
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false The ID of the credential holding information about a user with write access to the JDBC data source.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataStoreId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false ID of the data store to connect to

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
schema string false The name of the specified database schema to write the results to.
statementType string true The statement type to use when writing the results. Deprecation Warning: Use of create_table is now discouraged. Use one of the other possibilities along with the parameter createTableIfNotExists set to true.
table string true The name of the specified database table to write the results to.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
type string true Type name for this intake type
updateColumns [string] false maxItems: 100
The column names to be updated if statementType is set to either update or upsert.
whereColumns [string] false maxItems: 100
The column names to be used in the where clause if statementType is set to update or upsert.

Enumerated Values

Property Value
anonymous [redacted]
anonymous [redacted]
statementType [createTable, create_table, insert, insertUpdate, insert_update, update]
type jdbc

LocalFileDataStreamer

{
  "async": true,
  "multipart": true,
  "type": "local_file"
}

Stream CSV data chunks from local file storage

Properties

Name Type Required Restrictions Description
async boolean¦null false The default behavior (async: true) will still submit the job to the queue and start processing as soon as the upload is started.Setting it to false will postpone submitting the job to the queue until all data has been uploaded.This is helpful if the user is on a bad connection and bottlednecked by the upload speed. Instead of blocking the queue this will allow others to submit to the queue until the upload has finished.
multipart boolean false specify if the data will be uploaded in multiple parts instead of a single file
type string true Type name for this intake type

Enumerated Values

Property Value
type [local_file, localFile]

LocalFileIntake

{
  "async": true,
  "multipart": true,
  "type": "local_file"
}

Stream CSV data chunks from local file storage

Properties

Name Type Required Restrictions Description
async boolean¦null false The default behavior (async: true) will still submit the job to the queue and start processing as soon as the upload is started.Setting it to false will postpone submitting the job to the queue until all data has been uploaded.This is helpful if the user is on a bad connection and bottlednecked by the upload speed. Instead of blocking the queue this will allow others to submit to the queue until the upload has finished.
multipart boolean false specify if the data will be uploaded in multiple parts instead of a single file
type string true Type name for this intake type

Enumerated Values

Property Value
type [local_file, localFile]

LocalFileOutput

{
  "type": "local_file"
}

Save CSV data chunks to local file storage

Properties

Name Type Required Restrictions Description
type string true Type name for this output type

Enumerated Values

Property Value
type [local_file, localFile]

LocalFileOutputAdaptor

{
  "type": "local_file"
}

Save CSV data chunks to local file storage

Properties

Name Type Required Restrictions Description
type string true Type name for this output type

Enumerated Values

Property Value
type [local_file, localFile]

MonitoringAggregation

{
  "retentionPolicy": "samples",
  "retentionValue": 0
}

Defines the aggregation policy for monitoring jobs.

Properties

Name Type Required Restrictions Description
retentionPolicy string false Monitoring jobs retention policy for aggregation.
retentionValue integer false Amount/percentage of samples to retain.

Enumerated Values

Property Value
retentionPolicy [samples, percentage]

MonitoringColumnsMapping

{
  "actedUponColumn": "string",
  "actualsTimestampColumn": "string",
  "actualsValueColumn": "string",
  "associationIdColumn": "string",
  "customMetricId": "string",
  "customMetricTimestampColumn": "string",
  "customMetricTimestampFormat": "string",
  "customMetricValueColumn": "string",
  "monitoredStatusColumn": "string",
  "predictionsColumns": [
    {
      "className": "string",
      "columnName": "string"
    }
  ],
  "reportDrift": true,
  "reportPredictions": true,
  "uniqueRowIdentifierColumns": [
    "string"
  ]
}

Column names mapping for monitoring

Properties

Name Type Required Restrictions Description
actedUponColumn string false Name of column that contains value for acted_on.
actualsTimestampColumn string false Name of column that contains actual timestamps.
actualsValueColumn string false Name of column that contains actuals value.
associationIdColumn string false Name of column that contains association Id.
customMetricId string false Id of custom metric to process values for.
customMetricTimestampColumn string false Name of column that contains custom metric values timestamps.
customMetricTimestampFormat string false Format of timestamps from customMetricTimestampColumn.
customMetricValueColumn string false Name of column that contains values for custom metric.
monitoredStatusColumn string false Column name used to mark monitored rows.
predictionsColumns any false Name of the column(s) which contain prediction values.

oneOf

Name Type Required Restrictions Description
» anonymous [PredictionColumMap] false maxItems: 100
Map containing column name(s) and class name(s) for multiclass problem.

xor

Name Type Required Restrictions Description
» anonymous string false Column name that contains the prediction for regressions problem.

continued

Name Type Required Restrictions Description
reportDrift boolean false True to report drift, False otherwise.
reportPredictions boolean false True to report prediction, False otherwise.
uniqueRowIdentifierColumns [string] false maxItems: 100
Column(s) name of unique row identifiers.

MonitoringOutputSettings

{
  "monitoredStatusColumn": "string",
  "uniqueRowIdentifierColumns": [
    "string"
  ]
}

Output settings for monitoring jobs

Properties

Name Type Required Restrictions Description
monitoredStatusColumn string true Column name used to mark monitored rows.
uniqueRowIdentifierColumns [string] true maxItems: 100
Column(s) name of unique row identifiers.

OverallConfigUpdate

{
  "timeoutAction": "block",
  "timeoutSec": 2
}

Overall moderation configuration to push (not specific to one guard)

Properties

Name Type Required Restrictions Description
timeoutAction string true Action to take if timeout occurs
timeoutSec integer true minimum: 2
Timeout value in seconds for any guard

Enumerated Values

Property Value
timeoutAction [block, score]

OverallModerationConfigurationResponse

{
  "entityId": "string",
  "entityType": "customModel",
  "timeoutAction": "block",
  "timeoutSec": 2,
  "updatedAt": "2019-08-24T14:15:22Z",
  "updaterId": "string"
}

Properties

Name Type Required Restrictions Description
entityId string true ID of custom model or playground for this configuration.
entityType string true Type of associated entity.
timeoutAction string true Action to take if timeout occurs
timeoutSec integer true minimum: 2
Timeout value in seconds for any guard
updatedAt string(date-time) false When the configuration was updated.
updaterId string¦null true ID of user who updated the configuration.

Enumerated Values

Property Value
entityType [customModel, customModelVersion, playground]
timeoutAction [block, score]

OverallModerationConfigurationUpdate

{
  "entityId": "string",
  "entityType": "customModel",
  "timeoutAction": "block",
  "timeoutSec": 0
}

Properties

Name Type Required Restrictions Description
entityId string true ID of custom model or playground for this configuration.
entityType string true Type of associated entity.
timeoutAction string true Action to take if timeout occurs
timeoutSec integer true minimum: 0
Timeout value in seconds for any guard

Enumerated Values

Property Value
entityType [customModel, customModelVersion, playground]
timeoutAction [block, score]

PPSImageMetadataResponse

{
  "baseImageId": "string",
  "created": "2019-08-24T14:15:22Z",
  "datarobotRuntimeImageTag": "string",
  "dockerImageId": "string",
  "filename": "string",
  "hash": "string",
  "hashAlgorithm": "SHA256",
  "imageSize": 0,
  "shortDockerImageId": "string"
}

Properties

Name Type Required Restrictions Description
baseImageId string true Internal base image entity id for troubleshooting purposes
created string(date-time) true ISO formatted image upload date
datarobotRuntimeImageTag string¦null false For internal use only.
dockerImageId string true A Docker image id (immutable, content-based) hash associated with the given image
filename string true The name of the file when the download requested
hash string true Hash of the image content, supposed to be used for verifying content after the download. The algorithm used for hashing is specified in hashAlgorithm field. Note that hash is calculated over compressed image data.
hashAlgorithm string true An algorithm name used for calculating content hash
imageSize integer true Size in bytes of the compressed PPS image data
shortDockerImageId string true A 12-chars shortened version of the dockerImageId as shown in 'docker images' command line command output

Enumerated Values

Property Value
hashAlgorithm SHA256

PredictionColumMap

{
  "className": "string",
  "columnName": "string"
}

Properties

Name Type Required Restrictions Description
className string true Class name.
columnName string true Column name that contains the prediction for a specific class.

PredictionEnvironmentInUseResponse

{
  "id": "string",
  "name": "string",
  "usedBy": [
    {
      "configurationId": "string",
      "deploymentId": "string",
      "name": "string"
    }
  ]
}

Properties

Name Type Required Restrictions Description
id string true ID of prediction environment.
name string true Name of prediction environment.
usedBy [DeploymentAndGuardResponse] true maxItems: 32
Guards using this prediction environment.

S3DataStreamer

{
  "credentialId": "string",
  "endpointUrl": "string",
  "format": "csv",
  "type": "s3",
  "url": "string"
}

Stream CSV data chunks from Amazon Cloud Storage S3

Properties

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false Use the specified credential to access the url

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
endpointUrl string(url) false Endpoint URL for the S3 connection (omit to use the default)
format string false Type of input file format
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
anonymous [redacted]
format [csv, parquet]
type s3

S3Intake

{
  "credentialId": "string",
  "endpointUrl": "string",
  "format": "csv",
  "type": "s3",
  "url": "string"
}

Stream CSV data chunks from Amazon Cloud Storage S3

Properties

Name Type Required Restrictions Description
credentialId string¦null false Use the specified credential to access the url
endpointUrl string(url) false Endpoint URL for the S3 connection (omit to use the default)
format string false Type of input file format
type string true Type name for this intake type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
format [csv, parquet]
type s3

S3Output

{
  "credentialId": "string",
  "endpointUrl": "string",
  "format": "csv",
  "partitionColumns": [
    "string"
  ],
  "serverSideEncryption": {
    "algorithm": "string",
    "customerAlgorithm": "string",
    "customerKey": "string",
    "kmsEncryptionContext": "string",
    "kmsKeyId": "string"
  },
  "type": "s3",
  "url": "string"
}

Saves CSV data chunks to Amazon Cloud Storage S3

Properties

Name Type Required Restrictions Description
credentialId string¦null false Use the specified credential to access the url
endpointUrl string(url) false Endpoint URL for the S3 connection (omit to use the default)
format string false Type of output file format
partitionColumns [string] false maxItems: 100
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/").
serverSideEncryption ServerSideEncryption false Configure Server-Side Encryption for S3 output
type string true Type name for this output type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
format [csv, parquet]
type s3

S3OutputAdaptor

{
  "credentialId": "string",
  "endpointUrl": "string",
  "format": "csv",
  "partitionColumns": [
    "string"
  ],
  "serverSideEncryption": {
    "algorithm": "string",
    "customerAlgorithm": "string",
    "customerKey": "string",
    "kmsEncryptionContext": "string",
    "kmsKeyId": "string"
  },
  "type": "s3",
  "url": "string"
}

Saves CSV data chunks to Amazon Cloud Storage S3

Properties

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false Use the specified credential to access the url

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
endpointUrl string(url) false Endpoint URL for the S3 connection (omit to use the default)
format string false Type of output file format
partitionColumns [string] false maxItems: 100
For Parquet directory-scoring only. The column names of the intake data of which to partition the dataset. Columns are partitioned in the order they are given. At least one value is required if scoring to a directory (meaning the output url ends with a slash ("/").
serverSideEncryption ServerSideEncryption false Configure Server-Side Encryption for S3 output
type string true Type name for this output type
url string(url) true URL for the CSV file

Enumerated Values

Property Value
anonymous [redacted]
format [csv, parquet]
type s3

Schedule

{
  "dayOfMonth": [
    "*"
  ],
  "dayOfWeek": [
    "*"
  ],
  "hour": [
    "*"
  ],
  "minute": [
    "*"
  ],
  "month": [
    "*"
  ]
}

The scheduling information defining how often and when to execute this job to the Job Scheduling service. Optional if enabled = False.

Properties

Name Type Required Restrictions Description
dayOfMonth [anyOf] true The date(s) of the month that the job will run. Allowed values are either [1 ... 31] or ["*"] for all days of the month. This field is additive with dayOfWeek, meaning the job will run both on the date(s) defined in this field and the day specified by dayOfWeek (for example, dates 1st, 2nd, 3rd, plus every Tuesday). If dayOfMonth is set to ["*"] and dayOfWeek is defined, the scheduler will trigger on every day of the month that matches dayOfWeek (for example, Tuesday the 2nd, 9th, 16th, 23rd, 30th). Invalid dates such as February 31st are ignored.

anyOf

Name Type Required Restrictions Description
» anonymous number false none

or

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dayOfWeek [anyOf] true The day(s) of the week that the job will run. Allowed values are [0 .. 6], where (Sunday=0), or ["*"], for all days of the week. Strings, either 3-letter abbreviations or the full name of the day, can be used interchangeably (e.g., "sunday", "Sunday", "sun", or "Sun", all map to [0]. This field is additive with dayOfMonth, meaning the job will run both on the date specified by dayOfMonth and the day defined in this field.

anyOf

Name Type Required Restrictions Description
» anonymous number false none

or

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
hour [anyOf] true The hour(s) of the day that the job will run. Allowed values are either ["*"] meaning every hour of the day or [0 ... 23].

anyOf

Name Type Required Restrictions Description
» anonymous number false none

or

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
minute [anyOf] true The minute(s) of the day that the job will run. Allowed values are either ["*"] meaning every minute of the day or[0 ... 59].

anyOf

Name Type Required Restrictions Description
» anonymous number false none

or

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
month [anyOf] true The month(s) of the year that the job will run. Allowed values are either [1 ... 12] or ["*"] for all months of the year. Strings, either 3-letter abbreviations or the full name of the month, can be used interchangeably (e.g., "jan" or "october"). Months that are not compatible with dayOfMonth are ignored, for example {"dayOfMonth": [31], "month":["feb"]}.

anyOf

Name Type Required Restrictions Description
» anonymous number false none

or

Name Type Required Restrictions Description
» anonymous string false none

ServerSideEncryption

{
  "algorithm": "string",
  "customerAlgorithm": "string",
  "customerKey": "string",
  "kmsEncryptionContext": "string",
  "kmsKeyId": "string"
}

Configure Server-Side Encryption for S3 output

Properties

Name Type Required Restrictions Description
algorithm string false The server-side encryption algorithm used when storing this object in Amazon S3 (for example, AES256, aws:kms).
customerAlgorithm string false Specifies the algorithm to use to when encrypting the object (for example, AES256).
customerKey string false Specifies the customer-provided encryption key for Amazon S3 to use in encrypting data. This value is used to store the object and then it is discarded; Amazon S3 does not store the encryption key. The key must be appropriate for use with the algorithm specified in customerAlgorithm. The key must be sent as an base64 encoded string.
kmsEncryptionContext string false Specifies the Amazon Web Services KMS Encryption Context to use for object encryption. The value of this header is a base64-encoded UTF-8 string holding JSON with the encryption context key-value pairs.
kmsKeyId string false Specifies the ID of the symmetric customer managed key to use for object encryption.

SnowflakeDataStreamer

{
  "catalog": "string",
  "cloudStorageCredentialId": "string",
  "cloudStorageType": "azure",
  "credentialId": "string",
  "dataStoreId": "string",
  "externalStage": "string",
  "query": "string",
  "schema": "string",
  "table": "string",
  "type": "snowflake"
}

Stream CSV data chunks from Snowflake

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to read input data from.
cloudStorageCredentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false The ID of the credential holding information about a user with read access to the cloud storage.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
cloudStorageType string false Type name for cloud storage
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false The ID of the credential holding information about a user with read access to the Snowflake data source.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataStoreId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false ID of the data store to connect to

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
externalStage string true External storage
query string false A self-supplied SELECT statement of the dataset you wish to score. Helpful for supplying a more fine-grained selection of data not achievable through specification of "table" and/or "schema" parameters exclusively.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
schema string false The name of the specified database schema to read input data from.
table string false The name of the specified database table to read input data from.
type string true Type name for this intake type

Enumerated Values

Property Value
anonymous [redacted]
cloudStorageType [azure, gcp, s3]
anonymous [redacted]
anonymous [redacted]
type snowflake

SnowflakeIntake

{
  "catalog": "string",
  "cloudStorageCredentialId": "string",
  "cloudStorageType": "azure",
  "credentialId": "string",
  "dataStoreId": "string",
  "externalStage": "string",
  "query": "string",
  "schema": "string",
  "table": "string",
  "type": "snowflake"
}

Stream CSV data chunks from Snowflake

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to read input data from.
cloudStorageCredentialId string¦null false The ID of the credential holding information about a user with read access to the cloud storage.
cloudStorageType string false Type name for cloud storage
credentialId string¦null false The ID of the credential holding information about a user with read access to the Snowflake data source.
dataStoreId string true ID of the data store to connect to
externalStage string true External storage
query string false A self-supplied SELECT statement of the dataset you wish to score. Helpful for supplying a more fine-grained selection of data not achievable through specification of "table" and/or "schema" parameters exclusively.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
schema string false The name of the specified database schema to read input data from.
table string false The name of the specified database table to read input data from.
type string true Type name for this intake type

Enumerated Values

Property Value
cloudStorageType [azure, gcp, s3]
type snowflake

SnowflakeOutput

{
  "catalog": "string",
  "cloudStorageCredentialId": "string",
  "cloudStorageType": "azure",
  "createTableIfNotExists": false,
  "credentialId": "string",
  "dataStoreId": "string",
  "externalStage": "string",
  "schema": "string",
  "statementType": "insert",
  "table": "string",
  "type": "snowflake"
}

Save CSV data chunks to Snowflake in bulk

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to write output data to.
cloudStorageCredentialId string¦null false The ID of the credential holding information about a user with write access to the cloud storage.
cloudStorageType string false Type name for cloud storage
createTableIfNotExists boolean false Attempt to create the table first if no existing one is detected, before writing data with the strategy defined in the statementType parameter.
credentialId string¦null false The ID of the credential holding information about a user with write access to the Snowflake data source.
dataStoreId string true ID of the data store to connect to
externalStage string true External storage
schema string false The name of the specified database schema to write results to.
statementType string true The statement type to use when writing the results.
table string true The name of the specified database table to write results to.
type string true Type name for this output type

Enumerated Values

Property Value
cloudStorageType [azure, gcp, s3]
statementType [insert, create_table, createTable]
type snowflake

SnowflakeOutputAdaptor

{
  "catalog": "string",
  "cloudStorageCredentialId": "string",
  "cloudStorageType": "azure",
  "createTableIfNotExists": false,
  "credentialId": "string",
  "dataStoreId": "string",
  "externalStage": "string",
  "schema": "string",
  "statementType": "insert",
  "table": "string",
  "type": "snowflake"
}

Save CSV data chunks to Snowflake in bulk

Properties

Name Type Required Restrictions Description
catalog string false The name of the specified database catalog to write output data to.
cloudStorageCredentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false The ID of the credential holding information about a user with write access to the cloud storage.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
cloudStorageType string false Type name for cloud storage
createTableIfNotExists boolean false Attempt to create the table first if no existing one is detected, before writing data with the strategy defined in the statementType parameter.
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false The ID of the credential holding information about a user with write access to the Snowflake data source.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataStoreId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false ID of the data store to connect to

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
externalStage string true External storage
schema string false The name of the specified database schema to write results to.
statementType string true The statement type to use when writing the results.
table string true The name of the specified database table to write results to.
type string true Type name for this output type

Enumerated Values

Property Value
anonymous [redacted]
cloudStorageType [azure, gcp, s3]
anonymous [redacted]
anonymous [redacted]
statementType [insert, create_table, createTable]
type snowflake

SupportedLlmListResponse

{
  "count": 0,
  "data": [
    {
      "description": "string",
      "id": "string",
      "llmType": "string",
      "model": "string",
      "name": "string",
      "provider": "string",
      "vendor": "string"
    }
  ],
  "next": "http://example.com",
  "previous": "http://example.com",
  "totalCount": 0
}

Properties

Name Type Required Restrictions Description
count integer false Number of items returned on this page.
data [SupportedLlmResponse] true maxItems: 200
List of supported LLMs for moderation.
next string(uri)¦null true URL pointing to the next page (if null, there is no next page).
previous string(uri)¦null true URL pointing to the previous page (if null, there is no previous page).
totalCount integer true The total number of items across all pages.

SupportedLlmResponse

{
  "description": "string",
  "id": "string",
  "llmType": "string",
  "model": "string",
  "name": "string",
  "provider": "string",
  "vendor": "string"
}

Properties

Name Type Required Restrictions Description
description string true maxLength: 1024
Description of this LLM.
id string true maxLength: 1024
Identifier for this LLM.
llmType string true maxLength: 1024
General category of this LLM.
model string true maxLength: 1024
Specific model of this LLM.
name string true maxLength: 1024
Display name of this LLM.
provider string true maxLength: 1024
Provider of access to this LLM.
vendor string true maxLength: 1024
Vendor of this LLM.

SynapseDataStreamer

{
  "cloudStorageCredentialId": "string",
  "credentialId": "string",
  "dataStoreId": "string",
  "externalDataSource": "string",
  "query": "string",
  "schema": "string",
  "table": "string",
  "type": "synapse"
}

Stream CSV data chunks from Azure Synapse

Properties

Name Type Required Restrictions Description
cloudStorageCredentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false The ID of the Azure credential holding information about a user with read access to the cloud storage.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false The ID of the credential holding information about a user with read access to the JDBC data source.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataStoreId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false ID of the data store to connect to

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
externalDataSource string true External datasource name
query string false A self-supplied SELECT statement of the dataset you wish to score. Helpful for supplying a more fine-grained selection of data not achievable through specification of "table" and/or "schema" parameters exclusively.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
schema string false The name of the specified database schema to read input data from.
table string false The name of the specified database table to read input data from.
type string true Type name for this intake type

Enumerated Values

Property Value
anonymous [redacted]
anonymous [redacted]
anonymous [redacted]
type synapse

SynapseIntake

{
  "cloudStorageCredentialId": "string",
  "credentialId": "string",
  "dataStoreId": "string",
  "externalDataSource": "string",
  "query": "string",
  "schema": "string",
  "table": "string",
  "type": "synapse"
}

Stream CSV data chunks from Azure Synapse

Properties

Name Type Required Restrictions Description
cloudStorageCredentialId string¦null false The ID of the Azure credential holding information about a user with read access to the cloud storage.
credentialId string¦null false The ID of the credential holding information about a user with read access to the JDBC data source.
dataStoreId string true ID of the data store to connect to
externalDataSource string true External datasource name
query string false A self-supplied SELECT statement of the dataset you wish to score. Helpful for supplying a more fine-grained selection of data not achievable through specification of "table" and/or "schema" parameters exclusively.If this job is executed with a job definition, then template variables are available which will be substituted for timestamps: {{ current_run_timestamp }}, {{ last_completed_run_time }}, {{ last_scheduled_run_time }}, {{ next_scheduled_run_time }}, {{ current_run_time }}
schema string false The name of the specified database schema to read input data from.
table string false The name of the specified database table to read input data from.
type string true Type name for this intake type

Enumerated Values

Property Value
type synapse

SynapseOutput

{
  "cloudStorageCredentialId": "string",
  "createTableIfNotExists": false,
  "credentialId": "string",
  "dataStoreId": "string",
  "externalDataSource": "string",
  "schema": "string",
  "statementType": "insert",
  "table": "string",
  "type": "synapse"
}

Save CSV data chunks to Azure Synapse in bulk

Properties

Name Type Required Restrictions Description
cloudStorageCredentialId string¦null false The ID of the credential holding information about a user with write access to the cloud storage.
createTableIfNotExists boolean false Attempt to create the table first if no existing one is detected, before writing data with the strategy defined in the statementType parameter.
credentialId string¦null false The ID of the credential holding information about a user with write access to the JDBC data source.
dataStoreId string true ID of the data store to connect to
externalDataSource string true External data source name
schema string false The name of the specified database schema to write results to.
statementType string true The statement type to use when writing the results.
table string true The name of the specified database table to write results to.
type string true Type name for this output type

Enumerated Values

Property Value
statementType [insert, create_table, createTable]
type synapse

SynapseOutputAdaptor

{
  "cloudStorageCredentialId": "string",
  "createTableIfNotExists": false,
  "credentialId": "string",
  "dataStoreId": "string",
  "externalDataSource": "string",
  "schema": "string",
  "statementType": "insert",
  "table": "string",
  "type": "synapse"
}

Save CSV data chunks to Azure Synapse in bulk

Properties

Name Type Required Restrictions Description
cloudStorageCredentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false The ID of the credential holding information about a user with write access to the cloud storage.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
createTableIfNotExists boolean false Attempt to create the table first if no existing one is detected, before writing data with the strategy defined in the statementType parameter.
credentialId any false Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string¦null false The ID of the credential holding information about a user with write access to the JDBC data source.

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
dataStoreId any true Either the populated value of the field or [redacted] due to permission settings

oneOf

Name Type Required Restrictions Description
» anonymous string false ID of the data store to connect to

xor

Name Type Required Restrictions Description
» anonymous string false none

continued

Name Type Required Restrictions Description
externalDataSource string true External data source name
schema string false The name of the specified database schema to write results to.
statementType string true The statement type to use when writing the results.
table string true The name of the specified database table to write results to.
type string true Type name for this output type

Enumerated Values

Property Value
anonymous [redacted]
anonymous [redacted]
anonymous [redacted]
statementType [insert, create_table, createTable]
type synapse

Updated November 23, 2024