# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/catalogItems/?offset=0&limit=0&initialCacheSize=500\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
offset
query
integer
true
Specifies the number of results to skip for pagination.
limit
query
integer
true
Sets the maximum number of results returned. Enter 0 to specify no limit.
initialCacheSize
query
integer
true
The initial cache size, for Mongo search only.
useCache
query
string
false
Sets whether to use the cache, for Mongo search only.
orderBy
query
string
false
The attribute sort order applied to the returned catalog list: 'catalogName', 'originalName', 'description', 'created', or 'relevance'. For all options other than 'relevance', prefix the attribute name with a dash to sort in descending order. e.g., orderBy='-catalogName'. Defaults to '-created'.
searchFor
query
string
false
A value to search for in the dataset's name, description, tags, column names, categories, and latest errors. The search is case insensitive. If no value is provided, or if the empty string is used, or if the string contains only whitespace, no filtering occurs. Partial matching is performed on the dataset name and description fields; all other fields require an exact match.
tag
query
any
false
Filter results to display only items with the specified catalog item tags, in lower case, with no spaces.
accessType
query
string
false
Access type used to filter returned results. Valid options are 'owner', 'shared', 'created', and 'any' (the default): 'owner' items are owned by the requester, 'shared' items have been shared with the requester, 'created' items have been created by the requester, and 'any' items matches all.
datasourceType
query
any
false
Data source types used for filtering.
category
query
any
false
Category type(s) used for filtering. Searches are case sensitive and support '&' and 'OR' operators.
filterFailed
query
string
false
Sets whether to exclude from the search results all catalog items that failed during import. If True, invalid catalog items will be excluded; default is False.
ownerUserId
query
any
false
Filter results to display only those owned by user(s) identified by the specified UID.
ownerUsername
query
any
false
Filter results to display only those owned by user(s) identified by the specified username.
type
query
string
false
Filter results by catalog type. The 'dataset' option matches both 'snapshot_dataset' and 'remote_dataset'.
isUxrPreviewable
query
boolean
false
Filter results by catalogType = 'snapshot_dataset' and catalogType = 'remote_dataset' and data_origin in ['snowflake', 'bigquery-v1']
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/catalogItems/{catalogId}/\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
# You can also use wget
curl-XPATCHhttps://app.datarobot.com/api/v2/catalogItems/{catalogId}/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/dataEngineQueryGenerators/\-H"Content-Type: application/json"\-H"Authorization: Bearer {access-token}"\-d'{CreateDataEngineQueryGenerator}'
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/dataEngineQueryGenerators/{dataEngineQueryGeneratorId}/\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/dataEngineWorkspaceStates/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{CreateWorkspaceState}'
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/dataEngineWorkspaceStates/fromDataEngineQueryGenerator/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{CreateWorkspaceStateFromQueryGenerator}'
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/dataEngineWorkspaceStates/{workspaceStateId}/\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/?limit=100&offset=0\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
category
query
string
false
If specified, only dataset versions that have the specified category will be included in the results. Categories identify the intended use of the dataset.
orderBy
query
string
false
Sorting order which will be applied to catalog list.
limit
query
integer
true
At most this many results are returned.
offset
query
integer
true
This many results will be skipped.
filterFailed
query
string
false
Whether datasets that failed during import should be excluded from the results. If True invalid datasets will be excluded.
datasetVersionIds
query
any
false
If specified will only return datasets that are associated with specified dataset versions. Cannot be used as the same time with experiment_container_ids.
# You can also use wget
curl-XPATCHhttps://app.datarobot.com/api/v2/datasets/\-H"Content-Type: application/json"\-H"Authorization: Bearer {access-token}"\-d'{BulkDatasetAction}'
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/fromDataEngineWorkspaceState/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{DatasetCreateFromWorkspaceState}'
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/fromDataSource/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{Datasource}'
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/fromFile/\-H"Content-Type: multipart/form-data"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{DatasetFromFile}'
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/fromHDFS/\-H"Content-Type: application/json"\-H"Authorization: Bearer {access-token}"\-d'{Hdfs}'
Create a dataset item and version from a recipe.During publishing, an immutable copy of the recipe is created, as well as a copy of the recipe's data source.
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/fromRecipe/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{CreateFromRecipe}'
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/fromStage/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{DatasetFromStage}'
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/fromURL/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{Url}'
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
# You can also use wget
curl-XPATCHhttps://app.datarobot.com/api/v2/datasets/{datasetId}/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/accessControl/?offset=0&limit=100\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
userId
query
string
false
Only return the access control information for a user with this user ID.
username
query
string
false
Only return the access control information for a user with this username.
# You can also use wget
curl-XPATCHhttps://app.datarobot.com/api/v2/datasets/{datasetId}/accessControl/\-H"Content-Type: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
Duplicate entry for a user in permission list or the request would leave the dataset without an owner.
None
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/datasets/{datasetId}/allFeaturesDetails/¶
Return detailed information on all the features and transforms for this dataset.If the Dataset Item has attribute snapshot = True, all optional fields also appear
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/allFeaturesDetails/?limit=100&offset=0&orderBy=featureType\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
limit
query
integer
true
At most this many results are returned. The default may change and a maximum limit may be imposed without notice.
offset
query
integer
true
This many results will be skipped.
orderBy
query
string
true
How the features should be ordered.
includePlot
query
string
false
Include histogram plot data in the response.
searchFor
query
string
false
A value to search for in the feature name. The search is case insensitive. If no value is provided, an empty string is used, or the string contains only whitespace, no filtering occurs.
featurelistId
query
string
false
ID of a featurelist. If specified, only returns features that are present in the specified featurelist.
includeDataQuality
query
string
false
Include detected data quality issue types in the response.
# You can also use wget
curl-XPATCHhttps://app.datarobot.com/api/v2/datasets/{datasetId}/deleted/\-H"Content-Type: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/featureHistograms/{featureName}/?binLimit=60\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
binLimit
query
integer
true
Maximum number of bins in the returned plot.
key
query
string
false
Only required for the Summarized categorical feature. Name of the top 50 key for which plot to be retrieved.
usePlot2
query
string
false
Use frequent values plot data instead of histogram for supported feature types.
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/featureTransforms/?limit=100&offset=0\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
limit
query
integer
true
At most this many results are returned. The default may change and a maximum limit may be imposed without notice.
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/{datasetId}/featureTransforms/\-H"Content-Type: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/featureTransforms/{featureName}/\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
datasetId
path
string
true
The dataset to select feature from.
featureName
path
string
true
The name of the feature. Note that DataRobot renames some features, so the feature name may not be the one from your original data. Non-ascii features names should be utf-8-encoded (before URL-quoting).
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/featurelists/?limit=100&offset=0&orderBy=name\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
limit
query
integer
true
At most this many results are returned. The default may change and a maximum limit may be imposed without notice.
offset
query
integer
true
This many results will be skipped.
orderBy
query
string
true
How the feature lists should be ordered.
searchFor
query
string
false
A value to search for in the featurelist name. The search is case insensitive. If no value is provided, an empty string is used, or the string contains only whitespace, no filtering occurs.
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/{datasetId}/featurelists/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
# You can also use wget
curl-XDELETEhttps://app.datarobot.com/api/v2/datasets/{datasetId}/featurelists/{featurelistId}/\-H"Authorization: Bearer {access-token}"
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/featurelists/{featurelistId}/\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
# You can also use wget
curl-XPATCHhttps://app.datarobot.com/api/v2/datasets/{datasetId}/featurelists/{featurelistId}/\-H"Content-Type: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/file/\-H"Accept: application/vnd.apache.parquet"\-H"Authorization: Bearer {access-token}"
Dataset cannot be downloaded. Possible reasons include "dataPersisted" being false for the dataset, the dataset not being a snapshot, and this dataset is too big to be downloaded (maximum download size depends on a config of your installation).
None
To perform this operation, you must be authenticated by means of one of the following methods:
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/permissions/\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/projects/?limit=100&offset=0\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/refreshJobs/?limit=100&offset=0\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Create a dataset refresh job that will automatically create dataset snapshots on a schedule.
Optionally if the limit of enabled jobs per user is reached the following metadata will be added
to the default error response payload:
datasetsWithJob (array) - The list of datasets IDs that have at least one enabled job.
errorType (string) - (New in version v2.21) The type of error that happened, possible values include (but are not limited to): Generic Limit Reached, Max Job Limit Reached for Dataset, and Max Job Limit Reached for User.
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/{datasetId}/refreshJobs/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
Refresh job could not be created. Possible reasons include, the job does not belong to the given dataset, credential ID required when Kerberos authentication enabled, or the schedule is not valid or cannot be understood.
None
To perform this operation, you must be authenticated by means of one of the following methods:
# You can also use wget
curl-XDELETEhttps://app.datarobot.com/api/v2/datasets/{datasetId}/refreshJobs/{jobId}/\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
datasetId
path
string
true
The dataset associated with the scheduled refresh job.
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/refreshJobs/{jobId}/\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
datasetId
path
string
true
The dataset associated with the scheduled refresh job.
Optionally if the limit of enabled jobs per user is reached the following metadata will be added
to the default error response payload:
datasetsWithJob (array) - The list of datasets IDs that have at least one enabled job.
errorType (string) - (New in version v2.21) The type of error that happened, possible values include (but are not limited to): Generic Limit Reached, Max Job Limit Reached for Dataset, and Max Job Limit Reached for User.
# You can also use wget
curl-XPATCHhttps://app.datarobot.com/api/v2/datasets/{datasetId}/refreshJobs/{jobId}/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
Refresh job could not be updated. Possible reasons include, the job does not belong to the given dataset, credential ID required when Kerberos authentication enabled, or the schedule is not valid or cannot be understood.
None
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/datasets/{datasetId}/refreshJobs/{jobId}/executionResults/¶
Paginated list of execution results for refresh job with the given ID and dataset with the given ID, sorted from newest to oldest.
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/refreshJobs/{jobId}/executionResults/\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
limit
query
integer
false
Maximum number of results returned. The default may change and a maximum limit may be imposed without notice.
offset
query
integer
false
Number of results that will be skipped.
datasetId
path
string
true
The dataset associated with the scheduled refresh job.
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/relationships/?limit=100&offset=0\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
limit
query
integer
true
At most this many results are returned.
offset
query
integer
true
This many results will be skipped.
linkedDatasetId
query
string
false
Providing linkedDatasetId will filter such that only relationships between datasetId (from the path) and linkedDatasetId will be returned.
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/{datasetId}/relationships/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
# You can also use wget
curl-XDELETEhttps://app.datarobot.com/api/v2/datasets/{datasetId}/relationships/{datasetRelationshipId}/\-H"Authorization: Bearer {access-token}"
# You can also use wget
curl-XPATCHhttps://app.datarobot.com/api/v2/datasets/{datasetId}/relationships/{datasetRelationshipId}/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/sharedRoles/?offset=0&limit=100\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
id
query
string
false
Only return the access control information for a organization, group or user with this ID.
name
query
string
false
Only return the access control information for a organization, group or user with this name.
Grant access, remove access or update roles for organizations, groups or users on this dataset. Up to 100 roles may be set per array in a single request.
# You can also use wget
curl-XPATCHhttps://app.datarobot.com/api/v2/datasets/{datasetId}/sharedRoles/\-H"Content-Type: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/?limit=100&offset=0\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
category
query
string
false
If specified, only dataset versions that have the specified category will be included in the results. Categories identify the intended use of the dataset.
orderBy
query
string
false
Sorting order which will be applied to catalog list.
limit
query
integer
true
At most this many results are returned.
offset
query
integer
true
This many results will be skipped.
filterFailed
query
string
false
Whether datasets that failed during import should be excluded from the results. If True invalid datasets will be excluded.
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/datasets/{datasetId}/versions/fromDataEngineWorkspaceState/¶
Create a new dataset version for a specified dataset from a Data Engine workspace state. The new dataset version should have the same schema as the specified dataset.
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/fromDataEngineWorkspaceState/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/datasets/{datasetId}/versions/fromDataSource/¶
Create a new version for the specified dataset from specified Data Source. The dataset must have been created from a compatible data source originally.
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/fromDataSource/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/fromFile/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/fromHDFS/\-H"Content-Type: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/datasets/{datasetId}/versions/fromLatestVersion/¶
Create a new version of the specified dataset from the latest dataset version. This will reuse the same source of the data that was previously used. Not supported for datasets that were previously loaded from an uploaded file. If the dataset is currently a remote dataset, it will be converted to a snapshot dataset.
NOTE:
if the current version uses a Data Source, the user and password must be specified so the data can be accessed.
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/fromLatestVersion/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/fromStage/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
The request cannot be processed. Possible reasons include the request did not contain data stage, dataset was previously created from a non data stage source
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/fromURL/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
# You can also use wget
curl-XDELETEhttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/{datasetVersionId}/\-H"Authorization: Bearer {access-token}"
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/{datasetVersionId}/\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
GET /api/v2/datasets/{datasetId}/versions/{datasetVersionId}/allFeaturesDetails/¶
Return detailed information on all the features and transforms for this dataset.If the Dataset Item has attribute snapshot = True, all optional fields also appear
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/{datasetVersionId}/allFeaturesDetails/?limit=100&offset=0&orderBy=featureType\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
limit
query
integer
true
At most this many results are returned. The default may change and a maximum limit may be imposed without notice.
offset
query
integer
true
This many results will be skipped.
orderBy
query
string
true
How the features should be ordered.
includePlot
query
string
false
Include histogram plot data in the response.
searchFor
query
string
false
A value to search for in the feature name. The search is case insensitive. If no value is provided, an empty string is used, or the string contains only whitespace, no filtering occurs.
featurelistId
query
string
false
ID of a featurelist. If specified, only returns features that are present in the specified featurelist.
includeDataQuality
query
string
false
Include detected data quality issue types in the response.
# You can also use wget
curl-XPATCHhttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/{datasetVersionId}/deleted/\-H"Content-Type: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/{datasetVersionId}/featureHistograms/{featureName}/?binLimit=60\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
binLimit
query
integer
true
Maximum number of bins in the returned plot.
key
query
string
false
Only required for the Summarized categorical feature. Name of the top 50 key for which plot to be retrieved.
usePlot2
query
string
false
Use frequent values plot data instead of histogram for supported feature types.
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/{datasetVersionId}/featurelists/?limit=100&offset=0&orderBy=name\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
limit
query
integer
true
At most this many results are returned. The default may change and a maximum limit may be imposed without notice.
offset
query
integer
true
This many results will be skipped.
orderBy
query
string
true
How the feature lists should be ordered.
searchFor
query
string
false
A value to search for in the featurelist name. The search is case insensitive. If no value is provided, an empty string is used, or the string contains only whitespace, no filtering occurs.
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/{datasetVersionId}/featurelists/{featurelistId}/\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
Parameters
Name
In
Type
Required
Description
datasetId
path
string
true
The ID of the dataset to retrieve featurelist for.
datasetVersionId
path
string
true
The ID of the dataset version to retrieve featurelists for.
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/{datasetVersionId}/file/\-H"Accept: application/vnd.apache.parquet"\-H"Authorization: Bearer {access-token}"
Dataset version cannot be downloaded. Possible reasons include dataPersisted being false for the dataset, the dataset not being a snapshot, and this dataset version is too big to be downloaded (maximum download size depends on a config of your installation).
None
To perform this operation, you must be authenticated by means of one of the following methods:
BearerAuth
POST /api/v2/datasets/{datasetId}/versions/{datasetVersionId}/fromVersion/¶
Create a new version of the specified dataset from the specified dataset version. This will reuse the same source of the data that was previously used. Not supported for datasets that were previously loaded from an uploaded file. If the dataset is currently a remote dataset, it will be converted to a snapshot dataset.
NOTE:
If the specified version uses a Data Source, the user and password must be specified so the data can be accessed.
# You can also use wget
curl-XPOSThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/{datasetVersionId}/fromVersion/\-H"Content-Type: application/json"\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"\-d'{undefined}'
# You can also use wget
curl-XGEThttps://app.datarobot.com/api/v2/datasets/{datasetId}/versions/{datasetVersionId}/projects/?limit=100&offset=0\-H"Accept: application/json"\-H"Authorization: Bearer {access-token}"
An array of strings describing the intended use of the dataset.
createdBy
string¦null
true
Username of the user who created the dataset.
creationDate
string(date-time)
true
The date when the dataset was created.
dataPersisted
boolean
true
If true, user is allowed to view extended data profile (which includes data statistics like min/max/median/mean, histogram, etc.) and download data. If false, download is not allowed and only the data schema (feature names and types) will be available.
datasetId
string
true
The ID of this dataset.
isDataEngineEligible
boolean
true
Whether this dataset can be a data source of a data engine query.
isLatestVersion
boolean
true
Whether this dataset version is the latest version of this dataset.
isSnapshot
boolean
true
Whether the dataset is an immutable snapshot of data which has previously been retrieved and saved to DataRobot.
An array of strings describing the intended use of the dataset.
columnCount
integer
true
The number of columns in the dataset.
createdBy
string¦null
true
Username of the user who created the dataset.
creationDate
string(date-time)
true
The date when the dataset was created.
dataPersisted
boolean
true
If true, user is allowed to view extended data profile (which includes data statistics like min/max/median/mean, histogram, etc.) and download data. If false, download is not allowed and only the data schema (feature names and types) will be available.
datasetId
string
true
The ID of this dataset.
datasetSize
integer
true
The size of the dataset as a CSV in bytes.
isDataEngineEligible
boolean
true
Whether this dataset can be a data source of a data engine query.
isLatestVersion
boolean
true
Whether this dataset version is the latest version of this dataset.
isSnapshot
boolean
true
Whether the dataset is an immutable snapshot of data which has previously been retrieved and saved to DataRobot.
The action to execute on the datasets. Has to be 'updateRoles' for this payload.
applyGrantToLinkedObjects
boolean
false
If true for any users being granted access to the dataset, grant the user read access to any linked objects such as DataSources and DataStores that may be used by this dataset. Ignored if no such objects are relevant for dataset. Will not result in access being lowered for a user if the user already has higher access to linked objects than read access. However, if the target user does not have sharing permissions to the linked object, they will be given sharing access without lowering existing permissions. May result in an error if user making call does not have sufficient permissions to complete grant. Default value is false.
roles
[oneOf]
true
maxItems: 100 minItems: 1
An array of RoleRequest objects. May contain at most 100 such objects.
The ID of the set of credentials to authenticate with the database.
doSnapshot
boolean
false
If true, create a snapshot dataset; if false, create a remote dataset. Creating snapshots from non-file sources requires an additional permission, Enable Create Snapshot Data Source.
Destination table information to create and materialize the recipe to. If None, the recipe will be materialized in DataRobot.
name
string
false
Name to be assigned to new Dataset.
persistDataAfterIngestion
boolean
false
If true, will enforce saving all data (for download and sampling) and will allow a user to view extended data profile (which includes data statistics like min/max/median/mean, histogram, etc.). If false, will not enforce saving data. The data schema (feature names and types) still will be available. Specifying this parameter to false and doSnapshot to true will result in an error.
recipeId
string
true
The identifier for the Wrangling Recipe to use as the source of data.
skipDuplicateDatesValidation
boolean
false
By default, if a recipe contains time series or a time series resampling operation, publishing fails if there are date duplicates to prevent data quality issues and ambiguous transformations. If set to True, then validation will be skipped.
This property is gated behind the feature flags ['TIME_SERIES_DATA_WRANGLING']. To enable this feature, you can contact your DataRobot representative or administrator.
useKerberos
boolean
false
If true, use kerberos authentication for database authentication.
whether the user should be able to share with other users. If true, the user will be able to grant any role up to and including their own to other users. If role is empty canShare is ignored.
canUseData
boolean
false
Whether the user should be able to view, download and process data (use to create projects, predictions, etc). For OWNER canUseData is always True. If role is empty canUseData is ignored.
role
string
true
the role to grant to the user, or "" (empty string) to remove the users access
username
string
true
username of the user to update the access role for.
If true for any users being granted access to the dataset, grant the user read access to any linked objects such as DataSources and DataStores that may be used by this dataset. Ignored if no such objects are relevant for dataset. Will not result in access being lowered for a user if the user already has higher access to linked objects than read access. However, if the target user does not have sharing permissions to the linked object, they will be given sharing access without lowering existing permissions. May result in an error if user making call does not have sufficient permissions to complete grant. Default value is false.
If true, create a snapshot dataset; if false, create a remote dataset. Creating snapshots from non-file sources requires an additional permission, Enable Create Snapshot Data Source.
workspaceStateId
string
true
The ID of the workspace state to use as the source of data.
If true, create a snapshot dataset; if false, create a remote dataset. Creating snapshots from non-file sources requires an additional permission, Enable Create Snapshot Data Source.
workspaceStateId
string
true
ID of the workspace state to use as the source of data.
The ID of the dataset version the feature belongs to.
dateFormat
string¦null
true
The date format string for how this feature was interpreted (or null if not a date feature). If not null, it will be compatible with https://docs.python.org/2/library/time.html#time.strftime .
featureType
string
true
Feature type.
id
integer
true
The number of the column in the dataset.
isZeroInflated
boolean¦null
false
whether feature has an excessive number of zeros
keySummary
any
false
Per key summaries for Summarized Categorical or Multicategorical columns
The number of rows in the sample used to calculate the statistics.
stdDev
any
false
Standard deviation of EDA sample of the feature.
oneOf
Name
Type
Required
Restrictions
Description
» anonymous
string
false
Standard deviation of EDA sample of the feature.
xor
Name
Type
Required
Restrictions
Description
» anonymous
number
false
Standard deviation of EDA sample of the feature.
continued
Name
Type
Required
Restrictions
Description
timeSeriesEligibilityReason
string¦null
false
why the feature is ineligible for time series projects, or 'suitable' if it is eligible.
timeSeriesEligibilityReasonAggregation
string¦null
false
why the feature is ineligible for aggregation, or 'suitable' if it is eligible.
timeSeriesEligible
boolean
false
whether this feature can be used as a datetime partitioning feature for time series projects. Only sufficiently regular date features can be selected as the datetime feature for time series projects. Always false for non-date features. Date features that cannot be used in datetime partitioning for a time series project may be eligible for an OTV project, which has less stringent requirements.
timeSeriesEligibleAggregation
boolean
false
whether this feature can be used as a datetime feature for aggregationfor time series data prep. Always false for non-date features.
timeStep
integer¦null
false
The minimum time step that can be used to specify time series windows. The units for this value are the timeUnit. When specifying windows for time series projects, all windows must have durations that are integer multiples of this number. Only present for date features that are eligible for time series projects and null otherwise.
timeStepAggregation
integer¦null
false
The minimum time step that can be used to aggregate using this feature for time series data prep. The units for this value are the timeUnit. Only present for date features that are eligible for aggregation in time series data prep and null otherwise.
timeUnit
string¦null
false
The unit for the interval between values of this feature, e.g. DAY, MONTH, HOUR. When specifying windows for time series projects, the windows are expressed in terms of this unit. Only present for date features eligible for time series projects, and null otherwise.
timeUnitAggregation
string¦null
false
The unit for the interval between values of this feature, e.g. DAY, MONTH, HOUR. Only present for date features eligible for aggregation, and null otherwise.
uniqueCount
integer¦null
false
Number of unique values.
upperQuartile
any
false
Upper quartile point of EDA sample of the feature.
oneOf
Name
Type
Required
Restrictions
Description
» anonymous
string
false
Upper quartile point of EDA sample of the feature.
xor
Name
Type
Required
Restrictions
Description
» anonymous
number
false
Upper quartile point of EDA sample of the feature.
The date(s) of the month that the job will run. Allowed values are either [1 ... 31] or ["*"] for all days of the month. This field is additive with dayOfWeek, meaning the job will run both on the date(s) defined in this field and the day specified by dayOfWeek (for example, dates 1st, 2nd, 3rd, plus every Tuesday). If dayOfMonth is set to ["*"] and dayOfWeek is defined, the scheduler will trigger on every day of the month that matches dayOfWeek (for example, Tuesday the 2nd, 9th, 16th, 23rd, 30th). Invalid dates such as February 31st are ignored.
anyOf
Name
Type
Required
Restrictions
Description
» anonymous
number
false
none
or
Name
Type
Required
Restrictions
Description
» anonymous
string
false
none
continued
Name
Type
Required
Restrictions
Description
dayOfWeek
[anyOf]
false
maxItems: 7
The day(s) of the week that the job will run. Allowed values are [0 .. 6], where (Sunday=0), or ["*"], for all days of the week. Strings, either 3-letter abbreviations or the full name of the day, can be used interchangeably (e.g., "sunday", "Sunday", "sun", or "Sun", all map to [0]. This field is additive with dayOfMonth, meaning the job will run both on the date specified by dayOfMonth and the day defined in this field.
anyOf
Name
Type
Required
Restrictions
Description
» anonymous
number
false
none
or
Name
Type
Required
Restrictions
Description
» anonymous
string
false
none
continued
Name
Type
Required
Restrictions
Description
hour
any
false
The hour(s) of the day that the job will run. Allowed values are [0 ... 23].
oneOf
Name
Type
Required
Restrictions
Description
» anonymous
string
false
none
xor
Name
Type
Required
Restrictions
Description
» anonymous
[string]
false
none
continued
Name
Type
Required
Restrictions
Description
minute
any
false
The minute(s) of the day that the job will run. Allowed values are [0 ... 59].
oneOf
Name
Type
Required
Restrictions
Description
» anonymous
string
false
none
xor
Name
Type
Required
Restrictions
Description
» anonymous
[string]
false
none
continued
Name
Type
Required
Restrictions
Description
month
[anyOf]
false
maxItems: 12
The month(s) of the year that the job will run. Allowed values are either [1 ... 12] or ["*"] for all months of the year. Strings, either 3-letter abbreviations or the full name of the month, can be used interchangeably (e.g., "jan" or "october"). Months that are not compatible with dayOfMonth are ignored, for example {"dayOfMonth": [31], "month":["feb"]}.
An array of strings describing the intended use of the dataset. The supported options are TRAINING, and PREDICTION.
oneOf
Name
Type
Required
Restrictions
Description
» anonymous
string
false
none
xor
Name
Type
Required
Restrictions
Description
» anonymous
[string]
false
none
continued
Name
Type
Required
Restrictions
Description
credentialId
string¦null
false
The ID of the set of credentials to use to run the scheduled job when the Kerberos authentication service is utilized. Required when useKerberos is true.
credentials
string
false
A JSON string describing the data engine queries credentials to use when refreshing.
enabled
boolean
false
Boolean for whether the scheduled job is active (true) or inactive (false).
Schedule describing when to refresh the dataset. The smallest schedule allowed is daily.
scheduleReferenceDate
string(date-time)
false
The UTC reference date in RFC-3339 format of when the schedule starts from. This value is returned in /api/v2/datasets/(datasetId)/refreshJobs/(jobId)/ to help build a more intuitive schedule picker. The default is the current time.
useKerberos
boolean
false
If true, the Kerberos authentication system is used in conjunction with a credential ID.
An array of strings describing the intended use of the dataset. The supported options are TRAINING, and PREDICTION.
oneOf
Name
Type
Required
Restrictions
Description
» anonymous
string
false
none
xor
Name
Type
Required
Restrictions
Description
» anonymous
[string]
false
none
continued
Name
Type
Required
Restrictions
Description
credentialId
string¦null
false
The ID of the set of credentials to use to run the scheduled job when the Kerberos authentication service is utilized. Required when useKerberos is true.
credentials
string
false
A JSON string describing the data engine queries credentials to use when refreshing.
enabled
boolean
false
Boolean for whether the scheduled job is active (true) or inactive (false).
Schedule describing when to refresh the dataset. The smallest schedule allowed is daily.
scheduleReferenceDate
string(date-time)
false
The UTC reference date in RFC-3339 format of when the schedule starts from. This value is returned in /api/v2/datasets/(datasetId)/refreshJobs/(jobId)/ to help build a more intuitive schedule picker. Required when schedule is being updated. The default is the current time.
useKerberos
boolean
false
If true, the Kerberos authentication system is used in conjunction with a credential ID.
An array of information about scheduled dataset refresh jobs. Results are based on updatedAt value and returned in descending order (latest returned first).
next
string(uri)¦null
true
URL pointing to the next page (if null, there is no next page).
previous
string(uri)¦null
true
URL pointing to the previous page (if null, there is no previous page).
Whether the org/group/user should be able to share with others. If true, the org/group/user will be able to grant any role up to and including their own to other orgs/groups/user. If role is NO_ROLEcanShare is ignored.
canUseData
boolean
false
Whether the user/group/org should be able to view, download and process data (use to create projects, predictions, etc). For OWNER canUseData is always True. If role is empty canUseData is ignored.
id
string
true
The org/group/user ID.
role
string
true
The role of the org/group/user on this dataset or "NO_ROLE" for removing access when used with route to modify access.
Whether the org/group/user should be able to share with others. If true, the org/group/user will be able to grant any role up to and including their own to other orgs/groups/user. If role is NO_ROLEcanShare is ignored.
canUseData
boolean
false
Whether the user/group/org should be able to view, download and process data (use to create projects, predictions, etc). For OWNER canUseData is always True. If role is empty canUseData is ignored.
name
string
true
Name of the user/group/org to update the access role for.
role
string
true
The role of the org/group/user on this dataset or "NO_ROLE" for removing access when used with route to modify access.
If true for any users being granted access to the dataset, grant the user read access to any linked objects such as DataSources and DataStores that may be used by this dataset. Ignored if no such objects are relevant for dataset. Will not result in access being lowered for a user if the user already has higher access to linked objects than read access. However, if the target user does not have sharing permissions to the linked object, they will be given sharing access without lowering existing permissions. May result in an error if user making call does not have sufficient permissions to complete grant. Default value is false.
operation
string
true
The name of the action being taken. The only operation is "updateRoles".
roles
[oneOf]
true
maxItems: 100 minItems: 1
An array of RoleRequest objects. May contain at most 100 such objects.
The ID of the set of credentials to authenticate with the database.
dataSourceId
string
true
The identifier for the DataSource to use as the source of data.
doSnapshot
boolean
false
If true, create a snapshot dataset; if false, create a remote dataset. Creating snapshots from non-file sources requires an additional permission, Enable Create Snapshot Data Source.
password
string
false
The password (in cleartext) for database authentication. The password will be encrypted on the server side in scope of HTTP request and never saved or stored. DEPRECATED: please use credentialId or credentialData instead.
persistDataAfterIngestion
boolean
false
If true, will enforce saving all data (for download and sampling) and will allow a user to view extended data profile (which includes data statistics like min/max/median/mean, histogram, etc.). If false, will not enforce saving data. The data schema (feature names and types) still will be available. Specifying this parameter to false and doSnapshot to true will result in an error.
List of names of features to be included in the new featurelist, all features listed must be part of the universe. All features for this dataset for the request to succeed.
The value to extract from the date column, of these options: [year|yearDay|month|monthDay|week|weekDay]. Required for transformation of a date column. Otherwise must not be provided.
name
string
true
The name of the new feature. Must not be the same as any existing features for this project. Must not contain '/' character.
parentName
string
true
The name of the parent feature.
replacement
any
false
The replacement in case of a failed transformation.
anyOf
Name
Type
Required
Restrictions
Description
» anonymous
string¦null
false
none
or
Name
Type
Required
Restrictions
Description
» anonymous
boolean¦null
false
none
or
Name
Type
Required
Restrictions
Description
» anonymous
number¦null
false
none
or
Name
Type
Required
Restrictions
Description
» anonymous
integer¦null
false
none
continued
Name
Type
Required
Restrictions
Description
variableType
string
true
The type of the new feature. Must be one of text, categorical (Deprecated in version v2.21), numeric, or categoricalInt. See the description of this method for more information.
The ID of the set of credentials to authenticate with the database.
credentials
string
false
A list of credentials to use if this is a Spark dataset that requires credentials.
password
string
false
The password (in cleartext) for database authentication. The password will be encrypted on the server-side HTTP request and never saved or stored. Required only if the previous data source was a data source. DEPRECATED: please use credentialId or credentialData instead.
useKerberos
boolean
false
If true, use Kerberos for database authentication.
useLatestSuccess
boolean
false
If true, use the latest version that was successfully ingested instead of the latest version, which might be in an errored state. If no successful version is present, the latest errored version is used and the operation fails.
user
string
false
The username for database authentication. Required only if the dataset was initially created from a data source. DEPRECATED: please use credentialId or credentialData instead.
The ID of the set of credentials to authenticate with the database.
credentials
string
false
A list of credentials to use if this is a Spark dataset that requires credentials.
password
string
false
The password (in cleartext) for database authentication. The password will be encrypted on the server-side HTTP request and never saved or stored. Required only if the previous data source was a data source. DEPRECATED: please use credentialId or credentialData instead.
useKerberos
boolean
false
If true, use Kerberos for database authentication.
user
string
false
The username for database authentication. Required only if the dataset was initially created from a data source. DEPRECATED: please use credentialId or credentialData instead.
An array of strings describing the intended use of the dataset.
columnCount
integer
true
The number of columns in the dataset.
createdBy
string¦null
true
Username of the user who created the dataset.
creationDate
string(date-time)
true
The date when the dataset was created.
dataEngineQueryId
string¦null
true
ID of the source data engine query.
dataPersisted
boolean
true
If true, user is allowed to view extended data profile (which includes data statistics like min/max/median/mean, histogram, etc.) and download data. If false, download is not allowed and only the data schema (feature names and types) will be available.
dataSourceId
string¦null
true
ID of the datasource used as the source of the dataset.
dataSourceType
string
true
The type of the datasource that was used as the source of the dataset.
datasetId
string
true
The ID of this dataset.
datasetSize
integer
true
The size of the dataset as a CSV in bytes.
description
string¦null
true
The description of the dataset.
eda1ModificationDate
string(date-time)
true
The ISO 8601 formatted date and time when the EDA1 for the dataset was updated.
eda1ModifierFullName
string
true
The user who was the last to update EDA1 for the dataset.
The URI to datasource. For example, file_name.csv, or jdbc:DATA_SOURCE_GIVEN_NAME/SCHEMA.TABLE_NAME, or jdbc:DATA_SOURCE_GIVEN_NAME/<query> for query based datasources, orhttps://s3.amazonaws.com/dr-pr-tst-data/kickcars-sample-200.csv, etc.
versionId
string
true
The object ID of the catalog_version the dataset belongs to.
The Google Cloud Platform (GCP) key. Output is the downloaded JSON resulting from creating a service account User Managed Key (in the IAM & admin > Service accounts section of GCP).Required if googleConfigId/configId is not specified.Cannot include this parameter if googleConfigId/configId is specified.
The date column that will be used as a datetime partition column in time series project.
defaultCategoricalAggregationMethod
string
true
Default aggregation method used for categorical feature.
defaultNumericAggregationMethod
string
true
Default aggregation method used for numeric feature.
defaultTextAggregationMethod
string
false
Default aggregation method used for text feature.
endToSeriesMaxDatetime
boolean
false
A boolean value indicating whether generates post-aggregated series up to series maximum datetime or global maximum datetime.
multiseriesIdColumns
[string]
false
maxItems: 1 minItems: 1
An array with the names of columns identifying the series to which row of the output dataset belongs. Currently, only one multiseries ID column is supported.
startFromSeriesMinDatetime
boolean
false
A boolean value indicating whether post-aggregated series starts from series minimum datetime or global minimum datetime.
target
string
false
The name of target for the output dataset.
timeStep
integer
true
minimum: 0 (exclusive)
Number of time steps for the output dataset.
timeUnit
string
true
Indicates which unit is a basis for time steps of the output dataset.
The Google Cloud Platform (GCP) key. Output is the downloaded JSON resulting from creating a service account User Managed Key (in the IAM & admin > Service accounts section of GCP).Required if googleConfigId/configId is not specified.Cannot include this parameter if googleConfigId/configId is specified.
googleConfigId
string
false
ID of Secure configurations shared by admin. This is deprecated.Please use configId instead. If specified, cannot include gcpKey.
An array of strings describing the intended use of the dataset.
oneOf
Name
Type
Required
Restrictions
Description
» anonymous
string
false
none
xor
Name
Type
Required
Restrictions
Description
» anonymous
[string]
false
none
continued
Name
Type
Required
Restrictions
Description
doSnapshot
boolean
false
If true, create a snapshot dataset; if false, create a remote dataset. Creating snapshots from non-file sources requires an additional permission, Enable Create Snapshot Data Source.
namenodeWebhdfsPort
integer
false
The port of HDFS name node.
password
string
false
The password (in cleartext) for authenticating to HDFS using Kerberos. The password will be encrypted on the server side in scope of HTTP request and never saved or stored.
persistDataAfterIngestion
boolean
false
If true, will enforce saving all data (for download and sampling) and will allow a user to view extended data profile (which includes data statistics like min/max/median/mean, histogram, etc.). If false, will not enforce saving data. The data schema (feature names and types) still will be available. Specifying this parameter to false and doSnapshot to true will result in an error
url
string(uri)
true
The HDFS url to use as the source of data for the dataset being created.
user
string
false
The username for authenticating to HDFS using Kerberos.
An array of strings describing the intended use of the dataset.
oneOf
Name
Type
Required
Restrictions
Description
» anonymous
string
false
none
xor
Name
Type
Required
Restrictions
Description
» anonymous
[string]
false
none
continued
Name
Type
Required
Restrictions
Description
doSnapshot
boolean
false
If true, create a snapshot dataset; if false, create a remote dataset. Creating snapshots from non-file sources requires an additional permission, Enable Create Snapshot Data Source.
persistDataAfterIngestion
boolean
false
If true, will enforce saving all data (for download and sampling) and will allow a user to view extended data profile (which includes data statistics like min/max/median/mean, histogram, etc.). If false, will not enforce saving data. The data schema (feature names and types) still will be available. Specifying this parameter to false and doSnapshot to true will result in an error.