Skip to content

On-premise users: click in-app to access the full platform documentation for your version of DataRobot.

Create custom inference models

Custom inference models are user-created, pre-trained models that you can upload to DataRobot (as a collection of files) via the Custom Model Workshop. You can then upload a model artifact to create, test, and deploy custom inference models to DataRobot's centralized deployment hub.

You can assemble custom inference models in either of the following ways:

  • Create a custom model without providing the model requirements and start_server.sh file on the Assemble tab. This type of custom model must use a drop-in environment. Drop-in environments contain the requirements and start_server.sh file used by the model. They are provided by DataRobot in the Custom Model Workshop. You can also create your own drop-in custom environment.

  • Create a custom model with the model requirements and start_server.sh file on the Assemble tab. This type of custom model can be paired with a custom or drop-in environment.

Be sure to review the guidelines for assembling a custom model before proceeding. If any files overlap between the custom model and the environment folders, the model's files will take priority.

Note

Once a custom model's file contents are assembled, you can test the contents locally for development purposes before uploading it to DataRobot. After you create a custom model in the Workshop, you can run a testing suite from the Assemble tab.

Create a new custom model

  1. To create a custom model, navigate to Model Registry > Custom Model Workshop and select the Models tab. This tab lists the models you have created. Click Add new model.

  2. In the Add Custom Inference Model window, enter the fields described in the table below:

    Element Description
    1 Model name Name the custom model.
    2 Target type / Target name Select the target type (binary classification, regression, multiclass, text generation (premium feature), anomaly detection, or unstructured) and enter the name of the target feature.
    3 Positive class label / Negative class label These fields only display for binary classification models. Specify the value to be used as the positive class label and the value to be used as the negative class label.
    For a multiclass classification model, these fields are replaced by a field to enter or upload the target classes in .csv or .txt format.
  3. Click Show Optional Fields and, if necessary, enter a prediction threshold, the coding language used to build the model, and a description.

  4. After completing the fields, click Add Custom Model.

  5. In the Assemble tab, under Model Environment on the right, select a model environment by clicking the Base Environment dropdown menu on the right and selecting an environment. The model environment is used for testing and deploying the custom model.

    Note

    The Base Environment pulldown menu includes drop-in model environments, if any exist, as well as custom environments that you can create.

  6. Under Model on the left, add content by dragging and dropping files or browsing. Alternatively, select a remote integrated repository.

    If you click Browse local file, you have the option of adding a Local Folder. The local folder is for dependent files and additional assets required by your model, not the model itself. Even if the model file is included in the folder, it will not be accessible to DataRobot unless the file exists at the root level. The root file can then point to the dependencies in the folder.

    Note

    You must also upload the model requirements and a start_server.sh file to your model's folder unless you are pairing the model with a drop-in environment.

Anomaly detection

You can create custom inference models that support anomaly detection problems. If you choose to build one, reference the DRUM template. (Log in to GitHub before clicking this link.) When deploying custom inference anomaly detection models, note that the following functionality is not supported:

  • Data drift
  • Accuracy and association IDs
  • Challenger models
  • Humility rules
  • Prediction intervals

Updated February 15, 2024