# Feature Discovery

> Feature Discovery - With DataRobot, you can automatically discover and generate new features from
> multiple datasets, without consolidating manually.

This Markdown file sits beside the HTML page at the same path (with a `.md` suffix). It summarizes the topic and lists links for tools and LLM context.

Companion generated at `2026-04-24T16:03:56.542682+00:00` (UTC).

## Primary page

- [Feature Discovery](https://docs.datarobot.com/en/docs/classic-ui/data/transform-data/feature-discovery/index.html): Full documentation for this topic (HTML).

## Sections on this page

- [Feature considerations](https://docs.datarobot.com/en/docs/classic-ui/data/transform-data/feature-discovery/index.html#feature-considerations): In-page section heading.
- [Batch prediction considerations](https://docs.datarobot.com/en/docs/classic-ui/data/transform-data/feature-discovery/index.html#batch-prediction-considerations): In-page section heading.
- [Feature Discovery compatibility](https://docs.datarobot.com/en/docs/classic-ui/data/transform-data/feature-discovery/index.html#feature-discovery-compatibility): In-page section heading.

## Related documentation

- [Classic UI documentation](https://docs.datarobot.com/en/docs/classic-ui/index.html): Linked from this page.
- [Data](https://docs.datarobot.com/en/docs/classic-ui/data/index.html): Linked from this page.
- [Transform data](https://docs.datarobot.com/en/docs/classic-ui/data/transform-data/index.html): Linked from this page.
- [End-to-end Feature Discovery](https://docs.datarobot.com/en/docs/classic-ui/data/transform-data/feature-discovery/enrich-data-using-feature-discovery.html): Linked from this page.
- [Feature Discovery projects](https://docs.datarobot.com/en/docs/classic-ui/data/transform-data/feature-discovery/fd-overview.html): Linked from this page.
- [Snowflake integration](https://docs.datarobot.com/en/docs/classic-ui/data/transform-data/feature-discovery/fd-snowflake.html): Linked from this page.
- [Feature Discovery settings](https://docs.datarobot.com/en/docs/classic-ui/data/transform-data/feature-discovery/fd-adv-opt.html): Linked from this page.
- [Time-aware feature engineering](https://docs.datarobot.com/en/docs/classic-ui/data/transform-data/feature-discovery/fd-time.html): Linked from this page.
- [Derived features](https://docs.datarobot.com/en/docs/classic-ui/data/transform-data/feature-discovery/fd-gen.html): Linked from this page.
- [Predictions](https://docs.datarobot.com/en/docs/classic-ui/data/transform-data/feature-discovery/fd-predict.html): Linked from this page.
- [CV partitioning](https://docs.datarobot.com/en/docs/reference/pred-ai-ref/data-partitioning.html): Linked from this page.
- [Allowed source IP addresses](https://docs.datarobot.com/en/docs/reference/data-ref/allowed-ips.html): Linked from this page.

## Documentation content

# Feature Discovery

To deploy AI across the enterprise and make the best use of predictive models, you must be able to access relevant features. Often, the starting point of your data does not contain the right set of features. Feature Discovery discovers and generates new features from multiple datasets so that you no longer need to perform manual feature engineering to consolidate various datasets into one.

See the associated [considerations](https://docs.datarobot.com/en/docs/classic-ui/data/transform-data/feature-discovery/index.html#feature-considerations) for important additional information.

Select topics from the following table to learn about the feature engineering workflow:

| Topic | Description |
| --- | --- |
| End-to-end Feature Discovery | An end-to-end example that shows you how to enrich data using Feature Discovery. |
| Feature Discovery projects | Create and configure projects with secondary datasets, including a simple use-case-based workflow overview. |
| Snowflake integration | Set up an integration that allows joint users to both execute data science projects in DataRobot and perform computations in Snowflake. |
| Feature Discovery settings | Configure advanced options for Feature Discovery projects, including feature engineering controls and feature reduction. |
| Time-aware feature engineering | Configure time-aware feature engineering. |
| Derived features | Introduction to the list of aggregations and the feature reduction process. |
| Predictions | Score data with models created using secondary datasets. |

## Feature considerations

When using Feature Discovery, consider the following:

- JDBC drivers must be compatible with Java 1.8 and later.
- For secondary datasets, only uploaded files and JDBC sources registered in theAI Catalogare supported.
- The following features are not supported in Feature Discovery projects:
- Maximum supported values:
- If the primary dataset is larger than 40MB,CV partitioningis disabled by default.
- Column names in Feature Discovery datasets cannot contain the following:
- When there is an error during project start, you cannot return to defining relationships. You must restart the configuration.
- There can be issues with the colors used in the visualization of linkages in the Feature Engineering relationship editor.
- You must allow the IP addresses listed on theAllowed source IP addressespage to connect to the DataRobot JDBC connector.

### Batch prediction considerations

- Only DataRobot models are supported; no external or custom model support.
- Model package export is not supported for Feature Discovery models.
- You cannot replace a Feature Discovery model with a non-Feature Discovery model or vice versa.
- When a Feature Discovery model is replaced with another Feature Discovery model, the configuration used by the new model becomes the default configuration.
- Feature discovery predictions will be slower than other DataRobot models because feature engineering is applied.
- When Feature Discovery generates features using secondary datasets, the hash values of all the feature values (ROW_HASH) are used to break any ties (when applicable). The value of hash changes when applied to different datasets, so if you make predictions with another secondary configuration, you may receive different predictions.

### Feature Discovery compatibility

The following table indicates which features are supported for Feature Discovery and describes any limitations.

| Feature | Supported? | Limitations |
| --- | --- | --- |
| Monotonicity | Yes | Limited to features from the primary dataset used to start the project. Note: Users can start the project without specifying constraints. They can then manually constrain models from the Leaderboard and the Repository on eligible blueprints using discovered/generated features. |
| Pairwise interaction in GA2M models | Yes | Limited to features from the primary dataset used to start the project. |
| Positive class assignment | Yes |  |
| Smart downsampling | Yes |  |
| Supervised feature reduction | Yes | Only applies if secondary datasets are provided. |
| Search for interactions | Yes | Automatically enabled. Cannot be disabled if secondary datasets are provided. |
| Only blueprints with Scoring Code support | No |  |
| Create blenders from top models | Yes |  |
| Include only SHAP-supported blueprints | Yes |  |
| Recommend and prepare a model for deployment | Yes |  |
| Challenger models in MLOps | No |  |
| Include blenders when recommending a model | Yes |  |
| Use accuracy-optimized metablueprint | Yes | These models are extremely slow. |
| Upperbound running time | Yes |  |
| Weight | Yes | Weight feature must be in the primary dataset used to start the project. |
| Offset | Yes | Offset feature must be in the primary dataset used to start the project. |
| Exposure | Yes | Exposure feature must be in the primary dataset used to start the project. |
| Random seed | Yes |  |
| Count of events | Yes | Count of events feature must be in the primary dataset used to start the project. |
