Glossary¶
The DataRobot glossary provides brief definitions of terms relevant to the DataRobot platform. These terms span all phases of machine learning, from data to deployment.
All Agentic API Code-first Data MLOps Modeling Predictions Time-aware
A¶
Accuracy over space¶
A model Leaderboard tab (Evaluate > Accuracy Over Space) and Location AI insight that provides a spatial residual mapping within an individual model.
Accuracy over time¶
A model Leaderboard tab (Evaluate > Accuracy Over Time) that visualizes how predictions change over time.
ACE scores¶
Also known as Alternating Conditional Expectations. A univariate measure of correlation between the feature and the target. ACE scores detect non-linear relationships, but as they are univariate, they do not detect interaction effects.
Actuals¶
Actual values for an ML model that let you track its prediction outcomes. To generate accuracy statistics for a deployed model, you compare the model's predictions to real-world actual values for the problem. Both the prediction dataset and the actuals dataset must contain association IDs, which let you match up corresponding rows in the datasets to gauge the model's accuracy.
Advanced tuning¶
The ability to manually set model parameters after the model build, supporting experimentation with parameter settings to improve model performance.
Agent¶
An AI-powered component within DataRobot designed to execute complex, multi-step tasks autonomously. An agent can be configured with specific goals, LLMs, and a set of tools, allowing it to perform actions like orchestrating a data preparation workflow, running a modeling experiment, or generating an analysis without direct human intervention. Agents exhibit autonomous behavior, can reason about their environment, make decisions, and adapt their strategies based on feedback. Multiple agents can be combined in an agentic workflow to solve more sophisticated business problems through collaboration and coordination.
Agent-based modeling¶
Computational modeling approaches that simulate complex systems by modeling individual agents and their interactions. Agent-based modeling enables the study of emergent behaviors and system-level properties that arise from individual agent behaviors. In DataRobot's platform, agent-based modeling capabilities allow users to simulate business processes, test agent strategies, and understand how different agent configurations affect overall system performance.
Agentic AI¶
A paradigm of artificial intelligence where AI systems are designed to act as autonomous agents that can perceive their environment, reason about goals, plan actions, and execute tasks with minimal human oversight. Agentic AI systems are characterized by their ability to make independent decisions, learn from experience, and adapt their behavior to achieve objectives. In DataRobot's platform, agentic AI enables sophisticated automation of complex data science workflows, allowing AI systems to handle end-to-end processes from data preparation to model deployment and monitoring.
Agentic workflow¶
Systems that leverage AI agents to perform tasks and make decisions within a workflow, often with minimal human intervention. Agentic workflows can be built in a local IDE using DataRobot templates and a CLI and managed with real-time LLM intervention and moderation with out-of-the-box and custom guards, including integration with NVIDIA's NeMo for content safety and topical rails in the UI or with code.
Agent Framework (AF) components¶
Agent Framework (AF) components provide modular building blocks for constructing sophisticated AI agents. AF components include reasoning engines, memory systems, action planners, and communication modules that can be combined to create custom agent architectures. In DataRobot's platform, AF components enable rapid development of specialized agents with specific capabilities while maintaining consistency and interoperability across different agent implementations.
Agent-to-Agent (A2A)¶
Agent-to-Agent (A2A) refers to communication protocols and frameworks that enable direct interaction and coordination between AI agents. A2A systems facilitate information sharing, task delegation, and collaborative problem-solving among multiple agents. In DataRobot's agentic workflows, A2A capabilities enable agents to work together seamlessly, share context and knowledge, and coordinate complex multi-agent operations while maintaining security and governance controls.
Aggregate image feature¶
Used with Visual AI, a set of image features where each individual element of that set is a constituent image feature. For example, the set of image features extracted from an image might include a set of features indicating:
- The colors of the individual pixels in the image.
- Where edges are present in the image.
- Where faces are present in the image.
From the aggregate it may be possible to determine the impact of that feature on the output of a data analytics model and compare that impact to the impacts of the model's other features.
AI catalog¶
A browsable and searchable collection of registered objects that contains definitions and relationships between various objects types. Items stored in the catalog include: data connections, data sources, data metadata.
AI tools¶
Software applications, libraries, and frameworks designed to support the development, deployment, and management of artificial intelligence systems. In DataRobot, AI tools include built-in capabilities for model building, evaluation, deployment, and monitoring, as well as integrations with external AI services and frameworks.
AIM¶
The second phase of Exploratory Data Analysis (i.e., EDA2), that determines feature importance based on cross-correlation with the target feature. That data determines the "informative features" used for modeling during Autopilot.
Alignment¶
The critical process of steering an AI model's outputs and behavior to conform to an organization's specific ethical guidelines, safety requirements, and business objectives. In DataRobot, alignment is practically applied through features like guardrails, custom system prompts, and content moderation policies. This practice helps to mitigate risks from biased, unsafe, or off-topic model responses, ensuring the AI remains a trustworthy and reliable tool for the enterprise.
Alternating conditional expectations¶
See ACE scores.
Anomaly detection¶
A form of unsupervised learning used to detect anomalies in data. Anomaly detection, also referred to as outlier or novelty detection, can be useful with data having a low percentage of irregularities or large amounts of unlabeled data. See also unsupervised learning.
Apps¶
See No-Code AI Apps.
ARIMA (AutoRegressive Integrated Moving Average)¶
A time series modeling approach available in DataRobot time series that analyzes historical patterns to forecast future values. DataRobot's ARIMA implementation automatically handles parameter selection and optimization, making it accessible for users without deep statistical expertise while maintaining the mathematical rigor of traditional ARIMA models.
Autoregressive¶
A modeling approach where predictions are made sequentially, with each prediction depending on previous outputs. In DataRobot, autoregressive models are commonly used in time series forecasting and natural language processing tasks, where the model learns patterns from historical data to predict future values or generate text one step at a time. This technique enables coherent sequence generation and is particularly effective for time-dependent data where temporal relationships are crucial for accurate predictions.
Asset¶
One of the components of a Use Case that can be added, managed, and shared within Workbench. Components include data, vector databases, experiments, playgrounds, apps, and notebooks.
Association ID¶
An identifier that functions as a foreign key for your prediction dataset so you can later match up actual values (or "actuals") with the predicted values from the deployed model. An association ID is required for monitoring the accuracy of a deployed model.
AUC (Area Under the Curve)¶
A common error metric for binary classification that considers all possible thresholds and summarizes performance in a single value on the ROC Curve. It works by optimizing the ability of a model to separate the 1s from the 0s. The larger the area under the curve, the more accurate the model.
Audit log¶
A chronological, immutable record of all significant activities performed within the DataRobot platform by users and automated processes. It is essential for security audits, compliance reporting, and troubleshooting. Sometimes referred to as an "event log".
Augmented intelligence¶
DataRobot's enhanced approach to artificial intelligence, which expands current model building and deployment assistance practices. The DataRobot platform fully automates and governs the AI lifecycle from data ingest to model training and predictions to model-agnostic monitoring and governance. Guardrails ensure adherence to data science best practices when creating machine learning models and AI applications. Transparency across user personas and access to data wherever it resides avoids lock-in practices.
Autonomy¶
The ability of an AI agent to operate independently and make decisions without constant human oversight. Autonomous agents can plan, execute, and adapt their behavior based on changing conditions and feedback. In DataRobot's agentic workflows, autonomous capabilities are balanced with human oversight through guardrails and monitoring to ensure safe and effective operation. Autonomy enables agents to handle complex, multi-step processes while maintaining alignment with business objectives and safety requirements.
Authentication¶
The process of verifying the identity of users, applications, or systems before granting access to DataRobot's APIs and services. DataRobot supports multiple authentication methods, including API keys for programmatic access, OAuth 2.0 for web applications, and Single Sign-On (SSO) integration with enterprise identity providers. Authentication ensures secure access to projects, deployments, and platform resources while maintaining audit trails for compliance and security monitoring.
Authorization¶
The process of determining what actions or resources users or systems are permitted to access after authentication.
Automated retraining¶
Retraining strategies for MLOps that refresh production models based on a schedule or in response to an event (for example, a drop in accuracy or data drift). Automated Retraining also uses DataRobot's AutoML create and recommend new challenger models. When combined, these strategies maximize accuracy and enable timely predictions.
AutoML (Automated Machine Learning)¶
A software system that automates many of the tasks involved in preparing a dataset for modeling and performing a model selection process to determine the performance of each with the goal of identifying the best performing model for a specific use case. Used for predictive modeling; see also time series for forecasting.
Autopilot (full Autopilot)¶
The DataRobot "survival of the fittest" modeling mode that automatically selects the best predictive models for the specified target feature and runs them at ever-increasing sample sizes. In other words, it runs more models in the early stages on a small sample size and advances only the top models to the next stage. In full Autopilot, DataRobot runs models at 16% (by default) of total data and advances the top 16 models, then runs those at 32%. Taking the top 8 models from that run, DataRobot runs on 64% of the data (or 500MB of data, whichever is smaller). See also Quick (Autopilot), Comprehensive, and Manual.
AutoTS (Automated time series)¶
A software system that automates all or most of the steps needed to build forecasting models, including featurization, model specification, model training, model selection, validation, and forecast generation. See also time series.
Average baseline¶
The average of the target in the Feature Derivation Window; used in time series modeling.
B¶
Backend¶
The server-side components of LLM and AI applications that handle data processing, model inference, business logic, and database operations.
Backtesting¶
The time-aware equivalent of cross-validation. Unlike cross-validation, however, backtests allow you to select specific time periods or durations for your testing instead of random rows, creating "trials" for your data.
Baseline model¶
Also known as a naive model. A simple model used as a comparison point to confirm that a generated ML or time series model is learning with more accuracy than a basic non-ML model.
For example, generated ML models for a regression project should perform better than a baseline model that predicts the mean or median of the target. Generated ML models for a time series project should perform better than a baseline model that predicts the future using the most recent actuals (i.e., using today's actual value as tomorrow's prediction).
For time series projects, baseline models are used to calculate the MASE metric (the ratio of the MAE metric over the baseline model).
Batch predictions¶
A method of making predictions with large datasets, in which you pass input data and get predictions for each row; predictions are written to output files. Users can make batch predictions with MLOps via the Predictions interface or can use the Batch Prediction API for automating predictions. Schedule batch prediction jobs by specifying the prediction data source and destination and determining when the predictions will be run.
Bias mitigation¶
Augments blueprints with a pre- or post-processing task intended to reduce bias across classes in a protected feature. Bias Mitigation is also a model Leadboard tab (Bias and Fairness > Bias Mitigation) where you can apply mitigation techniques after Autopilot has finished.
Bias vs accuracy¶
A Leaderboard tab that generates a chart to show the tradeoff between predictive accuracy and fairness, removing the need to manually note each model's accuracy score and fairness score for the protected features.
Bias (AI bias)¶
Systematic prejudice in AI model outputs that reflects unfair treatment of certain groups or individuals. AI bias can manifest in various forms, including gender bias, racial bias, or socioeconomic bias, and can result from biased training data, model architecture, or deployment contexts. DataRobot provides tools and practices to detect, measure, and mitigate bias in AI systems.
Blind history¶
"Blind history", used in time-aware modeling, captures the gap created by the delay of access to recent data (e.g., "most recent" may always be one week old). It is defined as the period of time between the smaller of the values supplied in the Feature Derivation Window and the forecast point. A gap of zero means "use data up to, and including, today;" a gap of one means "use data starting from yesterday" and so on.
Blender¶
A model that potentially increases accuracy by combining the predictions of between two and eight models. DataRobot can be configured to automatically create blender models as part of Autopilot, based on the top three regular Leaderboard models (for AVG, GLM, and ENET blenders). You can also create blenders manually (aka ensemble models).
Blueprint¶
A blueprint is a graphical representation of the many steps involved in transforming input predictors and targets into a model. It represents the high-level end-to-end procedure for fitting the model, including any preprocessing steps, algorithms, and post-processing. Each box in a blueprint may represent multiple steps. You can view the graphical representation of a blueprint by clicking on a model on the Leaderboard. See also user blueprints.
C¶
Caching strategies¶
Techniques for storing frequently accessed LLM responses, embeddings, or intermediate results to improve performance and reduce computational costs.
Canary deployment¶
A deployment strategy for LLM and AI models that gradually rolls out new versions to a small subset of users before full deployment, allowing for early detection of issues.
"Can't operationalize" period¶
The "can't operationalize" period, used in time series modeling, defines the gap of time immediately after the Forecast Point and extending to the beginning of the Forecast Window. It represents the time required for a model to be trained, deployed to production, and to start making predictions—the period of time that is too near-term to be useful. For example, predicting staffing needs for tomorrow may be too late to allow for taking action on that prediction.
Catalog¶
See AI Catalog.
Centroid¶
The center of a cluster generated using unsupervised learning. A centroid is the multi-dimensional average of a cluster, where the dimensions are observations (data points).
CFDS (Customer Facing Data Scientist)¶
A DataRobot employee responsible for the technical success of user and potential users. They assist with tasks like structuring data science problems to complete integration of DataRobot. CFDS are passionate about ensuring user success.
Chain-of-thought¶
A prompting technique that encourages language models to break down complex problems into step-by-step reasoning processes. In DataRobot's agentic workflows, chain-of-thought prompting enhances agent reasoning capabilities by requiring explicit intermediate steps in decision-making, leading to more transparent and reliable outcomes. This technique improves problem-solving accuracy and enables better debugging and validation of agent behavior in multi-step tasks.
Challenger models¶
Models that you can compare to a currently deployed model (the "champion" model) to continue model comparison post-deployment. Submit a challenger model to shadow a deployed model and replay predictions made against the champion to determine if there is a superior DataRobot model that would be a better fit.
Champion model¶
A model recommended by DataRobot—for a deployment (predictions) or for time series segmented modeling.
In MLOps, you can replace the champion selected for a deployment yourself, or you can set up Automated Retraining, where DataRobot compares challenger models with the champion model and replaces the champion model if a challenger outperforms the champion.
In the segmented modeling workflow, DataRobot builds a model for each segment. DataRobot recommends the best model for each segment—the segment champion. The segment champions roll up into a Combined Model. For each segment, you can select a different model as champion, which is then used in the Combined Model.
Channel¶
The connection between an output port of one module and an input port of another module. Data flows from one module's output port to another module's input port via a channel, represented visually by a line connecting the two.
Chatting¶
Sending prompts (and as a result, LLM payloads) to LLM endpoints based on a single LLM blueprint and receiving a response from the LLM. In this case, context from previous prompts/responses is sent along with the payload.
Chunking¶
The action of taking a body of unstructured text and breaking it up into smaller pieces of unstructured text (tokens).
Citation¶
The chunks of text from the vector database used during the generation of LLM responses.
CI/CD pipelines¶
Continuous Integration (CI) and Continuous Deployment (CD) pipelines that automate the building, testing, and deployment of LLM and AI applications to ensure reliable and consistent releases.
Circuit breaker¶
A crucial MLOps reliability pattern that safeguards a deployed model by monitoring for high error rates or latency. If a predefined failure threshold is breached, the circuit breaker automatically and temporarily redirects or pauses traffic to the unhealthy model instance. This action prevents a single failing model from causing a cascade failure across an application and allows the system time to recover, ensuring high availability for production AI services.
Classification¶
A DataRobot modeling approach that predicts categorical outcomes from your target feature. DataRobot supports three classification types: binary classification for two-class problems (like "churn" vs "retain"), multiclass classification for multiple discrete outcomes (like "buy", "sell", "hold"), and unlimited multiclass for projects with numerous possible classes. DataRobot automatically selects appropriate classification algorithms from the Repository and provides specialized evaluation metrics like AUC and confusion matrices to assess model performance. See also regression.
CLI¶
Command Line Interface (CLI) tools that enable programmatic interaction with DataRobot's agentic workflows and platform services. CLI tools provide scriptable access to agent configuration, workflow execution, and platform management functions. In DataRobot's agentic ecosystem, CLI tools support automation of agent deployment, monitoring, and maintenance tasks, enabling integration with CI/CD pipelines and automated workflows.
Clustering¶
A form of unsupervised learning used to group similar data and identify natural segments.
Cognitive architecture¶
The underlying structural framework that defines how AI agents process information, make decisions, and interact with their environment. Cognitive architectures specify the components, processes, and relationships that enable intelligent behavior in agents. In DataRobot's agentic workflows, cognitive architectures provide the foundation for agent reasoning, memory management, learning, and decision-making capabilities, enabling sophisticated autonomous behavior.
Codespace¶
A fully configured Integrated Development Environment (IDE) hosted on the cloud. It provides tools for you to write, test, and debug code. It also offers file storage so that notebooks inside a codespace can reference Python utility scripts and other assets.
Coefficients¶
A model Leaderboard tab (Describe > Coefficients) that provides a visual indicator of information that can help you refine and optimize your models.
Combined model¶
The final model generated in a time series segmented modeling workflow. With segmented modeling, DataRobot builds a model for each segment and combines the segment champions into a single Combined Model that you can deploy.
Common event¶
A data point is a common event if it occurs in a majority of weeks in data (for example, regular business days and hours would be common, but an occasional weekend data point would be uncommon).
Compliance documentation¶
Automated model development documentation that can be used for regulatory validation. The documentation provides comprehensive guidance on what constitutes effective model risk management.
Compliance reporting¶
The generation of reports and documentation required for regulatory compliance in LLM and AI deployments, including data usage, model performance, and security measures.
Composable ML¶
A code-centric feature, designed for data scientists, that allows applying custom preprocessing and modeling methods to create a blueprint for model training. Using built-in and custom tasks, you can compose and then integrate the new blueprint with other DataRobot features to augment and improve machine learning pipelines.
Comprehensive¶
A modeling mode that runs all Repository blueprints on the maximum Autopilot sample size to ensure more accuracy for models.
Computer vision¶
Use of computer systems to analyze and interpret image data, used with Visual AI. Computer vision tools generally use models that incorporate principles of geometry to solve specific problems within the computer vision domain. For example, computer vision models may be trained to perform object recognition (recognizing instances of objects or object classes in images), identification (identifying an individual instance of an object in an image), detection (detecting specific types of objects or events in images), etc.
Computer vision tools/techniques¶
Tools—for example, models, systems—that perform image preprocessing, feature extraction, and detection/segmentation functions.
Connected vector database¶
An external vector database accessed via a direct connection to a supported provider for vector database creation. The data source is stored locally in the Data Registry, configuration settings are applied, and the created vector database is written back to the provider. Connected vector databases maintain real-time synchronization with the platform and provide seamless access to embeddings and text chunks for grounding LLM responses.
Configuration management¶
The practice of managing LLM and AI system configurations across different environments (development, staging, production) to ensure consistency and reduce deployment errors.
Confusion matrix¶
A table that reports true versus predicted values. The name "confusion matrix" refers to the fact that the matrix makes it easy to see if the model is confusing two classes (consistently mislabeling one class as another class). The confusion matrix is available as part of the ROC Curve, Eureqa, and Confusion Matrix for multiclass model visualizations in DataRobot.
Connection instance¶
A connection that is configured with metadata about how to connect to a source system (e.g., instance of a Snowflake connection).
Console¶
Console is a central hub for deployment management activity. Its dashboard provides access to deployed models for further monitoring and mitigation. It also provides access to prediction activities and allows you to view, create, edit, delete, or share serverless and external prediction environments
Constraints¶
A model Leaderboard tab (Describe > Constraints) that allows you to review monotonically constrained features if feature constraints were configured in Advanced Options prior to modeling.
Container orchestration¶
The automated management of containerized LLM and AI applications, including deployment, scaling, networking, and availability, typically using platforms like Kubernetes.
Context window¶
The limited amount of information, measured in tokens, that a large language model can hold in its active memory for a single chat conversation turn. This 'memory' includes the user's prompt, any recent conversation history provided, and data retrieved via Retrieval Augmented Generation (RAG). The size of the context window is a critical parameter in an LLM blueprint, as it dictates the model's ability to handle long documents or maintain coherence over extended dialogues; any information outside this window is not considered when generating the next response.
Conversation memory¶
The ability of an AI system to remember and reference previous interactions within a conversation session (meaning that the session contains one or more chat conversation turns). Conversation memory enables contextual continuity, allowing the AI to maintain awareness of earlier exchanges and build upon previous responses. In DataRobot's chat interfaces, conversation memory helps maintain coherent, contextually relevant dialogues.
Cost allocation¶
The process of assigning LLM and AI service costs to different teams, projects, or business units for budgeting and chargeback purposes.
Credentials¶
Information used to authenticate and authorize actions against data connections. The most common connection is through username and password, but alternate authentication methods include LDAP, Active Directory, and Kerberos.
Cross-class accuracy¶
A model Leaderboard tab (Bias and Fairness > Cross-Class Accuracy) that helps to shows why the model is biased, and where in the training data it learned the bias from. Bias and Fairness settings must be configured.
Cross-class data disparity¶
A model Leaderboard tab (Bias and Fairness > Cross-Class Data Disparity) that calculates, for each protected feature, evaluation metrics and ROC curve-related scores segmented by class. Bias and Fairness settings must be configured.
Cross-Validation (CV)¶
DataRobot's validation approach that tests model performance by creating multiple training and validation partitions from your data. DataRobot automatically implements five-fold cross-validation by default, building separate models on different data subsets and using the remaining data for validation. This process generates more reliable performance estimates than single validation splits, and DataRobot displays the average cross-validation scores on the Leaderboard to help you select the best model. See also validation.
Custom inference models¶
User-created, pre-trained models uploaded as a collection of files via the Custom Model Workshop. Upload a model artifact to create, test, and deploy custom inference models to the centralized deployment hub in DataRobot. An inference model can have a predefined input/output schema or it can be unstructured. To customize prior to model training, use custom tasks.
Custom model environment¶
A versioned, containerized environment (e.g., a Docker image) that includes all the necessary libraries, packages, and dependencies required to run a custom model or task within DataRobot. Administrators manage these environments to ensure reproducibility and governance.
Custom model workshop¶
In the Model Registry, a location where you can upload user-created, pre-trained models as a collection of files. You can use these model artifacts to create, test, and deploy custom inference models to centralized deployment hub in DataRobot.
Custom task¶
A data transformation or ML algorithm, for example, XGBoost or One-hot encoding, that can be used as a step in an ML blueprint inside DataRobot and used for model training. Tasks are written in Python or R and are added via the Custom Model Workshop. Once saved, the task can be used when modifying a blueprint with Composable ML. To deploy a pre-trained model where re-training is not required, use custom inference models.
CV¶
See Cross Validation.
D¶
Data classification¶
The process of categorizing data based on sensitivity, regulatory requirements, and business value to determine appropriate handling, storage, and access controls for LLM and AI systems. DataRobot provides automated PII detection and data governance features to help organizations classify and protect sensitive information in their datasets.
Data drift¶
The difference between values in new inference data used to generate predictions for models in production and the training data initially used to train the deployed model. Predictive models learn patterns in training data and use that information to predict target values for new data. When the training data and the production data change over time, causing the model to lose predictive power, the data surrounding the model is said to be drifting. Data drift can happen for a variety of reasons, including data quality issues, changes in feature composition, and even changes in the context of the target variable.
Data management¶
The umbrella term related to loading, cleaning, transforming, and storing data within DataRobot. It also refers to the practices that companies follow when collecting, storing, using, and deleting data.
Data preparation¶
The process of transforming raw data to the point where it can be run through machine learning algorithms to uncover insights or make predictions. Also called "data preprocessing," this term covers a broad range of activities like normalizing data, standardizing data, statistically or mathematically transforming data, processing and/or preprocessing data, and feature engineering.
Data Quality Handling Report¶
A model Leaderboard tab (Describe > Data Quality Handling Report) that analyzes the training data and provides the following information for each feature: feature name, variable type, row count, percentage, and data transformation information.
Data Registry¶
In Workbench, a central catalog for datasets, allowing you to link them to specific Use Cases.
Data residency¶
The physical or geographical location where LLM and AI data is stored and processed, often subject to regulatory requirements and compliance standards. DataRobot supports various deployment options including cloud, on-premises, and hybrid configurations to meet specific data residency requirements.
Data retention policies¶
Policies that define how long LLM and AI data should be kept, when it should be archived, and when it should be deleted to comply with regulations and manage storage costs.
Data wrangling¶
Data preparation operations of a scope that ties to creating a dataset at an appropriate unit of analysis for a given machine learning use case.
DataRobot Classic¶
The original DataRobot value-driven AI product. It provides a complete AI lifecycle platform, leveraging machine learning that has broad interoperability and end-to-end capabilities for ML experimentation and production. DataRobot Classic is being migrated to the new user interface, Workbench.
DataRobot User Models (DRUM)¶
A tool that allows you to test Python, R, and Java custom models and tasks locally. The test allows you to verify that a custom model can successfully run and make predictions in DataRobot before uploading it.
Dataset¶
Data, a file or the content of a data source, at a particular point in time. A data source can produce multiple datasets; an AI Catalog dataset has exactly one data source. In AI Catalog, a dataset is materialized data that is stored with a catalog version record. There may be multiple catalog version records associated with an entity, indicating that DataRobot has reloaded or refreshed the data. The older versions are stored to support existing projects, new projects use the most recent version. A dataset can be in one of two states:
- A "snapshotted" (or materialized) dataset is an immutable snapshot of data that has previously been retrieved and saved.
- A "remote" (or unmaterialized) dataset has been configured with a location from which data is retrieved on-demand (AI Catalog).
Data connection¶
A configured connection to a database—it has a name, a specified driver, and a JDBC URL. You can register data connections with DataRobot for ease of re-use. A data connection has one connector but can have many data sources.
Data source¶
A configured connection to the backing data (the location of data within a given endpoint). A data source specifies, via SQL query or selected table and schema data, which data to extract from the data connection to use for modeling or predictions. Examples include the path to a file on HDFS, an object stored in S3, and the table and schema within a database. A data source has one data connection and one connector but can have many datasets. It is likely that the features and columns in a datasource do not change over time, but that the rows within change as data is added or deleted.
Data stage¶
Intermediary storage that supports multipart upload of large datasets, reducing the chance of failure when working with large amounts of data. Upon upload, the dataset is uploaded in parts to the data stage, and once the dataset is whole and finalized, it is pushed to the AI Catalog or Batch Predictions. At any time after the first part is uploaded to the data stage, the system can instruct Batch Predictions to use the data from the data stage to fill in predictions.
Data store¶
A general term used to describe a remote location where your data is stored. A data store may contain one or more databases, or one or more files of varying formats.
Data/time partitioning¶
The only valid partitioning method for time-aware projects. With date/time, rows are assigned to backtests chronologically instead of, for example, randomly. Backtests are configurable, including number, start and end times, and sampling method.
Dashboard¶
Visual monitoring interfaces that provide real-time insights into LLM and AI system performance, health, and operational metrics for administrators and stakeholders. DataRobot provides comprehensive dashboards for monitoring model performance, data drift, prediction accuracy, and system health across all deployments.
Deep learning¶
DataRobot's implementation of neural network architectures that process data through multiple computational layers. These algorithms power DataRobot's Visual AI capabilities for image analysis and are available as blueprints in the model Repository. Users can monitor training progress and layer performance through the Training Dashboard visualization, making deep learning accessible without requiring expertise in neural network architecture design.
Deploying (from a playground)¶
LLM blueprints and all their associated settings are registered in Registry and can be deployed with DataRobot's production suite of products.
Deployment inventory¶
The central hub for managing deployments. Located on the Deployments page, the inventory serves as a coordination point for stakeholders involved in operationalizing models. From the inventory, you can monitor deployed model performance and take action as necessary, managing all actively deployed models from a single point.
Detection/segmentation¶
A computer vision technique that involves the selection of a subset of the input image data for further processing (for example, one or more images within a set of images or regions within an image).
Directed acyclic graph (DAG)¶
A mathematical structure used to represent workflows where nodes represent tasks or operations and edges represent dependencies between them. In AI workflows, DAGs ensure that tasks are executed in the correct order without circular dependencies, enabling efficient orchestration of complex multi-step processes like data preprocessing, model training, and deployment pipelines.
Disaster recovery¶
Plans and procedures for recovering LLM and AI services after system failures, natural disasters, or other catastrophic events to ensure business continuity. DataRobot provides backup and restore capabilities, along with high availability configurations to minimize downtime and ensure continuous model serving.
Distributed tracing¶
A technique for monitoring and troubleshooting LLM and AI applications by tracking requests as they flow through multiple services and components.
Downloads tab¶
A model Leaderboard tab (Predict > Downloads) where you can download model artifacts.
Downsampling¶
See Smart downsampling.
Driver¶
The software that allows the DataRobot application to interact with a database; each data connection is associated with one driver (created and installed by your administrator). The driver configuration saves the JAR file storage location in DataRobot and any additional dependency files associated with the driver. DataRobot supports JDBC drivers.
Dynamic dataset¶
A dynamic dataset is a "live" connection to the source data, however, DataRobot samples the data for profile statistics (EDA1). The catalog stores a pointer to the data and pulls it upon request, for example, when you create a project.
E¶
EDA (Exploratory Data Analysis)¶
The DataRobot approach to analyzing and summarizing the main characteristics of a dataset. Generally speaking, there are two stages of EDA:
- EDA1 provides summary statistics based on a sample of data. In EDA1, DataRobot counts, categorizes, and applies automatic feature transformations (where appropriate) to data.
- EDA2 is a recalculation of the statistics collected in EDA1 but using the entire dataset, excluding holdout. The results of this analysis are the criteria used for model building.
Embedding¶
A numerical (vector) representation of text, or a collection of numerical representations of text. The action of generating embeddings means taking a chunk of unstructured text and using a text embedding model to convert the text to a numerical representation. The chunk is the input to the embedding model and the embedding is the "prediction" or output of the model.
Episodic memory¶
Memory systems that store specific experiences, events, and contextual information about past interactions and situations. Episodic memory enables AI agents to recall specific instances, learn from particular experiences, and apply contextual knowledge to similar situations. In DataRobot's agentic workflows, episodic memory allows agents to remember specific user interactions, successful task executions, and contextual details that inform future decision-making.
Endpoint¶
A specific URL where a service can be accessed. In machine learning, an endpoint is typically used to send data to a deployed model and receive predictions. It is the primary interface for interacting with a model programmatically via an API.
Ensemble models¶
See blender.
Environment¶
A Docker container where a custom task runs.
Environment management¶
The practice of managing different environments (development, staging, production) for LLM and AI systems to ensure proper testing, deployment, and operational procedures.
ESDA¶
Exploratory Spatial Data Analysis (ESDA) is the exploratory data phase for Location AI. DataRobot provides a variety of tools for conducting ESDA within the DataRobot AutoML environment, including geometry map visualizations, categorical/numeric thematic maps, and smart aggregation of large geospatial datasets.
Eureqa¶
Model blueprints for Eureqa generalized additive models (Eureqa GAM), Eureqa regression, and Eureqa classification models. These blueprints use a proprietary Eureqa machine learning algorithm to construct models that balance predictive accuracy against complexity.
Event streaming¶
Real-time data processing systems that handle continuous streams of events from LLM and AI applications for monitoring, analytics, and operational insights.
EWMA (Exponentially Weighted Moving Average)¶
A moving average that places a greater weight and significance on the most recent data points, measuring trend direction over time. The "exponential" aspect indicates that the weighting factor of previous inputs decreases exponentially. This is important because otherwise a very recent value would have no more influence on the variance than an older value.
Experiment¶
An asset of a Use Case that is the result of having run the DataRobot modeling process. A Use Case can have zero or more experiments.
Experiment tracking¶
The process of recording and managing metadata, parameters, and results from machine learning experiments to enable reproducibility and comparison.
Exploratory data insights¶
See EDA.
External stage¶
A designated location in a cloud storage provider (such as Amazon S3 or Azure) that is configured to act as an intermediary for loading and unloading data with a Snowflake database. When preparing data for a project in DataRobot, users may interact with an external stage to efficiently ingest large datasets from Snowflake or to publish transformed data back to the cloud environment.
F¶
Fairness score¶
A numerical computation of model fairness against the protected class, based on the underlying fairness metric.
Fairness threshold¶
The measure of whether a model performs within appropriate fairness bounds for each protected class. It does not affect the fairness score or performance of any protected class.
Fairness value¶
Fairness scores normalized against the most favorable protected class (i.e., the class with the highest fairness score).
Favorable outcome¶
A value of the target that is treated as the favorable outcome for the model, used in bias and fairness modeling. Predictions from a binary classification model can be categorized as being a favorable outcome (i.e., good/preferable) or an unfavorable outcome (i.e., bad/undesirable) for the protected class.
FDW¶
See Feature Derivation Window.
Feature¶
A column in a dataset, also called "variable" or "feature variable." The target feature is the name of the column in the dataset that you would like to predict.
Feature Derivation Window¶
Also known as FDW; used in time series modeling. A rolling window of past values that models use to derive features for the modeling dataset. Consider the window relative to the Forecast Point, it defines the number of recent values the model can use for forecasting.
Feature Discovery¶
A DataRobot capability that discovers and generates new features from multiple datasets, eliminating the need to perform manual feature engineering to consolidate multiple datasets into one. A relationship editor visualizes these relationships and the end product is additional, derived features that result from the created linkages.
Feature Effects¶
A model Leaderboard tab (Understand > Feature Effects) that shows the effect of changes in the value of each feature on the model's predictions. It displays a graph depicting how a model "understands" the relationship between each feature and the target, with the features sorted by Feature Impact.
Feature engineering¶
The generation of additional features in a dataset, which as a result, improve model accuracy and performance. Time series and Feature Discovery both rely on feature engineering as the basis of their functionality.
Feature extraction¶
Models that perform image preprocessing (or image feature extraction and image preprocessing) are also known as "image feature extraction models" or "image-specific models."
Feature Extraction and Reduction (FEAR)¶
The feature generation process for time series modeling (e.g., lags, moving averages). It extracts new features (now) and then reduces the set of extracted features (later). See Time series feature derivation.
Feature flag¶
A DataRobot mechanism that allows administrators to enable or disable specific features for certain users, organizations, or the entire platform. Feature flags are used to manage phased rollouts, beta testing, and custom configurations. Toggling a feature flag is performed by DataRobot Support for SaaS customers.
Feature Impact¶
A measurement that identifies which features in a dataset have the greatest effect on model decisions. In DataRobot, the measurement is reported as a visualization available from the Leaderboard.
Feature imputation¶
A mechanism in time series modeling that uses forward filling to enable imputation for all features (target and others) when using the time series data prep tool. This results in a dataset with no missing values (with the possible exception of leading values at the start of each series where there is no value to forward fill).
Feature list¶
A subset of features from a dataset used to build models. DataRobot creates several lists during EDA2 including all informative features, informative features excluding those with a leakage risk, a raw list of all original features, and a reduced list. Uses can create project-specific lists as well.
Few-shot learning¶
A capability of a model to learn to perform a task from a small number of examples provided in the prompt.
Few-shot prompting¶
A technique where a few examples are provided in the prompt (either in an input or system prompt) to guide the model's behavior and improve its performance on specific tasks. Few-shot prompting helps models understand the desired output format and style without requiring fine-tuning, making it useful for quick adaptation to new tasks or domains.
Fine-tuning¶
The process of adapting pre-trained foundation models to specific tasks or domains by continuing training on targeted datasets. In DataRobot's platform, fine-tuning enables users to customize large language models for particular use cases, improving performance on domain-specific tasks while preserving general capabilities. Unlike prompt engineering which works with existing model weights, fine-tuning modifies the model's internal parameters to create specialized versions optimized for particular applications, industries, or data types.
Fitting¶
See model fitting.
Forecast Distance¶
A unique time step—a relative position—within the Forecast Window in a time series modeling project. A model outputs one row for each Forecast Distance.
Forecast Point¶
In time series modeling, the point you are making a prediction from; a relative time "if it was now..."; DataRobot trains models using all potential forecast points in the training data. In production, it is typically the most recent time.
Forecast vs Actual¶
A model Leaderboard tab (Evaluate > Forecast vs Actual) commonly used in time series projects that allows you to compare how different predictions behave from different forecast points to different times in the future. Although similar to the Accuracy Over Time chart, which displays a single forecast at a time, the Forecast vs Actual chart shows multiple forecast distances in one view.
Forecast Window¶
Also known as FW; used in time series modeling. Beginning from the Forecast Point, defines the range (the Forecast Distance) of future predictions—"this is the range of time I care about." DataRobot then optimizes models for that range and ranks them on the Leaderboard on the average across that range.
Forecasting¶
Predictions based on time, into the future; use inputs from recent rows to predict future values. Forecasting is a subset of predictions, using trends in observation to characterize expected outcomes or expected responses.
Foundation model¶
A powerful, large-scale AI model, like GPT or Claude, that provides broad, general-purpose capabilities learned from massive datasets. In the DataRobot platform, these models act as the core component or 'foundation' of an LLM blueprint. Rather than being a ready-made solution, a foundation model is the versatile starting point that can be customized for specific business needs through techniques like prompting, RAG, or fine-tuning.
Fast API¶
A modern, high-performance web framework for building APIs with Python. FastAPI provides automatic API documentation, type validation, and high performance through async support. In DataRobot's ecosystem, FastAPI is used for building custom API endpoints, microservices, and integration layers that support agentic workflows and custom model deployments.
Frozen run¶
A process that "freezes" parameter settings from a model's early, small sample size-based run. Because parameter settings based on smaller samples tend to also perform well on larger samples of the same data.
Function calling¶
The capability of large language models to invoke external functions, tools, or APIs based on user requests and conversation context. In DataRobot's agentic workflows, function calling enables agents to perform actions beyond text generation, such as data retrieval, mathematical computations, API interactions, and system operations. This allows agents to execute complex tasks, integrate with enterprise systems, and provide dynamic responses based on real-time information. Function calling transforms conversational AI into actionable systems that can manipulate data and interact with external services.
FW¶
See Forecast Window.
G¶
Generative AI (GenAI)¶
A type of artificial intelligence that generates new content based on learned patterns from training data. In DataRobot's platform, GenAI capabilities include text generation, content creation, and intelligent responses through LLM blueprints. Unlike traditional predictive models that analyze existing data, GenAI creates novel outputs through prompting and can be integrated into DataRobot workflows for content generation, analysis, and automated decision-making processes.
Governance lens¶
A filtered view of DataRobot's deployment inventory on the Deployments page, summarizing the social and operational aspects of a deployment. These include the deployment owner, how the model was built, the model's age, and the humility monitoring status.
GPU (graphics processing unit)¶
A specialized processor designed for parallel computing tasks, particularly effective for deep learning and AI workloads. GPUs excel at matrix operations and parallel processing, making them ideal for training complex models on large datasets. In DataRobot, GPU acceleration is available for supported deep learning blueprints and can significantly reduce training time for models that process text, images, or other computationally intensive tasks.
Guardrails¶
Safety mechanisms that prevent AI systems from generating harmful or inappropriate content. Guardrails include content filtering, output validation, and behavioral constraints that ensure AI responses align with safety guidelines and organizational policies. In DataRobot, guardrails can be configured and help maintain responsible AI practices and prevent the generation of unsafe or unethical content.
Grid search¶
An exhaustive search method used for hyperparameters.
Grounding¶
The process of ensuring that language model responses are based on specific, verifiable data sources rather than relying solely on training data. In DataRobot's platform, grounding is achieved through Retrieval Augmented Generation (RAG) workflows that connect LLMs to vector databases containing relevant documents, knowledge bases, or enterprise data. This technique improves response accuracy, reduces hallucinations, and ensures that AI outputs are contextualized with current, relevant information from trusted sources.
Group¶
A collection of users who share common permissions and access to projects, deployments, and other resources within an organization. Groups simplify user management by allowing administrators to manage permissions for multiple users at once.
H¶
Hallucination¶
When a language model generates information that is plausible-sounding but factually incorrect or not grounded in the provided data.
Health checks¶
Automated monitoring systems that verify the health and availability of LLM and AI services by periodically checking their status and responsiveness.
High availability¶
System design principles and practices that ensure LLM and AI services remain available and operational even during hardware failures, software issues, or other disruptions.
High code¶
A development approaches that emphasizes custom programming and fine-grained control over application behavior. High-code solutions provide maximum flexibility and customization capabilities for complex requirements. In DataRobot's agentic workflows, high-code capabilities enable advanced users to create highly specialized agents with custom logic, integrate with complex enterprise systems, and implement sophisticated decision-making algorithms.
Holdout¶
A subset of data that is unavailable to models during the training and validation process. Use the Holdout score for a final estimate of model performance only after you have selected your best model. See also Validation.
HTTP Status Codes¶
Standard response codes returned by DataRobot APIs to indicate the success or failure of requests. Common codes include 200 (success), 400 (bad request), 401 (unauthorized), 404 (not found), and 500 (server error). These codes help developers understand API responses and troubleshoot integration issues when working with DataRobot's REST APIs.
Human in the loop (HILT)¶
Integration patterns that incorporate human oversight, validation, and intervention into AI agent workflows. Human-in-the-loop systems enable humans to review agent decisions, provide feedback, correct errors, and guide agent behavior at critical decision points. In DataRobot's agentic workflows, human-in-the-loop capabilities ensure quality control, enable learning from human expertise, and maintain human authority over sensitive or high-stakes decisions.
Humility¶
A user-defined set of rules for deployments that allow models to be capable of recognizing, in real-time, when they make uncertain predictions or receive data they have not seen before. Unlike data drift, model humility does not deal with broad statistical properties over time—it is instead triggered for individual predictions, allowing you to set desired behaviors with rules that depend on different triggers.
I¶
Image data¶
A sequence of digital images (e.g., video), a set of digital images, a single digital image, and/or one or more portions of any of these—data used as part of Visual AI. A digital image may include an organized set of picture elements ("pixels") stored in a file. Any suitable format and type of digital image file may be used, including but not limited to raster formats (e.g., TIFF, JPEG, GIF, PNG, BMP, etc.), vector formats (e.g., CGM, SVG, etc.), compound formats (e.g., EPS, PDF, PostScript, etc.), and/or stereo formats (e.g., MPO, PNS, JPS).
Image preprocessing¶
A computer vision technique, part of Visual AI. Some examples include image re-sampling, noise reduction, contrast enhancement, and scaling (e.g., generating a scale space representation). Extracted features may be:
- Low-level: raw pixels, pixel intensities, pixel colors, gradients, textures, color histograms, motion vectors, edges, lines, corners, ridges, etc.
- Mid-level: shapes, surfaces, volumes, etc.
- High-level: objects, scenes, events, etc.
Incremental learning¶
A model training method specifically tailored for large datasets—those between 10GB and 100GB—that chunks data and creates training iterations. After model building begins, compare trained iterations and optionally assign a different active version or continue training. The active iteration is the basis for other insights and is used for making predictions.
Infrastructure as Code (IaC)¶
The practice of managing and provisioning LLM and AI infrastructure through machine-readable definition files rather than physical hardware configuration or interactive configuration tools.
In-context learning¶
The ability of LLMs to learn from examples provided in the prompt without requiring fine-tuning. In-context learning allows models to adapt their behavior based on the context and examples given in the current conversation, enabling them to perform new tasks or follow specific instructions without additional training.
Inference data¶
Data that is scored by applying an algorithmic model built from a historical dataset in order to uncover practical insights. See also Scoring data.
In-sample predictions¶
Predictions made on data that the model has already seen during its training process. This typically occurs when a model is trained on a very high percentage of the available data (e.g., above 80%), leaving little or no "unseen" data for validation. In such cases, the validation score is calculated from the same data used for training, which can result in an overly optimistic assessment of model performance. In DataRobot, these scores are marked with an asterisk on the Leaderboard to indicate that they may not reflect true generalization performance. Compare to stacked (out-of-sample) predictions.
Integration patterns¶
Common architectural patterns and best practices for integrating LLM and AI services with existing systems, applications, and data sources.
Instruction tuning¶
Training LLMs to follow specific instructions or commands by fine-tuning them on instruction-response pairs. Instruction tuning improves a model's ability to understand and execute user requests, making it more useful for practical applications where following directions is important.
Irregular data¶
Data in which no consistent spacing and no time step is detected. Used in time-aware modeling.
J¶
JSON¶
A lightweight data format commonly used in DataRobot APIs for exchanging structured data between services. JSON is used throughout the DataRobot platform for configuration files, API responses, data transfer operations, and storing model metadata. The format provides a standardized way to represent complex data structures in a human-readable format that can be easily processed by applications.
K¶
KA¶
See Known in advance features.
Kernel¶
Provides programming language support to execute the code in a notebook.
Knowledge cutoff¶
The date after which an LLM's training data ends, limiting its knowledge of historical events, information, and developments that occurred after that point. Knowledge cutoff dates are important for understanding the temporal scope of a model's information and determining when additional context or real-time data sources may be needed.
Known in advance features¶
Also known as KA; used in time series modeling. A variable for which you know the value in advance and does not need to be lagged, such as holiday dates. Or, for example, you might know that a product will be on sale next week and so you can provide the pricing information in advance.
L¶
Large language model (LLM)¶
A deep learning model trained on extensive text datasets that can understand, generate, and process human language. In DataRobot's platform, LLMs form the core of LLM blueprints and can be configured with various settings, system prompts, and vector databases to create customized AI applications. These models enable DataRobot users to build intelligent chatbots, content generators, and analysis tools that can understand context and provide relevant responses.
Latency¶
The time delay between sending a request to a model or API and receiving a response, often measured in milliseconds.
Leaderboard¶
The list of trained blueprints (models) for a project, ranked according to a project metric.
Leakage¶
See target leakage.
Learning curves¶
A graph to help determine whether it is worthwhile to increase the size of a dataset. The Learning Curve graph illustrates, for the top-performing models, how model performance varies as the sample size changes.
License¶
A commercial agreement that grants access to the DataRobot platform. The license defines the scope of usage, including the number of authorized users, available features, and limits on computational resources.
Lift chart¶
Depicts how well a model segments the target population and how capable it is of predicting the target to help visualize model effectiveness.
Linkage keys¶
(Feature Discovery) The features in the primary dataset used as keys to join and create relationships.
LLM blueprint¶
The saved blueprint, available to be used for deployment. LLM blueprints represent the full context for what is needed to generate a response from an LLM; the resulting output can be compared within the playground. This information is captured in the LLM blueprint settings.
LLM blueprint components¶
The entities that make up the LLM blueprint settings, this refers to the vector database, embedding model user to generate the vector database, LLM settings, system prompt, etc. These components can either be offered natively within DataRobot or can be brought in from external sources.
LLM blueprint settings¶
The parameters sent to the LLM to generate a response (in conjunction with the user-entered prompt). They include a single LLM, LLM settings, optionally a system prompt, and optionally a vector database. If no vector database is assigned, then the LLM uses its learnings from training to generate a response. LLM blueprint settings are configurable so that you can experiment with different configurations.
LLM gateway¶
A centralized service in DataRobot that manages access to multiple large language models from external providers with support for unified authentication, rate limiting, and request routing. The LLM gateway enables organizations to standardize their interactions with various LLM providers while maintaining security, monitoring, and cost controls across all model usage.
LLM payload¶
The bundle of contents sent to the LLM endpoint to generate a response. This includes the user prompt, LLM settings, system prompt, and information retrieved from the vector database.
LLM responses¶
Generated text from the LLM based on the payload sent to an LLM endpoint.
LLM settings¶
Parameters that define how an LLM intakes a user prompt and generates a response. They can be adjusted within the LLM blueprint to alter the response. These parameters are currently represented by the "Temperature", "Token selection probability cutoff (Top P)", and "Max completion tokens" settings.
Load balancing¶
The distribution of incoming requests across multiple LLM and AI service instances to optimize resource utilization, maximize throughput, minimize response time, and avoid overload.
Location AI¶
DataRobot's support for geospatial analysis by natively ingesting common geospatial formats and recognizing coordinates, allowing ESDA, and providing spatially-explicit modeling tasks and visualizations.
Log¶
A model Leaderboard tab (Describe > Log) that displays the status of successful operations with green INFO tags, along with information about errors marked with red ERROR tags.
Log aggregation¶
The centralized collection and storage of logs from multiple LLM and AI services to enable comprehensive monitoring, analysis, and troubleshooting.
Loss function¶
A method of evaluating how well a specific algorithm models the given data. It computes a number representing the "cost" of the model's predictions being wrong; the goal of training is to minimize this value.
Low code¶
A development approach that minimizes the amount of manual coding required to build applications and workflows. Low-code platforms provide visual interfaces, drag-and-drop components, and pre-built templates that enable rapid development. In DataRobot's agentic workflows, low-code capabilities allow users to create sophisticated AI agents and workflows through configuration interfaces rather than extensive programming, making agentic AI accessible to non-technical users.
M¶
Majority class¶
If you have a categorical variable (e.g., true/false or cat/mouse ), the value that's more frequent is the majority class. For example, if a dataset has 80 rows of value cat and 20 rows of value mouse, then cat is the majority class. See also minority class.
Make Predictions tab¶
A model Leaderboard tab (Predict > Make Predictions) that allows you to make predictions before deploying a model to a production environment.
Management agent¶
A downloadable client included in the MLOps agent tarball (accessed via API keys and tools) that allows you to manage external models (i.e., those running outside of DataRobot MLOps). This tool provides a standard mechanism to automate model deployment to any type of infrastructure. The management agent sends periodic updates about deployment health and status via the API and reports them as MLOps events on the Service Health page.
Manual¶
A modeling mode that causes DataRobot to complete EDA2 and prepare data for modeling, but does not execute model building. Instead, users select specific models to build from the model Repository.
Materialized¶
Data that DataRobot has pulled from the data asset and is currently keeping a copy of in the catalog. See also snapshot and unmaterialized data.
Metadata¶
Details of the data asset, such as creation and modification dates, number and types of features, snapshot status, and more.
Metric¶
See optimization metric.
Metrics collection¶
The systematic gathering of performance, business, and operational metrics from LLM and AI systems to enable monitoring, analysis, and decision-making.
Minority class¶
If you have a categorical variable (e.g., true/false or cat/mouse ), the value that's less frequent is the minority class. For example, if a dataset has 80 rows of value cat and 20 rows of value mouse, then mouse is the minority class. See also majority class.
MLOps (Machine Learning Operations)¶
A scalable and governed means to rapidly deploy and manage ML applications in production environments.
Multi-agent flow¶
A workflow pattern where multiple AI agents collaborate to solve complex problems by dividing tasks among specialized agents. Each agent has specific capabilities and responsibilities, and they communicate and coordinate to achieve the overall objective. Multi-agent flows enable more sophisticated problem-solving by leveraging the strengths of different specialized agents. See also Agentic workflow.
MLOps agent¶
The downloadable package (tarball) that contains two clients: the Monitoring Agent and the Management Agent. The MLOps Agent enables you to monitor and manage external models (i.e., those running outside of DataRobot MLOps) by providing these tools for deployment, monitoring, and reporting. See also Monitoring Agent and Management Agent.
Model Context Protocol (MCP) server¶
A Model Context Protocol (MCP) server provides standardized interfaces for AI agents to interact with external systems and data sources. MCP servers enable secure, controlled access to tools, databases, APIs, and other resources that agents need to accomplish their tasks. In DataRobot's agentic workflows, MCP servers facilitate seamless integration between agents and enterprise systems while maintaining security and governance controls.
Model¶
A trained machine learning model that can make predictions on new data. In DataRobot, models are built using various algorithms and can predict outcomes like customer churn, sales forecasts, or fraud detection.
Model approval workflows¶
Structured processes for reviewing, validating, and approving LLM and AI models before deployment to production, ensuring quality, compliance, and business alignment.
Model catalog¶
A centralized repository that provides a comprehensive view of all available LLM and AI models, including their versions, metadata, performance metrics, and deployment status.
Model comparison¶
A Leaderboard tab that allows you to compare two models using different evaluation tools, helping identify the model that offers the highest business returns or candidates for blender models.
Model alignment¶
Techniques to ensure AI models behave according to human values and intentions. Model alignment involves training and fine-tuning processes that help models produce outputs that are helpful, honest, and harmless, reducing risks of harmful or unintended behaviors in production environments.
Model deprecation¶
The process of phasing out and retiring old LLM and AI models from production use, including communication to stakeholders and migration strategies.
Model fitting¶
A measure of how well a model generalizes similar data to the data on which it was trained. A model that is well-fitted produces more accurate outcomes. A model that is overfitted matches the data too closely. A model that is underfitted doesn't match closely enough.
Model Info¶
A model Leaderboard tab (Describe > Model Info) that displays an overview for a given model, including model file size, prediction time, and sample size.
Model lineage¶
The complete history and provenance of LLM and AI models, including their training data, algorithms, parameters, and evolution over time for audit and compliance purposes. DataRobot tracks model lineage through the Model Registry, maintaining detailed records of training data, feature engineering steps, model versions, and deployment history for comprehensive audit trails.
Model overview¶
A page within an experiment that displays the model Leaderboard, and once a model is selected, displays visualizations for that model.
Model package¶
Archived model artifacts with associated metadata stored in the Model Registry. Model packages can be created manually or automatically, for example, through the deployment of a custom model. You can deploy, share, and permanently archive model packages.
Model Registry¶
An organizational hub for the variety of models used in DataRobot. Models are registered as deployment-ready model packages; Registry lists each package available for use. Each package functions the same way, regardless of the origin of its model. The Model Registry also contains the Custom Model Workshop where you can create and deploy custom models. Model packages can be created manually or automatically depending on the type of model.
Model scoring¶
The process of applying an optimization metric to a partition of the data and assigning a numeric score that can be used to evaluate a model performance.
Model versioning¶
The systematic tracking and management of different versions of LLM and AI models to enable rollbacks, comparisons, and controlled deployments.
Modeling¶
The process of building predictive models using machine learning algorithms. This involves training algorithms on historical data to identify patterns and relationships that can be used to predict future outcomes. DataRobot automates much of this process through AutoML, allowing users to build, evaluate, and deploy predictive models efficiently.
Modeling dataset¶
A transform of the original dataset that pre-shifts data to future values, generates lagged time series features, and computes time-series analysis metadata. Commonly referred to as feature derivation, it is used by time series but not OTV. See the time series feature engineering reference for a list of operators used and feature names created by the feature derivation process. See also FEAR.
Modeling mode¶
A setting that controls the sample percentages of the training set that DataRobot uses to build models. DataRobot offers four modeling modes: Autopilot, Quick (the default), Manual, and Comprehensive.
Moderation¶
The process of monitoring and filtering model outputs to ensure they comply with safety, ethical, and policy guidelines.
Monitoring agent¶
A downloadable client included in the MLOps agent tarball (accessed via API keys and tools) that allows you to monitor external models (i.e., those running outside of DataRobot MLOps). With this functionality, predictions and information from these models can be reported as part of deployments. You can use this tool to monitor accuracy, data drift, prediction distribution, latency, and more, regardless of where the model is running.
Monotonic modeling¶
A method to force certain XGBoost models to learn only monotonic (always increasing or always decreasing) relationships between specific features and the target.
Multiclass¶
See classification.
Multilabel¶
A classification task where each row in a dataset is associated with one, several, or zero labels. Common multilabel classification problems are text categorization (a movie is both "crime" and "drama") and image categorization (an image shows a house and a car).
Multimodal¶
A model type that supports multiple var types at the same time, in the same model.
Multiseries¶
Datasets that contain multiple time series (for example, to forecast the sales of multiple stores) based on a common set of input features.
N¶
Naive model¶
See baseline model.
NAT¶
Neural Architecture Transfer (NAT) enables efficient transfer of learned representations and architectures between different AI models and tasks. NAT techniques allow agents to leverage pre-trained components and adapt them for specific use cases without full retraining. In DataRobot's agentic workflows, NAT capabilities enable rapid deployment of specialized agents by transferring knowledge from general-purpose models to domain-specific applications.
NextGen¶
The updated DataRobot user interface comprised of Workbench for experiment based iterative workflows, Registry for model evolution tracking and the centralized management of versioned models, and Console for monitoring and managing deployed models. NextGen also provides the gateway for creating agentic workflows, GenAI experiments, Notebooks, and apps.
N-gram¶
A sequence of words, where N is the number of words. For example, "machine learning" is a 2-gram. Text features are divided into n-grams to prepare for Natural Language Processing (NLP).
NIM¶
NVIDIA Inference Microservice (NIM) is a containerized AI model that provides optimized, high-performance inference with low latency and efficient resource utilization. In DataRobot's platform, a NIM can be integrated into an agentic workflow to provide specialized AI capabilities, enabling agents to leverage state-of-the-art models for specific tasks while maintaining optimal performance and scalability.
No-Code AI Apps¶
A no-code interface to create AI-powered applications that enable core DataRobot services without having to build models and evaluate their performance. Applications are easily shared and do not require consumers to own full DataRobot licenses in order to use them.
Notebook¶
An interactive, computational environment that hosts code execution and rich media. DataRobot provides its own in-app environment to create, manage, and execute Jupyter-compatible hosted notebooks.
Nowcasting¶
A method of time series modeling that predicts the current value of a target based on past and present data. Technically, it is a forecast window in which the start and end times are 0 (now).
O¶
Offset¶
Feature(s) that should be treated as a fixed component for modeling (coefficient of 1 in generalized linear models or gradient boosting machine models). Offsets are often used to incorporate pricing constraints or to boost existing models.
One-shot learning¶
A capability of a model to learn to perform a task from only a single example.
Orchestration¶
The coordination of multiple AI components, tools, and workflows to achieve complex objectives. Orchestration involves managing the flow of data and control between different AI services, ensuring proper sequencing, error handling, and resource allocation. In DataRobot, orchestration enables the creation of sophisticated multi-step AI workflows that combine various capabilities and tools.
Parameter efficient fine-tuning (PEFT)¶
Methods to fine-tune large models using fewer parameters than full fine-tuning. PEFT techniques, such as LoRA (Low-Rank Adaptation) and adapter layers, allow for efficient model customization while maintaining most of the original model's performance and reducing computational requirements.
Operation¶
A single data manipulation instruction that specifies to either transform, filter, or pivot one or more records into zero or more records (e.g., find and replace or compute new feature).
Optimization metric¶
An error metric used in DataRobot to determine how well a model predicts actual values. After you choose a target feature, DataRobot selects an optimization metric based on the modeling task.
Ordering feature¶
The primary date/time feature that DataRobot will use for modeling. Options are detected during EDA1.
Organization¶
A top-level entity in DataRobot that represents a single customer or tenant. It serves as a container for all users, groups, projects, deployments, and other assets, enabling centralized billing and resource management.
OTV¶
Also known as out-of-time validation. A method for modeling time-relevant data. With OTV you are not forecasting, as with time series. Instead, you are predicting the target value on each individual row.
Overfitting¶
A modeling issue where predictive models perform exceptionally well on training data but poorly on new data. DataRobot addresses overfitting through automated techniques like regularization, early stopping, and cross-validation. The platform's built-in safeguards help prevent overfitting by monitoring validation performance and automatically adjusting model complexity, ensuring models generalize well to unseen data while maintaining predictive accuracy.
P¶
Partition¶
Segments of training data, broken down to maximize accuracy. The segments (splits) of the dataset. See also training, validation, cross-validation, and holdout.
Per-class bias¶
A model Leadboard tab (Bias and Fairness > Per-Class Bias) that helps to identify if a model is biased, and if so, how much and who it's biased towards or against. Bias and Fairness settings must be configured.
Permissions¶
A set of rights that control what actions a user or group can perform within DataRobot. Permissions are managed through roles and determine access to features like creating projects, deploying models, and managing system settings.
PID (project identifier)¶
An internal identifier used for uniquely identifying a project.
PII¶
Personal identifiable information, including name, pictures, home address, SSN or other identifying numbers, birth date, and more. DataRobot automates the detection of specific types of personal data to provide a layer of protection against the inadvertent inclusion of this information in a dataset.
Pipeline¶
A sequence of data processing and modeling steps, often automated, that transforms raw data into predictions or insights.
Playground¶
The place where you create and interact with LLM blueprints (LLMs and their associated settings), comparing the response of each to help determine which to use in production. Many LLM blueprints can live within a playground. A playground is an asset of a Use Case; multiple playgrounds can exist in a single Use Case.
Playground compare¶
The place to add LLM blueprints to the playground for comparison, submit prompts to these LLM blueprints, and evaluate the rendered responses. With RAG, a single prompt is sent to an LLM to generate a single response, without referencing previous prompts. This allows users to compare responses from multiple LLM blueprints.
Port¶
An interface that connects a DataRobot entity (a notebook, custom model, or custom app) to another network.
Portable Prediction Server (PPS)¶
A DataRobot execution environment for DataRobot model packages (.mlpkg files) distributed as a self-contained Docker image. It can be run disconnected from main installation environments.
Predicting¶
For non-time-series modeling. Use information in a row to determine the target for that row. Prediction uses explanatory variables to characterize expected outcomes or expected responses (e.g., a specific event in the future, gender, fraudulent transactions).
Prediction data¶
Data that contains prediction requests and results from the model.
Prediction environment¶
An environment configured to manage deployment predictions on an external system, outside of DataRobot. Prediction environments allow you to configure deployment permissions and approval processes. Once configured, you can specify a prediction environment for use by DataRobot models running on the Portable Prediction Server and for remote models monitored by the MLOps monitoring agent.
Prediction explanations¶
A visualization that helps to illustrate what drives predictions on a row-by-row basis—they provide a quantitative indicator of the effect variables have on a model, answering why a given model made a certain prediction. It helps to understand why a model made a particular prediction so that you can then validate whether the prediction makes sense. See also SHAP, XEMP.
Prediction intervals¶
Prediction intervals help DataRobot assess and describe the uncertainty in a single record prediction by including an upper and lower bound on a point estimate (e.g., a single prediction from a machine learning model). The prediction intervals provide a probable range of values that the target may fall into on future data points.
Prediction point¶
The point in time when you made or will make a prediction. Plan your prediction point based on the production model (for example, "one month before renewal" or "loan application submission time"). Once defined, create that entry in the training data to help avoid lookahead bias. With Feature Discovery, you define the prediction point to ensure the derived features only use data prior to that point.
Prediction server¶
The dedicated, scalable infrastructure responsible for hosting deployed models and serving real-time prediction requests via an API. It is optimized for low-latency and high-throughput scoring.
Prepared dataset¶
A dataset that has been materialized in its source after publishing a recipe.
Primary dataset¶
(Feature Discovery) The dataset used to start a project.
Primary features¶
(Feature Discovery) Features in the project's primary dataset.
Privacy controls¶
Mechanisms and policies for managing personal data in LLM and AI systems, including data anonymization, consent management, and compliance with privacy regulations.
Project¶
A referenceable item that includes a dataset, which is the source used for training, and any models built from the dataset. Projects can be created and accessed from the home page, the project control center, and the AI Catalog. They can be shared to users, groups, and an organization.
Prompt¶
The input entered during chatting used to generate the LLM response.
Prompt engineering¶
The practice of designing and refining input prompts to guide a language model toward producing desired outputs.
Prompt injection¶
A security vulnerability where malicious prompts can override system instructions or safety measures. Prompt injection attacks attempt to manipulate AI systems into generating inappropriate content or performing unintended actions by crafting inputs that bypass the model's intended constraints and guidelines.
Prompt template¶
See system prompt.
Pulumi¶
Infrastructure as Code (IaC) platform that enables developers to define and manage cloud infrastructure using familiar programming languages. Pulumi supports multiple cloud providers and provides a unified approach to infrastructure management. In DataRobot's agentic workflows, Pulumi enables automated provisioning and management of infrastructure resources needed for agent deployment, scaling, and monitoring across different environments.
Protected class¶
One categorical value of the protected feature, used in bias and fairness modeling.
Protected feature¶
The dataset column to measure fairness of model predictions against. Model fairness is calculated against the protected features from the dataset. Also known as "protected attribute."
Publish¶
Execution of the sequence of operations specified in a recipe resulting in the materialization of a dataset in a data source.
Q¶
Queue¶
The system that manages the execution of jobs, such as model training and batch predictions. The queue prioritizes and allocates tasks to available workers based on system load and user permissions, ensuring efficient use of computational resources.
Quick (Autopilot)¶
A shortened version of the full Autopilot modeling mode that runs models directly at 64%. With Quick, the 16% and 32% sample sizes are not executed. DataRobot selects models to run based on a variety of criteria, including target and performance metric, but as its name suggests, chooses only models with relatively short training runtimes to support quicker experimentation.
R¶
Rate limiting¶
A technique used to control the number of requests a client can make to an API within a specified time period, preventing abuse and ensuring fair usage.
Rating table¶
A model Leaderboard tab (Describe > Rating Table) where you can export the model's complete, validated parameters.
Real-time predictions¶
Method of making predictions when low latency is required. Use the Prediction API for real-time deployment predictions on a dedicated and/or a standalone prediction server.
Receiver Operating Characteristic Curve¶
See ROC Curve.
Recipe¶
A user-defined sequence of transformation operations that are applied to the data. A recipe is uniquely identified and versioned by the system. It includes metadata identifying the input data's source and schema, the output data's schema, the Use Case Container ID, and user ID.
Registry¶
Registry is a centralized location where you access versioned, deployment-ready model packages. From there, you can create custom models and jobs, generate compliance documentation, and deploy models to production.
Regression¶
A DataRobot modeling approach that predicts continuous numerical values from your target feature. DataRobot's regression capabilities handle various continuous outcomes like sales forecasts, price predictions, or risk scores. The platform automatically selects from regression algorithms in the Repository and provides evaluation metrics like RMSE, MAE, and R² to measure prediction accuracy. See also classification.
Regularization¶
A technique used to prevent model overfitting by adding a penalty term to the loss function. Common types are L1 (Lasso) and L2 (Ridge) regularization.
Regular data¶
Data is regular if rows in the dataset fall on an evenly spaced time grid (e.g., there's one row for every hour across the entire dataset). See also time step and semi-regular data.
Reinforcement learning from human feedback (RLHF)¶
A training method that uses human feedback to improve model behavior. RLHF involves collecting human preferences on model outputs and using reinforcement learning techniques to fine-tune the model to produce responses that align with human values and preferences, improving safety and usefulness.
ReAct¶
A Reasoning and Acting (ReAct) framework combines reasoning capabilities with action execution in AI agents. ReAct enables agents to think through problems step-by-step, plan actions, execute them, and observe results to inform subsequent reasoning. In DataRobot's agentic workflows, ReAct capabilities allow agents to perform complex problem-solving by iteratively reasoning about situations, taking actions, and learning from outcomes to achieve their goals.
Relationships¶
(Feature Discovery) Relationships between datasets. Each relationship involves a pair of datasets, and a join key from each dataset. A key comprises one or more columns of a dataset. The keys from both datasets are ordered, and must have the same number of columns. The combination of keys is used to determine how two datasets are joined.
Remote models¶
Models running outside of DataRobot in external prediction environments, often monitored by the MLOps monitoring agent to report statistics back to DataRobot.
Repository¶
A library of modeling blueprints available for a selected project (based on the problem type). These models may be selected and built by DataRobot and also can be user-executed.
Resource optimization¶
The practice of optimizing LLM and AI resource usage for cost efficiency while maintaining performance and reliability requirements.
Resource provisioning¶
The allocation and management of computing resources (CPU, memory, storage, GPU) for LLM and AI workloads to ensure optimal performance and cost efficiency.
Response time optimization¶
Techniques and strategies for improving LLM response times, including caching, model optimization, and infrastructure improvements.
Retrieval¶
The process of finding relevant information from a knowledge base or database. In the context of RAG workflows, retrieval involves searching through vector databases or other knowledge sources to find the most relevant content that can be used to ground and inform AI responses, improving accuracy and reducing hallucination.
Retrieval Augmented Generation (RAG)¶
The process of sending a payload to an LLM that contains the prompt, system prompt, LLM settings, vector database (or subset of vector database), and the LLM returning corresponding text based on this payload. It includes the process of retrieving relevant information from a vector database and sending that along with the prompt, system prompt, and LLM settings to the LLM endpoint to generate a response grounded in the data in the vector database. This operation may optionally also incorporate orchestration to execute a chain of multiple prompts.
Retrieval Augmented Generation (RAG) workflow¶
An AI system that runs RAG, which includes data preparation, vector database creation, LLM configuration, and response generation. RAG workflows typically involve steps such as document chunking, embedding generation, similarity search, and context-aware response generation, all orchestrated to provide accurate, grounded responses to user queries. See also Retrieval Augmented Generation (RAG).
REST (Representational State Transfer)¶
An architectural style for designing networked applications, commonly used for web APIs, that uses standard HTTP methods (GET, POST, PUT, DELETE) to access and manipulate resources.
ROC Curve¶
Also known as Receiver Operating Characteristic Curve. A visualization that helps to explore classification, performance, and statistics related to a selected model at any point on the probability scale. In DataRobot, the visualization is available from the Leaderboard.
Role¶
Roles—Owner, Consumer, and Editor—describe the capabilities provided to each user for a given dataset. This supports the scenarios when the user creating a data source or data connection and the enduser are not the same, or there are multiple endusers of the asset.
Role-based access control (RBAC)¶
A security model that restricts access to LLM and AI systems based on the roles of individual users, providing granular permission management and security control. DataRobot implements RBAC through user groups, permissions, and organization-level access controls to ensure secure and appropriate access to features and assets across the platform.
S¶
Sample¶
The process of selecting a subset of data from a larger dataset for analysis, modeling, or preview purposes. DataRobot samples data in various contexts:
- EDA1 sampling: DataRobot samples up to 500MB of data for initial exploratory data analysis. If the dataset is under 500MB, it uses the entire dataset; otherwise, it uses a 500MB random sample.
- Live sample: During data wrangling, DataRobot retrieves a configurable number of rows (default 10,000) using different sampling methods (Random, First-N Rows, or Date/time for time series data) to provide interactive preview and analysis capabilities.
- Feature Impact sampling: For calculating feature importance, DataRobot samples training records (default 2,500 rows, maximum 100,000) using different sampling strategies based on data characteristics (random sampling for balanced data, smart downsampling for imbalanced data).
- Model evaluation sampling: Various model insights and evaluations use sampled data to balance computational efficiency with statistical accuracy.
Sample size¶
The percentage of the total training data used to build models. The percentage is based on the selected modeling mode or can be user-selected.
Scoring¶
See Model scoring, Scoring data.
Scoring code¶
A method for using DataRobot models outside of the application. It is available for select models from the Leaderboard as a downloadable JAR file containing Java code that can be used to score data from the command line.
An exportable JAR file, available for select models, that runs in Java. Scoring Code JARs contain prediction calculation logic identical to the DataRobot API—the code generation mechanism tests each model for accuracy as a part of the generation process.
Scoring data¶
The dataset provided to a deployed model to generate predictions. This is also known as inference data. For example, to predict housing prices, the scoring data would be a file containing new listings with all the model's required features (e.g., square footage, number of bedrooms) but without the final price.
SDK (Software Development Kit)¶
A collection of tools and libraries provided by a hardware or software vendor to enable developers to create applications for a specific platform. (e.g., the DataRobot Python SDK).
Seasonality¶
Repeating highs and lows observed at different times of year, within a week, day, etc. Periodicity. For example, temperature is very seasonal (hot in the summer, cold in the winter, hot during the day, cold at night). Applicable to time series modeling.
Secondary dataset¶
(Feature Discovery) A dataset that is added to a project and part of a relationship with the primary dataset.
Secondary features¶
(Feature Discovery) Features derived from a project's secondary datasets.
Secure Single Sign-On Protocol (SSSOP)¶
Secure Single Sign-On Protocol (SSSOP) that provides authentication and authorization services for AI agents and workflows. SSSOP ensures secure access control across distributed agent systems while maintaining user privacy and session management. In DataRobot's agentic platform, SSSOP enables seamless authentication for agents accessing external systems and provides audit trails for compliance and security monitoring.
Segmented analysis¶
A deployment utility that filters data drift and accuracy statistics into unique segment attributes and values. Useful for identifying operational issues with training and prediction request data.
Segmented modeling¶
A method of modeling multiseries projects by generating a model for each segment. DataRobot selects the best model for each segment (the segment champion) and includes the segment champions in a single Combined Model that you can deploy.
Segment ID¶
A column in a dataset used to group series into segments for a multiseries project. A segment ID is required for the segmented modeling workflow, where DataRobot builds a separate model for each segment. See also Segmented modeling.
Semantic layer¶
A semantic layer is a business representation of the source data that maps complex data to common business terms, helping you more easily understand what the data means and the information it represents.
Semantic memory¶
Memory systems that store general knowledge, facts, concepts, and relationships that are not tied to specific experiences. Semantic memory enables AI agents to maintain domain knowledge, understand concepts, and apply general principles to new situations. In DataRobot's agentic workflows, semantic memory allows agents to maintain knowledge about business processes, domain expertise, and general problem-solving strategies.
Semantic search¶
Search method that finds content based on meaning rather than exact keyword matches. Semantic search uses vector embeddings to understand the intent and context of queries, enabling more accurate and relevant results even when the exact words don't match. This approach is particularly useful in RAG systems for finding the most relevant information to ground AI responses.
Short-term memory¶
Temporary storage systems that AI agents use to maintain context and information during active task execution. Short-term memory enables agents to remember recent interactions, maintain conversation context, and track progress on current tasks. In DataRobot's agentic workflows, short-term memory allows agents to maintain coherence across multi-step processes and provides continuity in user interactions.
Long-term memory¶
Persistent storage systems that AI agents use to retain knowledge, experiences, and learned patterns across multiple sessions and tasks. Long-term memory enables agents to build upon previous experiences, maintain learned behaviors, and accumulate domain knowledge over time. In DataRobot's agentic workflows, long-term memory allows agents to improve performance through experience and maintain consistency across different use cases.
Semi-regular data¶
Data is semi-regular if most time steps are regular but there are some small gaps (e.g., business days, but no weekends). See also regular data and time steps.
Series ID¶
A column in a dataset used to divide a dataset into series for a multiseries project. The column contains labels indicating which series each row belongs to. See also Multiseries modeling.
Service health¶
A performance monitoring component for deployments that tracks metrics about a deployment's ability to respond to prediction requests quickly and reliably. Useful for identifying bottlenecks and assessing prediction capacity.
Service mesh¶
A dedicated infrastructure layer for managing communication between LLM and AI microservices, providing features like load balancing, service discovery, and security. Service meshes enable fine-grained control over service-to-service communication, including traffic management, observability, and policy enforcement for complex AI application architectures.
Streaming¶
Real-time generation of text where output is displayed as it's being generated. Streaming provides immediate feedback to users by showing AI responses as they are produced, rather than waiting for the complete response. This approach improves user experience by reducing perceived latency and allowing users to see progress in real-time.
Single agent flow¶
A workflow pattern where a single AI agent handles all aspects of a task from start to finish. The agent receives input, processes it through its capabilities, and produces output without requiring coordination with other agents. Single agent flows are suitable for straightforward tasks that can be completed by one specialized agent.
SHAP (Shapley Values)¶
A fast, open-source methodology for computing Prediction Explanations for tree-based, deep learning, and linear-based models. SHAP estimates how much each feature contributes to a given prediction differing from the average. It is additive, making it easy to see how much top-N features contribute to a prediction. See also Prediction Explanations, XEMP.
Sidecar model¶
A structural component that supports the LLM that is serving back responses. It helps to make determinations about whether a prompt was toxic, an injection attack, etc. With DataRobot, it uses hosted custom metrics to accomplish the monitoring.
Single Sign-On (SSO)¶
An authentication method that allows users to log in to DataRobot using their existing corporate identity provider (e.g., Okta, Azure AD). SSO simplifies user access by eliminating the need for separate DataRobot-specific credentials.
Slim run¶
A technique to improve time and memory use, slim runs apply to datasets exceeding 800MB. When triggered, models do no calculate an internal cross-validation and so do not have stacked predictions.
Smart downsampling¶
A technique to reduce total dataset size by reducing the size of the majority class, enabling you to build models faster without sacrificing accuracy. When enabled, all analysis and model building is based on the new dataset size after smart downsampling.
Snapshot¶
An asset created from a data source. For example, with a database it represents either the entire database or a selection of (potentially joined) tables, taken at a particular point in time. It is taken from a live database but creates a static, read-only copy of data. DataRobot creates a snapshot of each data asset type, while allowing you to disable the snapshot when importing the data.
Speed vs accuracy¶
A Leaderboard tab that generates an analysis plot to show the tradeoff between runtime and predictive accuracy and help you choose the best model with the lowest overhead.
Stability¶
A model Leaderboard tab (Evaluate > Stability) that provides an at-a-glance summary of how well a model performs on different backtests. The backtesting information in this chart is the same as that available from the Model Info tab.
Stacked predictions¶
A method for building multiple models on different subsets of the data. The prediction for any row is made using a model that excluded that data from training. In this way, each prediction is effectively an "out-of-sample" prediction. See an example in the predictions documentation. Compare to "in-sample" predictions. See also slim run.
Stationarity¶
The mean of the series does not change over time. A stationary series does not have a trend or seasonal variation. Applicable to time series modeling. See also trend.
Stop sequence¶
A specific token or set of tokens that signals a language model to stop generating further output.
Supervised learning¶
Predictive modeling approach where your dataset includes a target feature with known values for training. DataRobot uses this labeled data to automatically build models that learn relationships between features and the target, enabling predictions on new data. This approach powers DataRobot's classification and regression projects, with the platform handling algorithm selection, hyperparameter tuning, and model evaluation automatically. See also unsupervised learning.
Syftr¶
A specialized agent framework component that provides secure, privacy-preserving data processing capabilities for AI agents. Syftr enables agents to work with sensitive data while maintaining confidentiality and compliance with privacy regulations. In DataRobot's agentic workflows, Syftr components allow agents to process encrypted or anonymized data, perform federated learning, and maintain data privacy throughout the agent lifecycle.
System prompt¶
The system prompt, an optional field, is a "universal" prompt prepended to all individual prompts. It instructs and formats the LLM response. The system prompt can impact the structure, tone, format, and content that is created during the generation of the response.
T¶
Target¶
The name of the column in the dataset that you would like to predict.
Target leakage¶
An outcome when using a feature whose value cannot be known at the time of prediction (for example, using the value for "churn reason" from the training dataset to predict whether a customer will churn). Including the feature in the model's feature list would incorrectly influence the prediction and can lead to overly optimistic models.
Task¶
An ML method, for example a data transformation such as one-hot encoding, or an estimation such as an XGBoost classifier, which is used to define a blueprint. There are hundreds of built-in tasks you can use, or you can define your own (custom) tasks.
Temperature¶
A parameter that controls the creativity and randomness of LLM responses. Lower temperature values (0.1-0.3) produce more focused, consistent outputs suitable for factual responses, while higher values (0.7-1.0) generate more creative and diverse content. DataRobot's playground interface allows you to experiment with different temperature values in LLM blueprint settings to find the optimal balance for your specific use case.
Terminal¶
A text-based interface used to interact with a server by entering commands.
Template¶
Pre-configured frameworks or structures that provide a starting point for creating agentic workflows, applications, or configurations. Templates in DataRobot include predefined agent configurations, workflow patterns, and code structures that accelerate development and ensure best practices. Templates can include agent goals, tool configurations, guardrails, and integration patterns, allowing users to quickly deploy sophisticated agentic systems without starting from scratch.
Throughput¶
The number of requests or predictions a system can process in a given period, often measured as requests per second (RPS) or tokens per second for LLMs.
Time-aware predictions¶
Assigns rows to backtests chronologically and makes row-by-row predictions. This method provides no feature engineering and can be used when forecasting is not needed.
Time-aware predictions with feature engineering¶
Assigns rows by forecast distance, builds separate models for each distance, and then makes row-by-row predictions. This method is best when combined with time-aware wrangling, which provides transparent and flexible feature engineering. Use when forecasting is not needed, but predictions based on forecast distance and full transparency of the transformation process is desired.
Time-aware wrangling¶
Perform time series feature engineering during the data preparation phase by creating recipes of operations and applying them first to a sample and then, when verified, to a full dataset—time-aware data. This method provides control over which time-based features are generated before modeling to allow adjustment before publishing, preventing the need to rerun modeling if what would otherwise be done automatically doesn't fit the use case.
Time series¶
A series of data points indexed in time order—ordinarily a sequence of measurements taken at successive, equally spaced intervals. Time series modeling is a recommended practice for data science problems where conditions may change over time.
Time series analysis¶
Methods for analyzing time series data in order to extract meaningful statistics and other characteristics of the data.
Time series forecasting¶
The use of a model to predict future values based on previously observed values. In practice, a forecasting model may combine time series features with other data.
Time step¶
The detected median time delta between rows in the time series; DataRobot determines the time unit. The time step consists of a number and a time-delta unit, for example (15, "minutes"). If a step isn't detected, the dataset is considered irregular and time series mode may be disabled. See also regular data and semi-regular data.
Token¶
The smallest unit of text that LLMs process when parsing prompts/generating responses. In DataRobot's platform, tokens are used to measure input/output size of chats and calculate usage costs for LLM operations. When you send prompts to LLM blueprints, the system tokenizes your text and tracks consumption for billing and performance monitoring. Token usage is displayed in DataRobot's playground and deployment interfaces to help you optimize costs and stay within platform limits.
Token usage¶
The number of tokens consumed by an LLM for input and output, often used for billing and cost management. Token usage is a key metric for understanding the computational cost of AI operations, as most LLM providers charge based on the number of tokens processed. Monitoring token usage helps optimize costs and resource allocation in AI applications.
Token usage tracking¶
The monitoring and recording of LLM token consumption to track costs, usage patterns, and optimize resource allocation. DataRobot provides token usage analytics and cost management features to help organizations monitor and control their LLM API expenses across different models and deployments.
Tokenization¶
The process of breaking text into smaller units called tokens, which can be words, subwords, or characters, for processing by a language model.
Tool¶
A software component or service that provides specific functionality to AI agents or workflows. Tools can perform various tasks such as data retrieval, computation, API calls, or specialized processing. In DataRobot's agentic workflows, tools are modular components that agents can invoke to extend their capabilities and perform complex operations beyond their core functionality.
Toolkit¶
A collection of tools, utilities, and resources designed to support the development and deployment of agentic AI systems. Toolkits provide standardized interfaces, common functionality, and best practices for building AI agents. In DataRobot's platform, toolkits include pre-built tools for data processing, model training, API integration, and workflow orchestration, enabling rapid development of sophisticated agentic applications.
Top-k¶
A decoding parameter that limits the model's next-token choices to the k most likely options, sampling from only those candidates to generate more focused or creative responses.
Top-p (nucleus sampling)¶
A decoding parameter that limits the model's next-token choices to the smallest set whose cumulative probability exceeds a threshold p, allowing for dynamic selection of likely tokens.
Toxicity¶
The presence of harmful, offensive, or inappropriate language in model outputs, which safety and moderation systems aim to detect and prevent.
Tracking agent¶
See MLOps agent.
Training¶
The process of building models on data in which the target is known.
Training dashboard¶
A model Leaderboard tab (Evaluate > Training dashboard) that provides, for each executed iteration, information about a model's training and test loss, accuracy, learning rate, and momentum to help you get a better understanding about what may have happened during model training.
Training data¶
The portion (partition) of data used to build models. See also validation, cross-validation, and holdout.
Transfer learning¶
A project training on one dataset, extracting information that may be useful, and applying that learning to another.
Trend¶
An increase or decrease over time. Trends can be linear or non-linear and can show fluctuation. A series with a trend is not stationary.
Tuning¶
A trial-and-error process by which you change some hyperparameters, run the algorithm on the data again, then compare performance to determine which set of hyperparameters results in the most accurate model. In DataRobot, this functionality is available from the Advanced Tuning tab.
U¶
Unit of analysis¶
(Machine learning) The unit of observation at which you are making a prediction.
Unlimited multiclass¶
See classification.
Unmaterialized¶
Data that DataRobot samples for profile statistics, but does not keep. Instead, the catalog stores a pointer to the data and only pulls it upon user request at project start or when running batch predictions. See also materialized data.
Unstructured text¶
Text that cannot fit cleanly into a table. The most typical example is large blocks of text typically in some kind of document or form.
Unsupervised learning¶
A DataRobot modeling approach to discovering patterns in datasets without requiring a target feature. DataRobot offers unsupervised learning through anomaly detection projects, which identify unusual data points, and clustering projects, which group similar records together. These capabilities help users explore data structure, identify outliers, and segment populations without needing labeled training data. DataRobot automatically selects appropriate unsupervised algorithms and provides visualizations to interpret results. See also supervised learning.
Use Case¶
A container that groups objects that are part of the Workbench experimentation flow.
User¶
A DataRobot account that can be assigned to a specific user. Users can be assigned to one or more organizations and have specific permissions within those organizations.
User blueprint¶
A blueprint (and extra metadata) that has been created by a user and saved to the AI Catalog, where it can be both shared and further modified. This is not the same as a blueprint available from the Repository or via models on the Leaderboard, though both can be used as the basis for creation of a user blueprint. See also blueprint.
V¶
Validation¶
The validation (or testing) partition is a subsection of data that is withheld from training and used to evaluate a model's performance. Since this data was not used to build the model, it can provide an unbiased estimate of a model's accuracy. You often compare the results of validation when selecting a model. See also cross-validation.
Variable¶
See feature.
Variance (Statistical)¶
The variability of model prediction for a given data point. High-variance models are often too complex and are sensitive to the specific data they were trained on, leading to overfitting.
Vector database¶
A specialized database that stores text chunks alongside their numerical representations (embeddings) for efficient similarity search. In DataRobot's platform, vector databases enable RAG operations by allowing LLM blueprints to retrieve relevant information from large document collections. When you upload documents to DataRobot, the system automatically chunks the text, generates embeddings, and stores them in a vector database that can be connected to LLM blueprints for grounded, accurate responses based on your specific content.
Visual AI¶
DataRobot's ability to combine supported image types, either alone or in combination with other supported feature types, to create models that use images as input. The feature also includes specialized insights (e.g., image embeddings, attention maps, neural network visualizer) to help visually assess model performance.
W¶
Word cloud¶
A model Leaderboard tab (Understand > Word Cloud) that displays the most relevant words and short phrases in word cloud format.
Workbench¶
Workbench is an experiment-based user interface optimized to support iterative workflow. It enables users to group and share everything they need to solve a specific problem from a single location. Workbench is organized by Use Case, and each Use Case contains zero or more datasets, vector databases, playgrounds, models, notebooks, and applications. Workbench is the new generation of DataRobot Classic.
Worker¶
The processing power behind the DataRobot platform, used for creating projects, training models, and making predictions. They represent the portion of processing power allocated to a task. DataRobot uses different types of workers for different phases of the project workflow, including DSS workers (Dataset Service workers), EDA workers, secure modeling workers, and quick workers.
Wrangle¶
A capability that enables you to import, explore, and transform data in an easy-to-use GUI environment.
Webhook¶
A user-defined HTTP callback that allows one system to send real-time data to another system when a specific event occurs.
X¶
XEMP (eXemplar-based Explanations of Model Predictions)¶
A methodology for computing Prediction Explanations that works for all models. See also Prediction Explanations, SHAP.
Y¶
YAML¶
A human-readable configuration format used in DataRobot for defining model parameters, deployment settings, and workflow configurations. YAML files are commonly used in DataRobot projects to specify custom model environments, deployment configurations, and automation workflows, providing a clear and structured way to manage complex settings.
Z¶
Z score¶
A metric measuring whether a given class of the protected feature is "statistically significant" across the population. used in bias and fairness modeling.
Zero-shot learning¶
A capability of a model to perform a task without having seen any examples of that task during training, relying on generalization from related knowledge.