# Confusion matrix

> Confusion matrix - The confusion matrix available on the DataRobot ROC Curve tab lets you evaluate
> accuracy by comparing actual versus predicted values.

This Markdown file sits beside the HTML page at the same path (with a `.md` suffix). It summarizes the topic and lists links for tools and LLM context.

Companion generated at `2026-04-24T16:03:56.585721+00:00` (UTC).

## Primary page

- [Confusion matrix](https://docs.datarobot.com/en/docs/classic-ui/modeling/analyze-models/evaluate/roc-curve-tab/confusion-matrix-classic.html): Full documentation for this topic (HTML).

## Sections on this page

- [Analyze the confusion matrix](https://docs.datarobot.com/en/docs/classic-ui/modeling/analyze-models/evaluate/roc-curve-tab/confusion-matrix-classic.html#analyze-the-confusion-matrix): In-page section heading.
- [Count differences](https://docs.datarobot.com/en/docs/classic-ui/modeling/analyze-models/evaluate/roc-curve-tab/confusion-matrix-classic.html#count-differences): In-page section heading.

## Related documentation

- [Classic UI documentation](https://docs.datarobot.com/en/docs/classic-ui/index.html): Linked from this page.
- [Modeling](https://docs.datarobot.com/en/docs/classic-ui/modeling/index.html): Linked from this page.
- [Model insights](https://docs.datarobot.com/en/docs/classic-ui/modeling/analyze-models/index.html): Linked from this page.
- [Evaluate](https://docs.datarobot.com/en/docs/classic-ui/modeling/analyze-models/evaluate/index.html): Linked from this page.
- [ROC Curve tools](https://docs.datarobot.com/en/docs/classic-ui/modeling/analyze-models/evaluate/roc-curve-tab/index.html): Linked from this page.
- [ROC Curve](https://docs.datarobot.com/en/docs/classic-ui/modeling/analyze-models/evaluate/roc-curve-tab/roc-curve-tab-use.html): Linked from this page.
- [data source](https://docs.datarobot.com/en/docs/classic-ui/modeling/analyze-models/evaluate/roc-curve-tab/threshold.html#select-data-for-visualizations): Linked from this page.
- [Prediction Distribution graph](https://docs.datarobot.com/en/docs/classic-ui/modeling/analyze-models/evaluate/roc-curve-tab/pred-dist-graph.html): Linked from this page.
- [here](https://docs.datarobot.com/en/docs/classic-ui/modeling/analyze-models/evaluate/roc-curve-tab/metrics-classic.html#metrics-explained): Linked from this page.
- [smart downsampling](https://docs.datarobot.com/en/docs/classic-ui/modeling/build-models/adv-opt/smart-ds.html): Linked from this page.
- [Classic](https://docs.datarobot.com/en/docs/classic-ui/modeling/build-models/adv-opt/additional.html): Linked from this page.
- [NextGen](https://docs.datarobot.com/en/docs/workbench/nxt-workbench/experiments/create-experiments/create-predictive/ml-adv-experiment.html#weight): Linked from this page.
- [Accuracy optimization metric](https://docs.datarobot.com/en/docs/reference/pred-ai-ref/opt-metric.html#accuracybalanced-accuracy): Linked from this page.

## Documentation content

# Confusion matrix

The [ROC Curve](https://docs.datarobot.com/en/docs/classic-ui/modeling/analyze-models/evaluate/roc-curve-tab/roc-curve-tab-use.html) tab provides a confusion matrix that lets you evaluate accuracy by comparing actual versus predicted values. The confusion matrix is a table that reports true versus predicted values. The name “confusion matrix” is used because the matrix shows whether the model is confusing two classes (consistently mislabeling one class as another class).

The confusion matrix facilitates more detailed analysis than relying on accuracy alone. Accuracy yields misleading results if the dataset is unbalanced (great variation in the number of samples in different classes), so it is not always a reliable metric for the real performance of a classifier.

To evaluate accuracy using the confusion matrix:

1. Select a model on the Leaderboard and navigate toEvaluate > ROC Curve.
2. Select adata sourceand set thedisplay threshold. The confusion matrix displays on the right side of the ROC Curve tab.

## Analyze the confusion matrix

The rows and columns of the confusion matrix below report true and false values of the [hospital readmission classification use case](https://docs.datarobot.com/en/docs/classic-ui/modeling/analyze-models/evaluate/roc-curve-tab/roc-curve-tab-use.html#classification-use-case-2).

> [!NOTE] Note
> The [Prediction Distribution graph](https://docs.datarobot.com/en/docs/classic-ui/modeling/analyze-models/evaluate/roc-curve-tab/pred-dist-graph.html) uses these same values and definitions.

- Each column of the matrix represents the instances in apredictedclass (predicted not readmitted, predicted readmitted).
- Each row represents the instances in anactualclass (actually not readmitted, actually readmitted). If you look at theActualaxis on the left in the example above,Truecorresponds to the blue row and represents the positive class (1 or readmitted), whileFalsecorresponds to the red row and represents the negative class (0 or not readmitted).
- Total correct predictions are comprised of TP + TN; total incorrect predictions are comprised of FP + FN. In the sample matrix above: ValueModel predictionTrue Negative (TN)459 patients predicted to not readmit that actually did not readmit. Correctly predicted False when actual was False.False Positive (FP)506 patients predicted to readmit, but actually did not readmit.  Incorrectly predicted True when actual was False.False Negative (FN)123 patients predicted to not readmit, but actually did readmit. Incorrectly predicted False when actual was True.True Positive (TP)512 patients predicted to readmit that actually readmitted. Correctly predicted True when actual was True.
- The matrix displays totals by row and column:
- To view total counts, hover over a cell in the matrix. The tooltip shows the marginal totals (row and column sums) rather than just the individual cell value. In this example, the predicted false values were 459 + 123, even though 123 were actually true. Use these values to help understand the distribution of your data:
- The key metrics you can derive from the matrix are calculated as follows (and explainedhere): AccuracyCalculationAccuracy(459 + 512) ÷ 1600 = 60.7% overall correct predictions.Precision512 ÷ (512 + 506) = 50.3% of positive predictions were correct.Recall/Sensitivity512 ÷ (512 + 123) = 80.6% of actual positives were caught.Specificity459 ÷ (459 + 506) = 47.6% of actual negatives were correctly identified.

### Count differences

When [smart downsampling](https://docs.datarobot.com/en/docs/classic-ui/modeling/build-models/adv-opt/smart-ds.html) is enabled, the confusion matrix totals may differ slightly from the size of the data partitions (validation, cross-validation, and holdout). This is largely due to a rounding error. In actuality, rows from the minority class are always assigned a "weight" of 1 (not to be confused with the weight set in Advanced options and therefore never removed during downsampling). Only rows from the majority class get a "weight" greater than 1 and are potentially downsampled.

When you do apply weights in Advanced options ( [Classic](https://docs.datarobot.com/en/docs/classic-ui/modeling/build-models/adv-opt/additional.html)) or as part of advanced experiment setup ( [NextGen](https://docs.datarobot.com/en/docs/workbench/nxt-workbench/experiments/create-experiments/create-predictive/ml-adv-experiment.html#weight)), the counts shown in the confusion matrix do not align with the training set row counts. This is because the [Accuracy optimization metric](https://docs.datarobot.com/en/docs/reference/pred-ai-ref/opt-metric.html#accuracybalanced-accuracy) uses the sum of weights. Specifically:

> Every cell of the confusion matrix will be the sum of the sample weights in that cell. If no weights are specified, the implied weight is 1, so the sum of the weights is also the count of observations.
