# Multiclass classification

## Performance metrics

We have already discussed binary classification at some length, and multiclass classification is similar, so here we will concentrate on the differences. The primary difference is that there are not just two possible outcomes, but three or more.

Assuming that all of the possible outcomes are equally interesting to you,
with no special emphasis on any of them, it's reasonable to choose
**Accuracy** as your performance metric, and to prefer the model
with the highest **Accuracy**.

When there are *N* possible outcomes, the confusion matrix has *N x N* elements.
The correct predictions are on the diagonal, and incorrect predictions are off-diagonal.
Although the summary performance table in the **Model Comparison** only mentions
**Accuracy** and **Classification Error**, you can still find the analogues of
**Precision** and **Recall** by looking at the **Performance** tab for each individual model.
**Precision** is computed for each row; **Recall** is computed for each column.

Actually A | Actually B | Actually C | Precision | |
---|---|---|---|---|

Predicted A | True A | True A / Predicted A | ||

Predicted B | True B | True B / Predicted B | ||

Predicted C | True C | True C / Predicted C | ||

Recall | True A / Actually A | True B / Actually B | True C / Actually C |

In the example below, a multiclass classification problem with three possible outcomes,
the model made two wrong predictions, indicated by a red mark in the confusion matrix,
when applied to the test set. Consequently, the **Recall** for the second column (5/7)
and the **Precision** for the third row (6/8) are less than 100%.