We often use a classification report to check the quality of classification algorithm predictions. A sample report is shown below which we generated in our previous post where we ran K-Means on Iris Dataset.

A above report shows the main classification metrics precision, recall and f1-score on a per-class basis. The metrics are defined in terms of true and false positives, and true and false negatives. Positive and negative in this case are generic names for the classes of a binary classification problem. There are four ways to check if the predictions are right or wrong:

**TN / True Negative:**case was negative and predicted negative**TP / True Positive:**case was positive and predicted positive**FN / False Negative:**case was positive but predicted negative**FP / False Positive:**case was negative but predicted positive

**Precision – What percent of your predictions were correct?**

Precision is the ability of a classifier not to label an instance positive that is actually negative. For each class it is defined as the ratio of true positives to the sum of true and false positives.

**TP – True Positives**

**FP – False Positives**

Precision – Accuracy of positive predictions.

Precision = TP/(TP + FP)

from sklearn.metrics import precision_score print("Precision score: {}".format(precision_score(y_true,y_pred)))

**Recall – What percent of the positive cases did you catch? **

Recall is the ability of a classifier to find all positive instances. For each class it is defined as the ratio of true positives to the sum of true positives and false negatives.

**FN – False Negatives**

Recall: Fraction of positives that were correctly identified.

Recall = TP/(TP+FN

from sklearn.metrics import recall_score print("Recall score: {}".format(recall_score(y_true,y_pred)))

**F1 score – What percent of positive predictions were correct? **

The F_{1} score is a weighted harmonic mean of precision and recall such that the best score is 1.0 and the worst is 0.0. Generally speaking, F_{1} scores are lower than accuracy measures as they embed precision and recall into their computation. As a rule of thumb, the weighted average of F_{1} should be used to compare classifier models, not global accuracy.

F1 Score = 2*(Recall * Precision) / (Recall + Precision)

from sklearn.metrics import f1_score print("F1 Score: {}".format(f1_score(y_true,y_pred)))

## Write a Comment