Accuracy-Precision-Recall
In Machine learning classification Accuracy is a simple metric measuring how many cases the model classified correctly divided by the number of cases.
Accuracy = Number of cases predicted correctly / Total number of Cases.
If all the predictions are correct then the accuracy score will be 1
If all the predictions are wrong the accuracy score will be 0
Is Accuracy a good metric to measure the efficiency of a Classification model ?
The answer is No. Why ?
Assume that for a sample size of 100 credit card transactions one transaction will be fraudulent. That is 1% of all the credit card transaction will be fraudulent. A classification model that blindly labels every transaction as genuine is 99% accurate in this case. It hides false positives and false negatives.
Confusion matrix.
First we need to understand the four terms : True Positive , True negative , False positive , false Negative.
True positive(TP) : Cases when classifier predicted true (Fraudulent transaction) and the correct class was also true(Actually fraudulent transaction).
True negative(TN): Cases when classifier predicted as False(Genuine transaction) and the correct class was also false(Genuine transaction).
False Positive(FP): Cases when classifier predicted true (Fraudulent transaction) but the correct class was false(Genuine transaction) . This is called Type 1 error.
False Negative(FN) : Cases when classifier predicted as False (Genuine transaction) but the correct class was true (Fraudulent transaction). This is called Type 2 error.
Type 2 errors are usually more serious than Type 1 errors. Having false negatives allows fraudulent transactions to take place, defeating the very purpose of a fraudulent detection system.
This kind of error can be avoided by evaluating the model's precision and recall .
Precision checks out of all cases detected as fraudulent how many were actually fraudulent transactions.
Precision = TP / Total true prediction. = TP / (TP + FP)
Recall checks out of all fraudulent transactions how many was actually detected as fraudulent
Recall = TP / Actual all fraudulent transactions = TP / (TP + FN)
Precision is about correctness and Recall is about completeness. To get a good classification model, it should be balanced
The F1-score combines precision and recall of a classifier into one metric by taking their harmonic mean