Logo
Pattern

Discover published sets by community

Explore tens of thousands of sets crafted by our community.

Evaluation Metrics for Classification

10

Flashcards

0/10

Still learning
StarStarStarStar

Accuracy

StarStarStarStar

The ratio of correctly predicted observations to the total observations. It is calculated as

Accuracy=TP+TNTP+FP+FN+TNAccuracy = \frac{TP+TN}{TP+FP+FN+TN}
where TP = True Positives, FP = False Positives, FN = False Negatives, and TN = True Negatives.

StarStarStarStar

PR Curve (Precision-Recall Curve)

StarStarStarStar

A graph showing the trade-off between precision and recall for different thresholds. A high area under the curve represents both high recall and high precision.

StarStarStarStar

Log-Loss (Cross-Entropy Loss)

StarStarStarStar

A performance metric that measures the penalty for incorrect predictions, where the penalty is logarithmically proportionate to the inverse of the predicted probability for the actual class. Calculated as

LogLoss=1Ni=1N[yilog(pi)+(1yi)log(1pi)]Log-Loss = -\frac{1}{N}\sum_{i=1}^{N}[y_i \log(p_i) + (1 - y_i)\log(1 - p_i)]
where NN is the number of observations, yiy_i is the binary indicator of the class label, and pip_i is the predicted probability.

StarStarStarStar

Precision

StarStarStarStar

The ratio of correctly predicted positive observations to the total predicted positive observations. Calculated as

Precision=TPTP+FPPrecision = \frac{TP}{TP + FP}

StarStarStarStar

Confusion Matrix

StarStarStarStar

A table used to describe the performance of a classification model, showing the actual and predicted classifications. It helps to visualize true positives, false positives, true negatives, and false negatives.

StarStarStarStar

Recall

StarStarStarStar

The ratio of correctly predicted positive observations to all observations in the actual class - true positive rate. Calculated as

Recall=TPTP+FNRecall = \frac{TP}{TP + FN}

StarStarStarStar

Receiver Operating Characteristic (ROC) Curve

StarStarStarStar

A graphical plot that illustrates the diagnostic ability of a binary classifier system as its discrimination threshold is varied. The curve is generated by plotting the True Positive Rate (Recall) against the False Positive Rate at various threshold settings.

StarStarStarStar

Area Under the Receiver Operating Characteristic Curve (AUC-ROC)

StarStarStarStar

The area under the ROC curve. It provides an aggregate measure of performance across all possible classification thresholds. The value ranges from 0 to 1, where 1 implies a perfect model and 0.5 denotes a model with no discriminative ability.

StarStarStarStar

F1 Score

StarStarStarStar

The weighted average of Precision and Recall. Calculated as

F1 Score=2×Precision×RecallPrecision+RecallF1 \ Score = 2 \times \frac{Precision \times Recall}{Precision + Recall}

StarStarStarStar

Matthews Correlation Coefficient (MCC)

StarStarStarStar

A coefficient that measures the quality of binary classifications, which generates a value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 is no better than random prediction, and -1 indicates total disagreement between prediction and observation. Calculated by the formula

MCC=TP×TNFP×FN(TP+FP)(TP+FN)(TN+FP)(TN+FN)MCC = \frac{TP \times TN - FP \times FN}{\sqrt{(TP + FP)(TP + FN)(TN + FP)(TN + FN)}}

Know
0
Still learning
Click to flip
Know
0
Logo

© Hypatia.Tech. 2024 All rights reserved.