Evaluation Metrics in Machine Learning
Evaluation Metrics in Machine Learning
TASK 1
MIP-ML-08 BATCH
Evaluation metrics in Machine Learning (ML) are important tools for checking how well
ML models work. These metrics use numbers to show how effective a model is at
handling new information and help decide which model is best. They look at things like
accuracy, precision, recall, and other measures to see if the model is performing well.
The goal is to make sure the model can learn from training and make accurate
predictions on new things it hasn't seen before. These metrics also help pick the best
model from many options and can find where a model needs to improve. This helps ML
experts adjust things to make the model more effective and ensures it does well in real-
life situations.
predictions or classifications.
• These metrics guide us in choosing the best model among several options by
perform better.
ML algorithms.
• They allow us to objectively assess and compare different models based on their
performance scores.
• These metrics are essential for making informed decisions about which ML
• Overall, evaluation metrics are key tools for assessing, improving, and optimizing
between predicted and actual values, giving an idea of how accurate the
predictions are.
• Root Mean Squared Error (RMSE): Similar to MAE but emphasizes larger
• R-squared (R2) Score: It shows how much of the variance in the data is
explained by the model, indicating how well the model fits the data.
include:
predictions.
• F1 Score: The F1 score combines precision and recall into a single value,
class imbalances.
• Confusion Matrix: A confusion matrix is a tabular representation of a
• It provides insights into how accurate the predictions of the model are on
average.
Where:
• N: number of datums
Example Graph: Mean Absolute Error
• RMSE is similar to MAE but gives more weight to larger errors, making it
sensitive to outliers.
• It indicates the goodness of fit of the regression model, showing how well
Where:
4.Accuracy:
• Accuracy is a simple and intuitive metric that measures the percentage of correct
• It is suitable for balanced datasets where the positive and negative classes are
similar in number.
• TP=True Positive
• TN=True Negative
• FP=False Positive
• FN=False Negative
5.Precision:
6.Recall:
• For instance, in healthcare, high recall ensures that the model doesn't miss
identifying patients with a disease, even if it means some healthy individuals are
Where:
TP=True Positive
FN=False Negative
7.F1 Score:
• Models with a similar balance between precision and recall are favored by the F1
score.
• The harmonic mean is particularly suitable for averaging ratios of values, making
the F1 score valuable in scenarios with imbalanced precision and recall values.
Example Graph: F1 Score
8.Confusion Matrix:
a classification problem.
• It displays the four possible combinations of true positives, true negatives, false
positives, and false negatives, offering insights into the model's performance and
Recap:
of a model's capabilities and areas for improvement. The table below summarizes the
key evaluation metrics discussed in this article, along with their descriptions and
formulas:
develop robust and accurate ML models that meet the desired performance standards
Thank you.