ExamGecko
Question list
Search
Search

List of questions

Search

Related questions











Question 71 - Professional Machine Learning Engineer discussion

Report
Export

Your task is classify if a company logo is present on an image. You found out that 96% of a data does not include a logo. You are dealing with data imbalance problem. Which metric do you use to evaluate to model?

A.
F1 Score
Answers
A.
F1 Score
B.
RMSE
Answers
B.
RMSE
C.
F Score with higher precision weighting than recall
Answers
C.
F Score with higher precision weighting than recall
D.
F Score with higher recall weighted than precision
Answers
D.
F Score with higher recall weighted than precision
Suggested answer: A

Explanation:

The F1 score is a metric that combines both precision and recall, and is suitable for evaluating imbalanced classification problems. Precision measures the fraction of true positives among the predicted positives, and recall measures the fraction of true positives among the actual positives. The F1 score is the harmonic mean of precision and recall, and it ranges from 0 to 1, with higher values indicating better performance. The F1 score is a good metric for imbalanced data because it balances both the false positives and the false negatives, and does not favor the majority class over the minority class.

The other options are not good metrics for imbalanced data. RMSE (root mean squared error) is a metric for regression problems, not classification problems. It measures the average squared difference between the predicted and the actual values, and is not suitable for binary outcomes. F score with higher precision weighting than recall, or F0.5 score, is a metric that gives more importance to precision than recall. This means that it penalizes false positives more than false negatives, which is not desirable for imbalanced data where the minority class is more important. F score with higher recall weighting than precision, or F2 score, is a metric that gives more importance to recall than precision. This means that it penalizes false negatives more than false positives, which might be suitable for some imbalanced data problems, but not for the logo detection problem. In this problem, both false positives and false negatives are equally important, as we want to accurately identify the presence or absence of a logo in an image. Therefore, the F1 score is a better metric than the F2 score.Reference:

Tour of Evaluation Metrics for Imbalanced Classification

Metrics for imbalanced data (simply explained)

asked 18/09/2024
brandon landaal
40 questions
User
Your answer:
0 comments
Sorted by

Leave a comment first