Precision, Recall, & F1 Score Intuitively Explained

preview_player
Показать описание
Classification performance metrics are an important part of any machine learning system. Here we discuss the most basic and common measures of model performance; accuracy, precision, recall, and F1 score.

Chapters:
0:00 Introduction
1:02 Basic Definitions
2:50 Accuracy
4:30 Precision
5:36 Recall
6:34 F1 Score
8:07 Conclusion
Рекомендации по теме
Комментарии
Автор

Many thanks for this intuitive yet detailed and well-reasoned explanation!

swedishguyonyoutube
Автор

Awesome explanation, thank you very much. It shines because it gives the motivation for each metric. Thanks!

jhfoleiss
Автор

These concepts "clicked" for me thanks to this video, really great work and thank you so much for sharing.

Maddy
Автор

Another good explanation of F1 score is that it considers FN and FP equally like:
Tp/(TP + 1/2(FN+FP)). So this is another way how it's a goos mix of precision and recall.

sarvagyagupta
Автор

Probably the best explanation on the net.

justsuyash
Автор

Great explanation because it includes motivation for the metrics and how they relate to each other.

vladimirblagojevic
Автор

Great explanation, thank you very much

RolandoLopezNieto
Автор

I couldnt get one thing, putting here if you are able to answer please ---> ' in the Slide of 'Precision', you mentioned that it solves problem with Accuracy cheating, where if 99% of people are healthy and 1% unhealthy, and if your model predict all as 'Healthy' then Precision would be 0, but how ? lets say 100 people, 99 healthy(positive), 1 unhealthy (negative), prediction made by model as All Positive (cheated), thus 99 (True-Positive), 1 ( False-Positive), Precision = TP / ( TP +FP ) = 99 / ( 99 + 1 ) = 0.99 = 99% right ? how 0 ?

samirkhan
Автор

Nice explanations. When you say negative labels are left out of picture in the precision calculation, I don’t quite get it because precision formula contains the term FP which represents a quantity from negative labels. Am I out of the track?

janithforex
Автор

ENJOYED THE VIDEO - keep up the great work !
I DO find myself with one question though... what about the "NULL" case did it actually get considered?

In one example you describe a type B as the "positive" case amongst types A, B, C and D; this makes SENSE, BUT...
... you then describe the "negatives" as being ONLY A, C and D... but what about the case of "NOT A, B, C OR D" ?
This is the "null" set, and it occurs quite regularly in Computer Vision OBJECT DETECTION models.

IMAGINE THIS SCENARIO:
An image of FRUIT on a TABLE - 3 red apples, 1 yellow apple, and 2 bananas.

SAY you're looking to detect APPLES, but your MODEL only detects the 3 RED APPLES and UTTERLY MISSES the 1 YELLOW APPLE?
THIS is a case of there BEING an APPLE, but you DIDN'T MIS-CLASSIFY it as the BANANA... YOU MISSED DETECTING IT COMPLETELY !

HOW would you describe THIS common scenario if you're only considering/assuming that your model WILL classify EVERYTHING in the image as EITHER an APPLE or a BANANA... but you DIDN'T expect it to UTTERLY IGNORE the yellow apple altogether?

It's been 2 years since you posted, so I'm not expecting a reply; I AM hoping that other viewers will ponder the explanations presented - there's a bit more going on...

Cheers,
-Mark Vogt, Solution Architect/Data Scientist - AVANADE

VOGTLANDOUTDOORS
Автор

Awesome! I love the cheating cases brought in!

elvenkim