Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Definition of “Precision” seems incorrect... #1049

Closed
Scoodood opened this issue Mar 10, 2020 · 2 comments · Fixed by #1074
Closed

Definition of “Precision” seems incorrect... #1049

Scoodood opened this issue Mar 10, 2020 · 2 comments · Fixed by #1074
Assignees
Labels
priority: high should be done before next release type: documentation writing and editing tasks for RTD

Comments

@Scoodood
Copy link

On this page at ClassificationReport

CE6AA3AE-37ED-4CA2-8D19-6AF2395DAC79

For binary classifier, the part “not to label an instance positive that is actually negative” sounded like “to correctly classified a negative”, which is the definition of True Negative right? I think the first sentence should be removed to avoid confusion.

@bbengfort
Copy link
Member

@Scoodood I agree that we could make it more clear; however I don't think we can necessarily simply remove the first sentence; do you have an edit to suggest?

@bbengfort bbengfort added the type: documentation writing and editing tasks for RTD label Apr 9, 2020
@bbengfort bbengfort self-assigned this Apr 9, 2020
@bbengfort bbengfort added the priority: high should be done before next release label Apr 9, 2020
@Scoodood
Copy link
Author

Scoodood commented Apr 9, 2020

Hi @bbengfort,

If we simplified the first sentence it will become
"Precision is the ability of a classifier to correctly label a negative instance as negative."

That definition is wrong, because it belongs to the True Negative Rate, not Precision. According to the wikipedia,
Precision = TP / (TP + FP) = Positive Predictive Value (PPV)

The denominator term TP + FP also means total predicted output. So the whole equation means for all cases that were labeled by the classifier as positive, what percent of those were actually correct. But your second sentence said this already.

Besides, the formula of TPR (True Positive Rate) and PPV is very similar, and yet their interpretation is different.
TPR = The ability of classifier to recognize positive instance
PPV = The ability of classifier to give relevant positive output

So if you insist to keep the first sentence, then I would like to suggest something like this
Precision measures the ability of a classifier to give relevant positive output.

But I really think your second sentence is clear enough even without the first sentence.

rebeccabilbro added a commit to rebeccabilbro/yellowbrick that referenced this issue Jun 10, 2020
rebeccabilbro added a commit that referenced this issue Jun 24, 2020
Update to PR Curve docs, closes #602 closes #1049
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
priority: high should be done before next release type: documentation writing and editing tasks for RTD
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants