Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Inconsistent behavior in averaging choices micro and samples. #10706

Open
amueller opened this issue Feb 26, 2018 · 2 comments
Open

Inconsistent behavior in averaging choices micro and samples. #10706

amueller opened this issue Feb 26, 2018 · 2 comments

Comments

@amueller
Copy link
Member

For the multi-class case, precision, recall and f-score with micro all produce accuracy, while with samples they produce an error.
That seem inconsistent. Using the definitions in the docs, they should also all be accuracy, I think.

I think I'd propose to deprecate micro averaging for multiclass.

The docs actually give an example of micro-average recall for multiclass, which is really weird imho.

@amueller
Copy link
Member Author

amueller commented Feb 26, 2018

The docs at the top actually recommend micro average for multi-class:

Micro-averaging may be preferred in multilabel settings,
including multiclass classification where a majority class is to be ignored.

That seems weird to me, given that it's just accuracy.

@jnothman
Copy link
Member

Wording could be clearer, but the intention there is that using multiclass with a majority class ignored (labels=np.setdiff1d(classes_, 'default class') will return something other than accuracy. I think it is hard to deprecate because of how it is used in classification_report and elsewhere.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

3 participants