-
Notifications
You must be signed in to change notification settings - Fork 48
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
minor: Causal fidelity metrics: Harmonization between timeseries and image - detailed evaluate #70
Conversation
You're trying to re-commit commits from last month, can you clean your PR with just the new commits pls ? ;) |
3b19788
to
ea31c39
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Most of my remarks question the relevance of some change to get more complete metrics. The idea is good but there is a few change to make I believe or to discuss at least. Otherwise, well done 👍 🎉
ea31c39
to
fdacaa8
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Cool PR, LGTM ! ;)
fdacaa8
to
22aebd4
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you for the modification!
minor: Causal fidelity metrics: Harmonization between timeseries and image - detailed evaluate
This pull request is divided in 3 contributions :
and a correction for the new version of pylint to work: 22aebd4 (we should use f-string).
Harmonization
The previously developed causal fidelity metrics for timeseries had several differences in their behavior compared to the initial Insertion and Deletion, thus an harmonization was necessary.
Detailed evaluate
Furthermore, through discussion, the necessity of accessing all values computed for insertion and deletion was necessary. The idea thus was to add a method (detailed_evaluate) with returns a dictionnary of such values. This method will then be called by the initial evaluate method that will return the auc.
The output of the detailed evaluate method can be used to draw score evolution curves, where the score depend on the number of features perturbed.
Interpretation and remarks in the documentation
There are remarks about: