You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Shuffle is important for ranking based metrics. For example, if you score all the candidate document with the same value (e.g., 0.0), and your data is organized by label. Then you may get a high MAP/NDCG performance. So, shuffle can reduce this illusion.
Thank you @thiziri
The trec_eval binary was very useful. I am using it for my purposes.
@faneshion
That sounds like a good reason. Is it incorporated into TREC/ is it standard practice to shuffle before evaluating? I need to establish solid benchmarks and hence need a commony accepted evaluation metric.
https://github.com/faneshion/MatchZoo/blob/50422a5bc973d27b8c507122aa3f41e633366893/matchzoo/metrics/evaluations.py#L19
@faneshion @bwanglzu
I am trying to understand the implementation of these metrics. Cannot understand why there is a random shuffle.
Also, could you point me to a good resource for an implementation of such metrics? I have gone through most found in google.
The text was updated successfully, but these errors were encountered: