-
Notifications
You must be signed in to change notification settings - Fork 470
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Alignment on modelcard metadata specification #39
Conversation
Cool thanks ! One note regarding metrics: they might need |
Very excited about this, and let's also ping the Paperswithcode team here for validation? |
Absolutely! Pinged them offline for a review. |
validation and integration? :) |
Thanks for looping us in! Yeah the format looks good. Since we already have the integration for wav2vec2 running, it should be pretty easy from our side to extend it to any result on any benchmark. A couple of things to consider:
|
One question re the front-end: can we use this schema to enable sorting by metric value on the Hub? For example, it would be cool to help users answer questions like "Which model achieves the highest metric value X on dataset Y?" I realise this overlaps with PWC's leaderboards (example), but still think there's value in providing this kind of overview to Hub users. |
@lewtun At some point we might want to display some sort of leaderboard-lite on the hf.co hub, but for now I feel like our main goal is on the data side, i.e. to ensure that as many models as possible contain the correct metadata in a format that's easily validated/leveraged by tools including Paperswithcode |
Will add an |
One question regarding dataset versioning: is this also something we want to include in the dataset card ? IMO this would be nice for reproducibility. |
Hi all, opening this PR so that we can all align on the metadata spec for model cards. This metadata is important as it closes the bridge between tasks, datasets, and metrics for a given checkpoint. This will eventually allow programmatic analysis and handling of model cards' metadata on the hub.
The metadata was drafted during the collaboration with papers-with-code in order to have a ranked leaderboard for the XLSR sprint. The following format was adopted:
This format should allow for multiple tasks and multiple metrics within each task.
Some existing examples of modelcards with this format are the modelcards uploaded during the XLSR sprint, like the following ydshieh/wav2vec2-large-xlsr-53-chinese-zh-cn-gpt
Looking forward to your feedback.
Transformers: @sgugger @patrickvonplaten
Datasets: @lhoestq
AutoNLP: @abhi1thakur @SBrandeis
Evaluation: @lewtun
@Pierrci @julien-c @thomwolf