Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Sharing Models through the Hugging Face Hub #86

Closed
osanseviero opened this issue Aug 24, 2021 · 7 comments
Closed

Sharing Models through the Hugging Face Hub #86

osanseviero opened this issue Aug 24, 2021 · 7 comments
Milestone

Comments

@osanseviero
Copy link

Hi CRFM team!

Mistral is very exciting! I see you currently share your model checkpoints through links to a hosted server. Would you be interested in sharing the pretrained models in the Hugging Face Hub? We already have a similar collaboration with the Stanford NLP group (see org).

The Hub offers free hosting of over 20K models, and it would make your work more accessible and visible to the rest of the community. Some of the benefits of sharing your models would be:

  • forget about the pain of managing the hosting
  • built-in versioning
  • commit history and diffs
  • repos provide useful metadata about their tasks, languages, metrics, etc that is useful for discoverability but also to understand the model

Creating the repos and adding new models should be a relatively straightforward process if you've used Git before. This is a step-by-step guide explaining the process in case you're interested. Please let us know if you would be interested and if you have any questions.

In a future we could also integrate this to our Inference API so users can play with the models directly in the browser with our widgets.

Happy to hear your thoughts,
Omar and the Hugging Face team

cc @lewtun @anton-l @LysandreJik

@siddk
Copy link
Contributor

siddk commented Aug 24, 2021

Hey @osanseviero - this is something we were talking about, but one clarification; would you be able to host all 610 checkpoints for each of the 10 runs (6100 checkpoints total, ~22TB)?

We figured this could get complicated (and expensive), but if you can do it, we can go through the process! Would also let us push our slightly tweaked/stable GPT-2 model definition up to HF as well!

@osanseviero
Copy link
Author

Hey @siddk! That's great to hear. Yes, we're up to host all your checkpoints if you would really like to share all of them. The way I would suggest to do it is to have one repository for each experiment and do a new commit for each checkpoint. Users would then be able to load the checkpoint from a given revision and the working widget would use the latest checkpoint, which should correspond to step 400,000.

@siddk
Copy link
Contributor

siddk commented Aug 24, 2021

Sounds great. I think we need to do a bit of clean-up on our side this week then, but we will start the process above early next week! Looking forward to working through this with you.

And a heads up - this probably won't be the last set of models we train 🙂 ! Looking forward to fostering a stronger relationship with HF as we keep exploring!

@dlwh
Copy link
Member

dlwh commented Mar 10, 2022

@siddk what's the definition of done here? is it having uploaded 6100 model checkpoints? (as opposed to the 10 that are there now?) Are we going to do that?

@siddk
Copy link
Contributor

siddk commented Mar 10, 2022

These should all be done - if you look at the different branches, you should see all 610 checkpoints (see here: https://huggingface.co/stanford-crfm/arwen-gpt2-medium-x21/tree/main).

@dlwh
Copy link
Member

dlwh commented Mar 10, 2022

ok, our readme.md says git clone https://huggingface.co/stanford-crfm/arwen-x21-checkpoint-400000 so we should probably fix

@dlwh
Copy link
Member

dlwh commented Mar 10, 2022

i'm closing this and i opened #123

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants