New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Adding TAPEX to HuggingFace Transformers #6
Comments
Hi Niels ! @NielsRogge Glad to hear that! Thank you very much for the kind words. I love and appreciate the great work you did integrating TAPAS into Huggingface, which draws more attention to the community ❤️. In fact, I have tried to convert fairseq model checkpoints to Huggingface Transformers and integrate them into the library. It will be a great honor for TAPEX to be integrated into 🤗 Transformers! But there are three reasons slowing down the process: (i). These days I am catching up with a paper submission deadline until November 15; (ii). TAPEX is still under review and I do not want to actively advertise it to influence the double-blind review. (iii). I have tried to train BART-large on the same dataset (e.g., Wikitablequestions) using Transformers, but it did not give a similar performance as with fairseq, which makes me very confused. I'm still trying to figure out the reason (may need your help in the near future :-D). Thanks again for your attention and effort on our work! I have joined the Microsoft org few days ago, thanks again! Best, |
Hi, Great to hear :) So I've managed to convert the TAPEX-base checkpoint to its HuggingFace counterpart (in a Where can I find this vocabulary file? Is the same vocabulary used during pre-training and fine-tuning? Similar to TAPAS, I'm considering making a
|
@NielsRogge Hi Niels! It looks that you have found the BPE files. Right they follow the same procedure with BART-base. BTW, we have an initial plan to release T5 based models (or more models) under the pre-training procedure of TAPEX to faciliate the community. What model names do you think are appropriate for the current TAPEX for future compatibility? For example, Thanks again for your effort! Best Regards, |
Oh great to hear! Maybe it makes sense to include the model architecture in the name of each checkpoint, so |
@NielsRogge I agree with that. Thanks for your quick reply! If there is any other thing I can help, please directly ping me or email me (qian.liu@buaa.edu.cn). Any discussion about TAPEX itself is also warmly welcome! In the next few months, I will be relatively free lol. |
Do you have time to add this model to the library? The modeling part is done, as the model is just a BART model. The conversion notebook can be found here. However, it would be great if you could implement the I can help you in the process. It will also make you familiar with HuggingFace Transformers (it's like a behind-the-scenes of the library). Let me know if you're interested :) |
@NielsRogge Sure! I'm interested to do that (I'm a earlier fans of huggingface transformers lol). I will add it to the library following the official guideline, which may cost nearly 1 week. I will be back here if I meet anything I cannot deal with, or I think the pull request needs reviewing. Thanks in advance, Niels! Best, |
Ok great, let me set up a Slack channel with your email address such that we can communicate over there rather than here. Is that ok for you? |
@NielsRogge It sounds good for me! |
@SivilTaram Hi, I just wanted to check in on the status of adding these models to Huggingface. Thank you! |
@vnik18 Hi! You can see the updated README to preview the fine-tuning script. And |
@SivilTaram Thank you for letting me know! |
TAPEX is merged into huggingface now. Now you can view https://github.com/huggingface/transformers/tree/main/examples/research_projects/tapex to have a try! Enjoy it ☕ . |
Hi!
First of all, congratulations on the great paper and results! If you need any help regarding converting the models to the HuggingFace API, let me know. My very first contribution to HuggingFace Transformers was actually TAPAS, the table question answering model from Google. We also create a subsequent table-question answering task on the hub, as well as an inference widget which let's users directly try out TAPAS in the browser.
We could do the same for TAPEX! TAPEX also looks way simpler as it's a generative model. The only thing required would be to write a conversion script (which I can help you with).
Also, are you interested in joining the Microsoft organization on the hub?
Btw, the README is also very comprehensive and well written. Wish more authors did that ;)
Kind regards,
Niels
ML Engineer @ HuggingFace
The text was updated successfully, but these errors were encountered: