-
Notifications
You must be signed in to change notification settings - Fork 26.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add GPT-J ONNX config to Transformers #16274
Conversation
The documentation is not available anymore as the PR was closed or merged. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you for adding this very clean implementation @chainyo 🔥 !
The PR looks good and I've just left a few minor suggestions.
I think one useful check would be to see if you're able to export the model with the LM head and do a simple greedy search decoding with ONNX Runtime of some prompt. We've had some users report issues with TensorRT in #15640 and I'm curious to know if the same exists with ONNX Runtime (I don't think so, but nice to check).
Co-authored-by: lewtun <lewis.c.tunstall@gmail.com>
Co-authored-by: lewtun <lewis.c.tunstall@gmail.com>
2f74d97
to
56a8507
Compare
So what is the next step to make it merged ? |
Hey @chainyo the last thing to check is that the slow tests pass with:
We only run these on the Apart from that, the PR looks really good - gently pinging @sgugger or @LysandreJik for final approval |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for your contribution!
* add GPT-J ONNX config to Transformers * remove token-classification features mapping Co-authored-by: lewtun <lewis.c.tunstall@gmail.com> * add question-answering features mapping Co-authored-by: lewtun <lewis.c.tunstall@gmail.com> * add GPT2 config init to GPT2 config + copie shebang for fix-copies Co-authored-by: ChainYo <t.chaigneau.tc@gmail.com> Co-authored-by: lewtun <lewis.c.tunstall@gmail.com>
What does this PR do?
I'm looking for contributing to
Transformers
repository by adding more OnnxConfig to available models on the hub.I have created a little organization ONNXConfig for all to track the models that needs support for ONNX.
This is the first contribution since CamemBERT OnnxConfig some months ago.
I took example on
GPT2
andGPT-Neo
OnnxConfig but I'm not sure if everything is good or ifGPT-J
needs special things to be added.So this PR is a work in progress. If anyone can send me ressources to read to understand if it lacks anything, it would be awesome! 🤗
Who can review?
Models GPT2 / GPT-Neo
@LysandreJik @michaelbenayoun