You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
New model addition - Google PG-19 Models
Model description
Model checkpoints finally released as discussed in "Efficient Content-Based Sparse Attention with Routing Transformers'
Aurko Roy, Mohammad Saffar, Ashish Vaswani, David Grangier (https://arxiv.org/abs/2003.05997)
Open source status
[X ] the model implementation is available: (same link as below)
[ X] the model weights are available: ( https://github.com/google-research/google-research/tree/master/routing_transformer)
[X ] who are the authors: (see above)
Good luck with your project.
The text was updated successfully, but these errors were encountered:
I posted a new model request last May which provides links to google Routing Transformer models trained on PG-19 Corpus.
Perhaps these can be converted to run with huggingface transformers?
https://github.com/google-research/google-research/tree/master/routing_transformer
huggingface/transformers#11686
New model addition - Google PG-19 Models
Model description
Model checkpoints finally released as discussed in "Efficient Content-Based Sparse Attention with Routing Transformers'
Aurko Roy, Mohammad Saffar, Ashish Vaswani, David Grangier (https://arxiv.org/abs/2003.05997)
Open source status
Good luck with your project.
The text was updated successfully, but these errors were encountered: