Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Routing Transformer weights available if useful #11

Open
GenTxt opened this issue Feb 25, 2022 · 1 comment
Open

Routing Transformer weights available if useful #11

GenTxt opened this issue Feb 25, 2022 · 1 comment

Comments

@GenTxt
Copy link

GenTxt commented Feb 25, 2022

I posted a new model request last May which provides links to google Routing Transformer models trained on PG-19 Corpus.

Perhaps these can be converted to run with huggingface transformers?

https://github.com/google-research/google-research/tree/master/routing_transformer

huggingface/transformers#11686

New model addition - Google PG-19 Models
Model description

Model checkpoints finally released as discussed in "Efficient Content-Based Sparse Attention with Routing Transformers'
Aurko Roy, Mohammad Saffar, Ashish Vaswani, David Grangier (https://arxiv.org/abs/2003.05997)
Open source status

[X ] the model implementation is available: (same link as below)
[ X] the model weights are available: ( https://github.com/google-research/google-research/tree/master/routing_transformer)
[X ] who are the authors: (see above)

Good luck with your project.

@vilmarzti
Copy link
Owner

Hi thanks, this is useful information :)
I have some other stuff to fix, but this will definitively come on to the list

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants