Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Maybe a link on how to set it up ? #6

Closed
PGTBoos opened this issue May 3, 2023 · 3 comments
Closed

Maybe a link on how to set it up ? #6

PGTBoos opened this issue May 3, 2023 · 3 comments

Comments

@PGTBoos
Copy link

PGTBoos commented May 3, 2023

Might be nice to try this, but how does one set it up in a python jupyter environment ?.
(a working step by step would be nice, or an example notebook)

I'm no newbie but never had the time to spend a few days into LLM's.
if a 3080ti can run it , ib be fun to give it a try.

@young-geng
Copy link
Contributor

Thanks for the suggestion! Since we train this with JAX using our EasyLM framework, we have detailed documentation for using it with EasyLM. As for using it with PyTorch, unfortunately we weren't able to experiment much as our compute resources are on TPU.

@PGTBoos
Copy link
Author

PGTBoos commented May 3, 2023 via email

@young-geng
Copy link
Contributor

You can also use the JAX version on GPU. That's how we test it locally.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants