-
Notifications
You must be signed in to change notification settings - Fork 2.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[QUESTION]Mamba-2-hybrid Weights #864
Comments
I think the model weights are released here: https://huggingface.co/collections/nvidia/ssms-666a362c5c3bb7e4a6bcfb9c |
Thanks! I've already found it. While when this question is posted, the weights haven't been set as public. Now, I'm looking for the tokenizer🤣. To run the example, a tokenizer is required. But I cannot find any. Any idea about this? |
I think the tokenizer path should point to the .model file in the huggingface repos. For example, I downloaded the
respectively. |
Wow, thank you so much for your guidance! It took me hours to find something like a tokenizer. Never used megatron before🙃. You did save my life!! |
Your question
An Empirical Study of Mamba-based Language Models](https://github.com/NVIDIA/Megatron-LM/tree/ssm/examples/mamba)
Hi! I'm impressed by this work and cannot wait to try the new mamba-2-hybrid. This paper mentioned that the weights are released on Huggingface. But I cannot find any. Wondering have they been released? If yes, where can I download them?
Thanks a lot for your folks' contribution to the community!
The text was updated successfully, but these errors were encountered: