-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature Request] Support for GALACTICA & EleutherAI Neo & Neo X models #10
Comments
Tried converting: The error starts with:
To further test/fuzz, I just added a vocab.json from GPT-JT
Added added_tokens.json to fuzz / find hints what would be needed for conversion in future:
Obviously there needs to be valid vocab and added_token files... I am figuring out how galactica works... |
I get similar errors as above when trying to convert neox 20b. How do I create the added_tokens.json file?
EDIT: Did a quick hack so that added_tokens.json is not required: https://github.com/trholding/ggml/blob/master/examples/gpt-j/convert-h5-to-ggml.py I get this error:
I suppose there is no point in editing configs as neox model is different at the model level I guess. EDIT I after a small change the neox conversion seemed to work, but OOM killed it. So I changed it again to make it suitable to convert neo 125 m, but again I got a error:
I did some blind changes: https://github.com/trholding/ggml/commits/master/examples/gpt-j/convert-h5-to-ggml.py Model gets converted but in the end results in this error:
I think I should give up... I have probably no idea what I am doing... |
Yeah, these models probably have different architecture compared to GPT-J, so it is not just a matter of converting the data. You have to also implement the missing layers and connect them correctly. Also there are probably some differences in the tokenizer. Every model can be ported to ggml, but it requires some work. I guess it would be better if I try to make the codebase easier to understand and document it. This way other people might wish to contribute. Otherwise, it's too much work for a single developer. |
Agreed, and a nice documentation with a Howto would be awesome :) |
Hello, any updates on where this is? I am interested in working to port this model. |
Looks like I won't have time to look into these in the near future. |
Closing as it is no longer relevant with current versions. |
Support for the GALACTICA & EleutherAI Neo, Neo X models would be an awesome addition.
GALACTICA seems like a ChatGPT for scientific stuff.
Info:
https://galactica.org/
https://huggingface.co/facebook/galactica-120b
https://the-decoder.com/galactica-is-an-open-source-language-model-for-scientific-progress/
https://huggingface.co/EleutherAI/gpt-neo-125M
https://huggingface.co/EleutherAI/gpt-neo-1.3B
https://huggingface.co/EleutherAI/gpt-neo-2.7B
https://huggingface.co/EleutherAI/gpt-neox-20b
The text was updated successfully, but these errors were encountered: