You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Here is a comment from the ReadMe regarding usage of convert.py:
# obtain the original LLaMA model weights and place them in ./models
However, it's not clear what format those weights need to be in...
In what format do I need to have the original weights? consolidated.00.pth or .bin?
If .bin, do the weights need to be float 32 or can they be bf16?
If .bin, is there a way to handle sharded weights? e.g. Llama 2 7B sharded into ~10 GB and ~3 GB? Can I specify each shard as a command line argument here:
# convert the 7B model to ggml FP16 format
python3 convert.py models/7B/
The text was updated successfully, but these errors were encountered:
Here is a comment from the ReadMe regarding usage of convert.py:
However, it's not clear what format those weights need to be in...
The text was updated successfully, but these errors were encountered: