You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
A new SOTA bitnet model, Bonsai 0.5B, has come out. Seems to outperform larger bitnet models like Falcon 1B, 3B, TriLM 700M. Seems like they are going to release a new line of bitnet models which is really exciting.
Support is needed for these models. They adopt a channel wise scaling factor compared to the tensor level ones. Maybe a separate kennel can be built to apply scales outside of the matmul kernels? Probably would yield similar inference speeds. Note that the hugging face does have a custom Q-linear layer that applies the scales.
They adopt a channel wise scaling factor compared to the tensor level ones. Maybe a separate kennel can be built to apply scales outside of the matmul kernels?
Hmm, channel-wise scales are not really convenient since they cannot be applied after the matmul, they have to be applied before. But that means they can be applied on the activations, though, if I understand correctly.
But maybe I'm reading the shapes wrong and the scales are row-wise, in which case TQ1_0 and TQ2_0 should already work (since they include a block-wise scale), and the weights simply have to be prepared beforehand so that the model can actually be converted.
EDIT: they apply the scales post-matmul, so this means the block-wise scales of TQ1_0 and TQ2_0 should work correctly with this model. It's only a matter of making the convert script do the right thing.
A new SOTA bitnet model, Bonsai 0.5B, has come out. Seems to outperform larger bitnet models like Falcon 1B, 3B, TriLM 700M. Seems like they are going to release a new line of bitnet models which is really exciting.
Support is needed for these models. They adopt a channel wise scaling factor compared to the tensor level ones. Maybe a separate kennel can be built to apply scales outside of the matmul kernels? Probably would yield similar inference speeds. Note that the hugging face does have a custom Q-linear layer that applies the scales.
HF: https://huggingface.co/deepgrove/Bonsai
Seems super promising --seems to be pretty performant and match full precision models like qwen2.5 0.5 from the looks of the report.
pinging @Eddie-Wang1120 + @compilade + other bitnet kernels contributors
Other posts and information:
https://www.reddit.com/r/LocalLLaMA/comments/1jgkqio/new_bitnet_model_from_deepgrove/
https://x.com/deepgrove_ai/status/1903103798735761518
The text was updated successfully, but these errors were encountered: