-
Notifications
You must be signed in to change notification settings - Fork 425
Commit
- Loading branch information
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -22,6 +22,7 @@ | |
"internlm", | ||
"qwen", | ||
"xverse", | ||
"deci_lm", | ||
This comment has been minimized.
Sorry, something went wrong.
This comment has been minimized.
Sorry, something went wrong.
LaaZa
Author
Contributor
|
||
] | ||
if compare_transformers_version("v4.28.0", op="ge"): | ||
SUPPORTED_MODELS.append("llama") | ||
|
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,31 @@ | ||
from logging import getLogger | ||
|
||
from ._base import * | ||
from ..utils.import_utils import compare_transformers_version | ||
|
||
if compare_transformers_version("v4.28.0", op="ge"): | ||
from ..nn_modules.fused_llama_attn import FusedLlamaAttentionForQuantizedModel | ||
from ..nn_modules.fused_llama_mlp import FusedLlamaMLPForQuantizedModel | ||
else: | ||
FusedLlamaAttentionForQuantizedModel = None | ||
FusedLlamaMLPForQuantizedModel = None | ||
|
||
logger = getLogger(__name__) | ||
|
||
|
||
class DeciLMGPTQForCausalLM(BaseGPTQForCausalLM): | ||
layer_type = "DeciLMDecoderLayer" | ||
layers_block_name = "model.layers" | ||
outside_layer_modules = ["model.embed_tokens", "model.norm"] | ||
inside_layer_modules = [ | ||
["self_attn.k_proj", "self_attn.v_proj", "self_attn.q_proj"], | ||
["self_attn.o_proj"], | ||
["mlp.up_proj", "mlp.gate_proj"], | ||
["mlp.down_proj"] | ||
] | ||
|
||
fused_attn_module_type = FusedLlamaAttentionForQuantizedModel | ||
fused_mlp_module_type = FusedLlamaMLPForQuantizedModel | ||
|
||
|
||
__all__ = ["DeciLMGPTQForCausalLM"] |
@LaaZa Can you change it from "deci_lm" to "deci",
![image](https://private-user-images.githubusercontent.com/20045975/292377294-64b78d86-3b65-4d26-8ff6-ef34f2de0bce.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MjE2NzIxMDksIm5iZiI6MTcyMTY3MTgwOSwicGF0aCI6Ii8yMDA0NTk3NS8yOTIzNzcyOTQtNjRiNzhkODYtM2I2NS00ZDI2LThmZjYtZWYzNGYyZGUwYmNlLnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNDA3MjIlMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjQwNzIyVDE4MTAwOVomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTFjYjVhZjk1ZmQzZThmMWVlNGM1OGRlYzFlMmM1MDE0YTlkMzIwMTJiMGVhODViMzdkNDMyMTYzZjkxZjZlZTcmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0JmFjdG9yX2lkPTAma2V5X2lkPTAmcmVwb19pZD0wIn0.U7G3APFxUi0HkMQGIyTx-_UcRAVN8Us_p6eA1PhuQJk)
getting this error, please check the ss.
Right now I have to edit the json and make it "deci_lm" to make it work.