You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
start launcher with gptq model, then try to load non gptq adapter
2023-11-22T19:34:06.185127Z ERROR lorax_client: router/client/src/lib.rs:33: Server error: 'QuantLinear' object has no attribute 'weight'
for tests if commented out the ID check
#if adapter_config.base_model_name_or_path != model_id:
# raise ValueError(f"Adapter '{adapter_id}' is not compatible with model '{model_id}'. "
# f"Use --model-id '{adapter_config.base_model_name_or_path}' instead.")
I am already thinking about better check, because it depends on model arch and parameters count instead of the id directly. so zephyr lora could be merged successfully to mistral instruct too.
Expected behavior
detect gptq model and use qweight instead of weight when merging
The text was updated successfully, but these errors were encountered:
Ah yeah, definitely agree the current check is too aggressive here. I was thinking of changing it to be a warning (like: adapter was trained on a different base model with the same architecture). I can take a quick look.
System Info
master branch
Information
Tasks
Reproduction
start launcher with gptq model, then try to load non gptq adapter
2023-11-22T19:34:06.185127Z ERROR lorax_client: router/client/src/lib.rs:33: Server error: 'QuantLinear' object has no attribute 'weight'
for tests if commented out the ID check
I am already thinking about better check, because it depends on model arch and parameters count instead of the id directly. so zephyr lora could be merged successfully to mistral instruct too.
Expected behavior
detect gptq model and use qweight instead of weight when merging
The text was updated successfully, but these errors were encountered: