-
Notifications
You must be signed in to change notification settings - Fork 2
Some Strange Bugs with llama3.1-8b-instruct #8
Copy link
Copy link
Closed
Description
Hello, your work is very meaningful, but I’ve encountered a few issues as shown below:
- When I use
llama3.1-8b-instructas the base model andrun 1_visualize.py, the output is:
Unfaithful acctivations:
['0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000']
Faithful acctivations:
['0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000']
Difference acctivations:
['0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000', '0.0000']
This doesn’t seem like a reasonable result. Could you please explain why this happens?
2. When I run tune.sh, I encounter the following error:
File "/mnt/sdc/xzs/experient/ParamMute/src/2_tuning/train.py", line 102, in <module>
main()
File "/mnt/sdc/xzs/experient/ParamMute/src/2_tuning/train.py", line 96, in main
trainer.train()
File "/mnt/sdc/xzs/experient/ParamMute/src/transformers/src/transformers/trainer.py", line 2164, in train
return inner_training_loop(
File "/mnt/sdc/xzs/experient/ParamMute/src/transformers/src/transformers/trainer.py", line 2524, in _inner_training_loop
tr_loss_step = self.training_step(model, inputs, num_items_in_batch)
File "/mnt/sdc/xzs/experient/ParamMute/src/transformers/src/transformers/trainer.py", line 3669, in training_step
loss = self.compute_loss(model, inputs, num_items_in_batch=num_items_in_batch)
File "/mnt/sdc/xzs/experient/ParamMute/src/2_tuning/trainer.py", line 155, in compute_loss
outputs = model(**inputs)
File "/home/xzs/data/miniconda3/envs/parammute/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/home/xzs/data/miniconda3/envs/parammute/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
File "/home/xzs/data/miniconda3/envs/parammute/lib/python3.10/site-packages/accelerate/utils/operations.py", line 819, in forward
return model_forward(*args, **kwargs)
File "/home/xzs/data/miniconda3/envs/parammute/lib/python3.10/site-packages/accelerate/utils/operations.py", line 807, in __call__
return convert_to_fp32(self.model_forward(*args, **kwargs))
File "/home/xzs/data/miniconda3/envs/parammute/lib/python3.10/site-packages/torch/amp/autocast_mode.py", line 44, in decorate_autocast
return func(*args, **kwargs)
File "/home/xzs/data/miniconda3/envs/parammute/lib/python3.10/site-packages/peft/peft_model.py", line 849, in forward
return self.get_base_model()(*args, **kwargs)
File "/home/xzs/data/miniconda3/envs/parammute/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/home/xzs/data/miniconda3/envs/parammute/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
File "/mnt/sdc/xzs/experient/ParamMute/src/transformers/src/transformers/models/llama/modeling_llama.py", line 1020, in forward
outputs = self.model(
File "/home/xzs/data/miniconda3/envs/parammute/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/home/xzs/data/miniconda3/envs/parammute/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
File "/mnt/sdc/xzs/experient/ParamMute/src/transformers/src/transformers/models/llama/modeling_llama.py", line 725, in forward
raise ValueError("You must specify exactly one of input_ids or inputs_embeds")
ValueError: You must specify exactly one of input_ids or inputs_embeds
My bash code follows your example exactly. I only changed model_name_or_path and train_file.
Your reply was very helpful for me to successfully reproduce the work. I look forward to your response!
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels