We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
train_sft.py训练指令: CUDA_VISIBLE_DEVICES=0 python src/train_sft.py --model_name_or_path /data1/projects/baichuan-7B/ --do_train --dataset alpaca_gpt4_zh --finetuning_type lora --output_dir output --overwrite_cache --per_device_train_batch_size 4 --gradient_accumulation_steps 4 --lr_scheduler_type cosine --logging_steps 10 --save_steps 1000 --learning_rate 5e-5 --num_train_epochs 3.0 --plot_loss --fp16
训练报错ValueError: Target modules ['q_proj', 'v_proj'] not found in the base model. Please check the target modules and try again.
有没有大佬知道怎么解决,谢谢!
The text was updated successfully, but these errors were encountered:
加参数:--lora_target W_pack
Sorry, something went wrong.
百川的attention应该是把(Wq,Wk,Wv) concat成一个W_pack了
No branches or pull requests
train_sft.py训练指令:
CUDA_VISIBLE_DEVICES=0 python src/train_sft.py
--model_name_or_path /data1/projects/baichuan-7B/
--do_train
--dataset alpaca_gpt4_zh
--finetuning_type lora
--output_dir output
--overwrite_cache
--per_device_train_batch_size 4
--gradient_accumulation_steps 4
--lr_scheduler_type cosine
--logging_steps 10
--save_steps 1000
--learning_rate 5e-5
--num_train_epochs 3.0
--plot_loss
--fp16
训练报错ValueError: Target modules ['q_proj', 'v_proj'] not found in the base model. Please check the target modules and try again.
有没有大佬知道怎么解决,谢谢!
The text was updated successfully, but these errors were encountered: