-
Notifications
You must be signed in to change notification settings - Fork 2.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Question]:chatglmv2无法正确初始化 #8352
Labels
question
Further information is requested
Comments
请问你用的是什么环境,我测了一下没什么问题:
|
您好,我的版本如下 |
解决了,谢谢大佬,把python版本降到3.9.16就好了 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
请提出你的问题
Some weights of ChatGLMv2ForCausalLM were not initialized from the model checkpoint at /home/.paddlenlp/models/THUDM/chatglm2-6b and are newly initialized: ['encoder.layers.2.self_attention.key.weight', 'encoder.layers.0.self_attention.value.bias', 'encoder.layers.0.self_attention.key.weight', 'encoder.layers.11.self_attention.key.weight', 'encoder.layers.15.self_attention.query.weight', 'encoder.layers.24.self_attention.value.bias', 'encoder.layers.7.self_attention.key.weight', 'encoder.layers.24.self_attention.key.weight', 'encoder.layers.19.self_attention.query.weight', 'encoder.layers.11.self_attention.query.weight', 'encoder.layers.20.self_attention.key.bias', 'encoder.layers.23.self_attention.query.bias', 'encoder.layers.25.self_attention.query.weight', 'encoder.layers.4.self_attention.key.bias', 'encoder.layers.6.self_attention.value.bias', 'encoder.layers.16.self_attention.value.bias', 'encoder.layers.17.self_attention.key.bias', 'encoder.layers.21.self_attention.query.weight', 'encoder.layers.24.self_attention.query.weight', 'encoder.layers.26.self_attention.query.bias', 'encoder.layers.23.self_attention.key.bias', 'encoder.layers.23.self_attention.query.weight', 'encoder.layers.21.self_attention.key.weight', 'encoder.layers.25.self_attention.key.weight', 'encoder.layers.27.self_attention.value.bias', 'encoder.layers.2.self_attention.key.bias', 'encoder.layers.25.self_attention.value.weight', 'encoder.layers.20.self_attention.value.bias', 'encoder.layers.18.self_attention.key.weight', 'encoder.layers.12.self_attention.query.bias', 'encoder.layers.14.self_attention.query.bias', 'encoder.layers.5.self_attention.key.bias', 'encoder.layers.24.self_attention.value.weight', 'encoder.layers.17.self_attention.query.weight', 'encoder.layers.7.self_attention.value.weight', 'encoder.layers.18.self_attention.value.weight', 'encoder.layers.22.self_attention.query.weight', 'encoder.layers.12.self_attention.key.weight', 'encoder.layers.17.self_attention.value.bias', 'encoder.layers.13.self_attention.query.bias', 'encoder.layers.22.self_attention.key.bias', 'encoder.layers.1.self_attention.key.bias', 'encoder.layers.5.self_attention.key.weight', 'encoder.layers.26.self_attention.query.weight', 'encoder.layers.12.self_attention.query.weight', 'encoder.layers.0.self_attention.query.weight', 'encoder.layers.16.self_attention.query.weight', 'encoder.layers.27.self_attention.query.bias', 'encoder.layers.3.self_attention.query.weight', 'encoder.layers.25.self_attention.key.bias', 'encoder.layers.1.self_attention.query.weight', 'encoder.layers.5.self_attention.value.bias', 'encoder.layers.21.self_attention.query.bias', 'encoder.layers.17.self_attention.value.weight', 'encoder.layers.10.self_attention.key.weight', 'encoder.layers.22.self_attention.key.weight', 'encoder.layers.19.self_attention.key.bias', 'encoder.layers.24.self_attention.query.bias', 'encoder.layers.24.self_attention.key.bias', 'encoder.layers.21.self_attention.key.bias', 'encoder.layers.22.self_attention.query.bias', 'encoder.layers.6.self_attention.key.weight', 'encoder.layers.4.self_attention.value.bias', 'encoder.layers.13.self_attention.query.weight', 'encoder.layers.11.self_attention.query.bias', 'encoder.layers.2.self_attention.value.weight', 'encoder.layers.9.self_attention.key.bias', 'encoder.layers.26.self_attention.key.bias', 'encoder.layers.2.self_attention.query.weight', 'encoder.layers.3.self_attention.value.weight', 'encoder.layers.15.self_attention.value.bias', 'encoder.layers.22.self_attention.value.bias', 'encoder.layers.27.self_attention.key.weight', 'encoder.layers.13.self_attention.value.weight', 'encoder.layers.1.self_attention.value.weight', 'encoder.layers.27.self_attention.query.weight', 'encoder.layers.14.self_attention.query.weight', 'encoder.layers.9.self_attention.query.weight', 'encoder.layers.25.self_attention.query.bias', 'encoder.layers.12.self_attention.value.weight', 'encoder.layers.4.self_attention.query.weight', 'encoder.layers.17.self_attention.query.bias', 'encoder.layers.14.self_attention.value.weight', 'encoder.layers.10.self_attention.query.weight', 'encoder.layers.18.self_attention.query.weight', 'encoder.layers.3.self_attention.query.bias', 'encoder.layers.8.self_attention.query.bias', 'encoder.layers.2.self_attention.value.bias', 'encoder.layers.9.self_attention.query.bias', 'encoder.layers.27.self_attention.value.weight', 'encoder.layers.1.self_attention.value.bias', 'encoder.layers.10.self_attention.query.bias', 'encoder.layers.7.self_attention.value.bias', 'encoder.layers.9.self_attention.value.bias', 'encoder.layers.27.self_attention.key.bias', 'encoder.layers.5.self_attention.query.weight', 'encoder.layers.17.self_attention.key.weight', 'encoder.layers.25.self_attention.value.bias', 'encoder.layers.8.self_attention.query.weight', 'encoder.layers.19.self_attention.query.bias', 'encoder.layers.22.self_attention.value.weight', 'encoder.layers.12.self_attention.value.bias', 'encoder.layers.20.self_attention.query.weight', 'encoder.layers.12.self_attention.key.bias', 'encoder.layers.26.self_attention.value.bias', 'encoder.layers.0.self_attention.value.weight', 'encoder.layers.8.self_attention.value.weight', 'encoder.layers.11.self_attention.value.bias', 'encoder.layers.7.self_attention.query.bias', 'encoder.layers.23.self_attention.key.weight', 'encoder.layers.21.self_attention.value.weight', 'encoder.layers.14.self_attention.key.weight', 'encoder.layers.9.self_attention.value.weight', 'encoder.layers.8.self_attention.key.weight', 'encoder.layers.7.self_attention.key.bias', 'encoder.layers.13.self_attention.key.bias', 'encoder.layers.6.self_attention.query.weight', 'encoder.layers.11.self_attention.key.bias', 'encoder.layers.3.self_attention.key.weight', 'encoder.layers.15.self_attention.value.weight', 'encoder.layers.3.self_attention.key.bias', 'encoder.layers.9.self_attention.key.weight', 'encoder.layers.16.self_attention.key.weight', 'encoder.layers.10.self_attention.key.bias', 'encoder.layers.1.self_attention.query.bias', 'encoder.layers.5.self_attention.value.weight', 'encoder.layers.20.self_attention.query.bias', 'encoder.layers.18.self_attention.query.bias', 'encoder.layers.20.self_attention.key.weight', 'encoder.layers.14.self_attention.value.bias', 'encoder.layers.13.self_attention.key.weight', 'encoder.layers.4.self_attention.value.weight', 'encoder.layers.7.self_attention.query.weight', 'encoder.layers.16.self_attention.value.weight', 'encoder.layers.10.self_attention.value.bias', 'encoder.layers.21.self_attention.value.bias', 'encoder.layers.23.self_attention.value.weight', 'encoder.layers.26.self_attention.key.weight', 'encoder.layers.18.self_attention.value.bias', 'encoder.layers.6.self_attention.query.bias', 'encoder.layers.8.self_attention.value.bias', 'encoder.layers.18.self_attention.key.bias', 'encoder.layers.4.self_attention.query.bias', 'encoder.layers.3.self_attention.value.bias', 'encoder.layers.4.self_attention.key.weight', 'encoder.layers.20.self_attention.value.weight', 'encoder.layers.8.self_attention.key.bias', 'encoder.layers.19.self_attention.value.bias', 'encoder.layers.11.self_attention.value.weight', 'encoder.layers.6.self_attention.value.weight', 'encoder.layers.0.self_attention.query.bias', 'encoder.layers.5.self_attention.query.bias', 'encoder.layers.2.self_attention.query.bias', 'encoder.layers.15.self_attention.key.weight', 'encoder.layers.0.self_attention.key.bias', 'encoder.layers.26.self_attention.value.weight', 'encoder.layers.19.self_attention.key.weight', 'encoder.layers.13.self_attention.value.bias', 'encoder.layers.19.self_attention.value.weight', 'encoder.layers.1.self_attention.key.weight', 'encoder.layers.23.self_attention.value.bias', 'encoder.layers.15.self_attention.query.bias', 'encoder.layers.14.self_attention.key.bias', 'encoder.layers.6.self_attention.key.bias', 'encoder.layers.16.self_attention.query.bias', 'encoder.layers.10.self_attention.value.weight', 'encoder.layers.15.self_attention.key.bias', 'encoder.layers.16.self_attention.key.bias']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
权重文件是下载的model_state.pdparams,但是模型无法正确初始化,进而无法作出正确的预测
The text was updated successfully, but these errors were encountered: