We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
回答如下:
显然这些问答是有问题的。
正常回答如下:
我想知道是只有在不添加LoRA模块、不使用量化时才会触发gpu_memory_utilization=0.9参数吗?这个参数与vllm的有关系吗?是什么原因导致模型回答异常?以下是部分日志:
感谢各位回答!
The text was updated successfully, but these errors were encountered:
No branches or pull requests
不量化、不添加LoRA,加载ChatGLM3-6B模型
回答如下:
显然这些问答是有问题的。
但是!!
正常回答如下:
Why
我想知道是只有在不添加LoRA模块、不使用量化时才会触发gpu_memory_utilization=0.9参数吗?这个参数与vllm的有关系吗?是什么原因导致模型回答异常?以下是部分日志:
感谢各位回答!
The text was updated successfully, but these errors were encountered: