-
-
Notifications
You must be signed in to change notification settings - Fork 10.5k
add QWen-7b support #685
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add QWen-7b support #685
Conversation
# Adapted from | ||
# https://huggingface.co/Qwen/Qwen-7B/blob/main/modeling_qwen.py | ||
# Copyright (c) Alibaba Cloud. | ||
# LICENSE: https://huggingface.co/Qwen/Qwen-7B/blob/main/LICENSE |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I noticed that here we introduced a new license different from this project. Will this license bring some potential legal risks for commercial use?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Qwen's license allows commercial use. The following is the original text from the Qwen license.
If you are commercially using the Materials, and your product or service has more than 100 million monthly active users, You shall request a license from Us.
I noticed that the baichuan model does not include the license of the original project. Do we need to add Qwen's license in this PR?
baichuan's original license: For commercial use, you need to send an email to apply for a license.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Maybe we can ask owners for some advice🤔.
Hi @WoosukKwon and @zhuohan123, should we follow the model's license when adding new models support in vLLM?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I believe we need to follow the model code's license instead of the model's licenses. For Baichuan (and LLaMA), the code is Apache 2 but the model has a special license. In this case we can include the code with no problem. However, for Qwen, both its code and model are with their restricted license. I believe we should include the link to Qwen's license on the top of Qwen's source file for safety.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you for your great contribution! LGTM!
I installed vllm from commit. and used qwen-7b-chat, generations are good to about 500 token lenght. then...
|
try set Line 37 in 66c54aa
stop=["<|im_end|>", "<|endoftext|>", "<|im_start|>"] |
Thank you! makes sense! |
请问用vllm, qwen最大长度可以是8k吗 |
Do you have a test comparison? Does it differ from the results without |
Co-authored-by: wq.chu <wq.chu@tianrang-inc.com>
https://github.com/QwenLM/Qwen-7B
tp=1 output:

tp=2 output:
