Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ENH: make deepseek_vl support streaming output #1444

Merged
merged 8 commits into from
May 10, 2024

Conversation

Minamiyama
Copy link
Contributor

QQ202458-102123.mp4

@XprobeBot XprobeBot added the enhancement New feature or request label May 8, 2024
@XprobeBot XprobeBot added this to the v0.11.0 milestone May 8, 2024
@sunhaha123
Copy link

Looks great! Could this model run on 24gb gpu?

@sunhaha123
Copy link

sunhaha123 commented May 9, 2024

Looks great! Could this model run on 24gb gpu?

seems fp16 works.
using:vl_gpt: MultiModalityCausalLM = AutoModelForCausalLM.from_pretrained( # type: ignore self.model_path, trust_remote_code=True, device_map=self._device, low_cpu_mem_usage=True, torch_dtype=torch.float16 )

@Minamiyama
Copy link
Contributor Author

Looks great! Could this model run on 24gb gpu?

yes, I run it on gtx4090

@qinxuye qinxuye merged commit 0cb0f0e into xorbitsai:main May 10, 2024
12 checks passed
@sunhaha123
Copy link

sunhaha123 commented May 10, 2024

  File "/home/echo/miniconda3/envs/xinfer/lib/python3.10/site-packages/xinference/model/llm/pytorch/deepseek_vl.py", line 190, in chat
    from ....thirdparty.deepseek_vl.serve.inference import generate
ModuleNotFoundError: [address=0.0.0.0:42331, pid=2580143] No module named 'xinference.thirdparty.deepseek_vl.serve'

I pip install after git pull from main branch. Occur this problem.

@Minamiyama
Copy link
Contributor Author

  File "/home/echo/miniconda3/envs/xinfer/lib/python3.10/site-packages/xinference/model/llm/pytorch/deepseek_vl.py", line 190, in chat
    from ....thirdparty.deepseek_vl.serve.inference import generate
ModuleNotFoundError: [address=0.0.0.0:42331, pid=2580143] No module named 'xinference.thirdparty.deepseek_vl.serve'

I pip install after git pull from main branch. Occur this problem.

the module codes is in source codes:
image

maybe you should build from source code, or put the codes in xinference/thirdparty/deepseek_vl to your package path (I solved it by this way)

@sunhaha123
Copy link

maybe you should build from source code, or put the codes in xinference/thirdparty/deepseek_vl to your package path (I solved it by this way)

It works. Still feeling strange, I do pip install from source code.

@qinxuye
Copy link
Contributor

qinxuye commented May 10, 2024

Did you update the main code?

@sunhaha123
Copy link

Did you update the main code?

git pull origin main, and pip install .[transformers]

@dimitribellini
Copy link

Dear Team,
I'm new to "xinference" and I do not understand so well on how to use your inference server and the Deepseek VL model.
Using the docker option I was able to run it and load the module using the engine option "transformer" but as soon a connect using OpenWebUI through the OpenAI API I received the message reported from @Minamiyama
How I can solve it? Your help it's very appreciate :-)
Thanks so much

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

6 participants