Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

I would like the --tensor-parallel-size setting of the vllm backend to be added to yaml. #2084

Closed
Taikono-Himazin opened this issue Apr 20, 2024 · 1 comment · Fixed by #2085
Labels
enhancement New feature or request

Comments

@Taikono-Himazin
Copy link
Contributor

Taikono-Himazin commented Apr 20, 2024

Is your feature request related to a problem? Please describe.

No similar issues were found.

Describe the solution you'd like

I would like it to be possible to set --tensor-parallel-size from yaml, similar to gpu_memory_utilization and trust_remote_code in the vllm backend.
https://docs.vllm.ai/en/latest/models/engine_args.html

The current settings only allow inference on a single GPU.
This setting is necessary to make this compatible with multiple GPUs.

@Taikono-Himazin Taikono-Himazin added the enhancement New feature or request label Apr 20, 2024
@localai-bot
Copy link
Contributor

Hello Taikono-Himazin,

Thank you for taking the time to submit an enhancement request to the LocalAI project.

Based on your description, it seems that adding the --tensor-parallel-size setting to the vllm backend configuration through a YAML file would indeed be a useful enhancement. This would allow users to leverage multiple GPUs for inference, which is an essential capability for handling larger workloads.

I will pass your suggestion along to the project maintainers. They will review it and decide whether or not to implement it. You may not receive a direct reply, but rest assured, your request has been noted and will be considered.

Remember, LocalAI is a community-driven project, so if you would like to contribute to its development, please feel free to join the community and helps.

Thank you once again for your feedback, and we look forward to seeing more contributions from you!


I am a bot, an experiment of @mudler.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants