Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[DOCS] add clarification in distilabel vLLM reference to specify dtype #326

Closed
kcentric opened this issue Feb 5, 2024 · 1 comment
Closed

Comments

@kcentric
Copy link
Contributor

kcentric commented Feb 5, 2024

Which page or section is this issue related to?

Currently the code snippet in the vLLM section of the guide (https://distilabel.argilla.io/latest/technical-reference/llms/#vllm) looks like:

llm = vLLM(
    model=LLM(model="argilla/notus-7b-v1"),
    task=TextGenerationTask(),
...

Running this as-is in a Colab notebook will result in "ValueError: Bfloat16 is only supported on GPUs with compute capability of at least 8.0. Your Tesla T4 GPU has compute capability 7.5." This is discussed in vLLM issues here.

Because everyone who uses Colab would likely use the T4 GPU (if they're on the free Colab at least), they'd face the same error every time if they copy our snippet from the Docs and try to test-run vLLM in their notebook.

I'd want to change the snippet to something like this:

llm = vLLM(
    model=LLM(model="argilla/notus-7b-v1", dtype="Bfloat16"),  # If using Tesla T4 on Colab, 
                                                      # specify dtype = float16 to prevent compute compatibility error
    task=TextGenerationTask(),

and add a brief clarification about it in the text with a link for anyone who wants to understand it further.

@davidberenstein1957
Copy link
Member

Hi @kcentric , feel free to create a PR for this.

@alvarobartt alvarobartt closed this as not planned Won't fix, can't repro, duplicate, stale May 7, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants