Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Don't build punica kernels by default #2605

Merged
merged 5 commits into from
Jan 26, 2024

Conversation

pcmoritz
Copy link
Collaborator

@pcmoritz pcmoritz commented Jan 25, 2024

The punica kernels take quite a while to build and most developers won't need them, so it will be better to not build them by default.

Maybe in the future the build can be optimized and then we can bring them back. E.g. if we had good incremental compilation that would probably solve the problem.

Fixes #2604

@Yard1
Copy link
Collaborator

Yard1 commented Jan 25, 2024

we should make sure CI builds it! let's set the env var there?

@pcmoritz
Copy link
Collaborator Author

Yes, I'm running the CI now, will see what fails and then fix it :)

Also we need to make sure the final wheel is built with LoRA, do you know where that's built?

@Yard1
Copy link
Collaborator

Yard1 commented Jan 25, 2024

.buildkite/test-template.j2 I think? @simon-mo would you know?

@simon-mo
Copy link
Collaborator

Final wheels is here https://github.com/vllm-project/vllm/blob/main/.github/workflows/publish.yml
For CI, the build step is all in Dockerfile.

@pcmoritz
Copy link
Collaborator Author

@simon-mo Let me know what you think, I believe I put the environment variables at the right places (let's see if the tests agree) :)

Copy link
Collaborator

@Yard1 Yard1 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@simon-mo simon-mo merged commit 390b495 into vllm-project:main Jan 26, 2024
17 checks passed
NikolaBorisov pushed a commit to deepinfra/vllm that referenced this pull request Jan 31, 2024
hongxiayang pushed a commit to hongxiayang/vllm that referenced this pull request Feb 13, 2024
alexm-neuralmagic pushed a commit to neuralmagic/nm-vllm that referenced this pull request Feb 13, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Build times too long since LoRA has been merged
3 participants