Skip to content

Conversation

imgaojun
Copy link
Contributor

@imgaojun imgaojun commented Dec 8, 2023

This PR addresses the issue where loading LLaMA model parameters into vLLM results in a KeyError due to the presence of rotary_emb.cos_cached and rotary_emb.sin_cached. These cached values from LLaMA are not required or directly managed by vLLM, leading to compatibility issues.

Related issue #1977

Changes Made:

  • Added conditional checks to ignore rotary_emb.cos_cached and rotary_emb.sin_cached during the loading process.
  • Ensured that other necessary parameters are loaded correctly, without interfering with the rotary embeddings.

allenhaozi added a commit to allenhaozi/vllm that referenced this pull request Dec 8, 2023
Copy link
Collaborator

@Yard1 Yard1 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants