Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add RoPE scaling arguments to engine #220

Open
AlpinDale opened this issue Jan 26, 2024 · 4 comments
Open

Add RoPE scaling arguments to engine #220

AlpinDale opened this issue Jan 26, 2024 · 4 comments
Labels
enhancement New feature or request

Comments

@AlpinDale
Copy link
Member

Currently, we auto-scale using the --max-model-len argument. It may be more appropriate to have specific options for the scaling factor, etc.

@AlpinDale AlpinDale added the enhancement New feature or request label Jan 26, 2024
@jagilley
Copy link

There are some models for long context tasks like storywriting that it'd be nice to use with a static RoPE scaling factor. +1 on this!

@sparsh35
Copy link

Hi getting error for bigger context models like Microsoft Phi 3 medium with respect to rope scaling factors with exl2 format.

@sparsh35
Copy link

It is something related to this I think, maybe not much needs to be done here, just implement this code , I will try to test if it doesn't breaks anything else , here is the git in vllm for this feature vllm-project/vllm#4638

@sparsh35
Copy link

vllm-project/vllm#4298 vllm has implemented rotatry scale embeddings like this

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

3 participants