Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Same generate params but got totally different result #706

Closed
luohao123 opened this issue Aug 8, 2023 · 4 comments
Closed

Same generate params but got totally different result #706

luohao123 opened this issue Aug 8, 2023 · 4 comments

Comments

@luohao123
Copy link

Hi, I have a params with do_sample False, the result on huggingface are pretty good.

But same set in vllm, the model likes a stupid bird.

sampling_params = SamplingParams(
        temperature=generating_args.temperature,
        top_p=generating_args.top_p,
        max_tokens=800,
        frequency_penalty=generating_args.repetition_penalty,
        use_beam_search=False,
        ignore_eos=False,
    )

Please let me know how to resolve the bias?

same args:

do_sample: Optional[bool] = field(default=True)
    temperature: Optional[float] = field(
        default=0.84,
        metadata={"help": "The value used to modulate the next token probabilities."},
    )
    top_p: Optional[float] = field(
        default=0.6,
        metadata={
            "help": "The smallest set of most probable tokens with probabilities that add up to top_p or higher are kept."
        },
    )
    top_k: Optional[int] = field(default=40)
    num_beams: Optional[int] = field(default=1)
    max_new_tokens: Optional[int] = field(default=800)
    repetition_penalty: Optional[float] = field(default=1.09)

Both set do_sample False. but the result are out of my scope, they are totally different.

@Zhuqln
Copy link

Zhuqln commented Aug 9, 2023

can you show the result about the difference?

@BaiMoHan
Copy link

BaiMoHan commented Aug 9, 2023

This problem may be related to this #712 #450

@luohao123
Copy link
Author

@Zhuqln hard to tell, but same question, vllm can not answer correctly.
this caused a massive performance reduce compare with hf transformers....

@hmellor
Copy link
Collaborator

hmellor commented Mar 8, 2024

Closing this issue as stale as there has been no discussion in the past 3 months.

If you are still experiencing the issue you describe, feel free to re-open this issue.

@hmellor hmellor closed this as completed Mar 8, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants