-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
OpenAI supports top_p = 0.0
and top_p = 1.0
but TGI fails with a validation error with either of these values.
#2222
Comments
text-generation-inference/router/src/validation.rs Lines 248 to 255 in 8511669
If you want to set top_p to 1.0, you can simply sending the top_p as none, which will result in the default value of 1.0 being applied. It seems like the equal condition in the code is causing an error. |
This doesn't provide a resolution to the issue. The docker container rejects Is there an easy way to "patch" the container and deploy using the patched version?
See also: guidance-ai/guidance#945 |
Don't error on OpenAI `top_p` valid values. * Closes: guidance-ai/guidance#945 * Closes: huggingface#2222
Thank you for bringing this to our attention and for making the PR 👍 As far as I can tell, there shouldn't be anything blocking for getting this merged in. I'll approve running the CI and can take over the merging of the PR. |
This issue is stale because it has been open 30 days with no activity. Remove stale label or comment or this will be closed in 5 days. |
@ErikKaum I ran into this issue as I switched out API base urls and suddenly my script broke as the new API uses TGI which doesnt allow |
Hi @cornzz 👋 I understand that it's annoying that it breaks the client. But I think still for now we're opting for a clear error VS discarding user input without letting the user know. But if there's a lot of demand for the Hopefully this makes sense to you 👍 |
Depending on the client software. It could result in a breakage that
prevents any use of TGI by a customer.
…-Michael Conrad
Telephone: +1.678.934.3989
Email: ***@***.***
Telegram: https://t.me/mconrad202
Mastodon: ***@***.***
On Tue, Sep 3, 2024 at 9:47 AM Erik Kaunismäki ***@***.***> wrote:
Hi @cornzz <https://github.com/cornzz> 👋
I understand that it's annoying that it breaks the client. But I think
still for now we're opting for a clear error VS discarding user input
without letting the user know.
But if there's a lot of demand for the top_p=1.0 == no top_p alternative
we're still open. One good way to get an "indication of demand" for that
would be e.g. to get thumbs up on a issue that's a feature request.
Hopefully this makes sense to you 👍
—
Reply to this email directly, view it on GitHub
<#2222 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ABH4XNG7FICQDFNQVTEVASLZUW4WRAVCNFSM6AAAAABKXPZBUWVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGMRWGU3TIMBXGQ>
.
You are receiving this because you commented.Message ID:
***@***.***>
|
Hey @ErikKaum, thanks for your quick response! Its not a problem, I was reusing a script and I am wondering why the authors set Still, and sorry if I am misunderstanding something, but what do you mean by discarding user input? Maybe I am missing something, but why can't |
Glad to hear that it's not a problem 👍
No worries. So I'm pretty sure by not having a This is where it get's processed as an
And then on the model logic side the branch is conditioned on text-generation-inference/server/text_generation_server/utils/logits_process.py Lines 26 to 49 in 6cb42f4
There might be something I missed here or misunderstood in your question. |
Ah okay, sorry then it was a misunderstanding on my side, I assumed from this comment above, that it defaults to |
i still dont understand why |
System Info
Information
Tasks
Reproduction
Fails
Works
Expected behavior
See also: #1896 where the patch did not address this issue even though raised as part of the ticket.
Impact
This generally breaks libraries like
guidance
where the library is hard coded to usetop_p=1.0
for the OpenAI interface.The text was updated successfully, but these errors were encountered: