Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Option to increase the number of input tokens for Falcon 40B Model #593

Closed
gabriead opened this issue Jul 12, 2023 · 4 comments
Closed

Option to increase the number of input tokens for Falcon 40B Model #593

gabriead opened this issue Jul 12, 2023 · 4 comments

Comments

@gabriead
Copy link

Feature request

I would like to increase the number of input tokens from currently 1024 to it's maximum of 2000 tokens.

Motivation

In order to answer questions given a specific context I want to use as many input tokens as possible.

Your contribution

I tried to extend the number of tokens in the openapi.json but that didn't help

@OlivierDehaene
Copy link
Member

You can use text-generation-launcher --help to see all the options available to you. The ones you might be interested in are --max-input-length and --max-total-tokens.

@gabriead
Copy link
Author

gabriead commented Jul 12, 2023

Hi @OlivierDehaene, thank's for your reply. I am using the generate() method from the client class. That does not provide any of the flags you mentioned. Where am I going wrong here? Could you please specify how I can use the text-generation launcher for inference?

@OlivierDehaene
Copy link
Member

Oh as a client you don't have the ability to modify these parameters. Are you using the Huggingface inference API?

@gabriead
Copy link
Author

gabriead commented Jul 13, 2023

I am not sure, what you mean by 'as a client'. I am using it like so:
from text_generation import Client client = Client("..") print(client.generate("...", max_new_tokens=...).generated_text)
And now I need to extend the number of input tokens to the maximum. Is there a way to integrate that into the framework?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants