You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Describe the feature or improvement you're requesting
The chatml.md document said every message is represented as
<|im_start|>{role}\n{text}<|im_end|>\n
That would add 5 more tokens beside the text itself.
This is correct for the gpt-3.5-turbo-0301 model. As the gpt-3.5-turbo-0613 and gpt-4 models, it seems that every message is represneted as the following form
The original ChatML documentation was true for the gpt-3.5-turbo-0301 model, but the details of ChatML have since changed for subsequent models. We don't have current plans to document the underlying structure, but are looking into better tools for estimating token count up front.
Describe the feature or improvement you're requesting
The chatml.md document said every message is represented as
That would add 5 more tokens beside the text itself.
This is correct for the gpt-3.5-turbo-0301 model. As the gpt-3.5-turbo-0613 and gpt-4 models, it seems that every message is represneted as the following form
For eample
Accorsing to the response, the prompt_tokens is 8, if we follow the rule described in the chatml.md, it's represened as
That would be 9 tokens. My guess is that there's no need for the '\n' after the <|im_end|> token.
How can I make sure for that? Thanks.
Additional context
No response
The text was updated successfully, but these errors were encountered: