Skip to content

Conversation

@seanhandley
Copy link
Contributor

No description provided.

@seanhandley seanhandley self-assigned this Oct 27, 2025
@seanhandley seanhandley requested review from a team as code owners October 27, 2025 15:36
@github-actions
Copy link

github-actions bot commented Oct 27, 2025

🔍 Preview links for changed docs

Comment on lines 60 to 68
EIS is billed per million "tokens" used. Tokens can be thought of loosely as "words" which are given to a machine learning model to operate upon. The model may also produce a number of tokens in response.

For example, the sentence:

"It was the best of times, it was the worst of times."

contains 52 characters, but would be tokenised into 14 tokens - one for each of the 12 words, one for the comma, and one for the period character.

This is because machine learning models use words to denote meaning.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'd revise to something like this:

Token-based billing

EIS is billed per million tokens used. Tokens are the fundamental units that language models process for both input and output.

Tokenizers convert text into numerical data by segmenting it into subword units. A token may be a complete word, part of a word, or a punctuation mark, depending on the model's trained tokenizer and the frequency patterns in its training data.

For example, the sentence "It was the best of times, it was the worst of times." contains 52 characters but would tokenize into approximately 14 tokens with a typical word-based approach, though the exact count varies by tokenizer.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks @leemthompo !

Copy link
Contributor

@leemthompo leemthompo left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

thanks for the docs updates @seanhandley 👍

@seanhandley seanhandley enabled auto-merge (squash) October 27, 2025 15:48
@leemthompo leemthompo disabled auto-merge October 27, 2025 15:49
Added information about token usage in conversations.
@leemthompo
Copy link
Contributor

@seanhandley added a bit of practical info to be super clear, lmkwyt

@seanhandley
Copy link
Contributor Author

Thanks @leemthompo !

One point here is that we do bill for output tokens for Chat models. But for embeddings models, we only bill for input tokens into the model - we don't bill for the embeddings that the model generates.

I think it's important to be specific here.

@leemthompo
Copy link
Contributor

good point, let me clarify that 👍

@leemthompo leemthompo changed the title Adding more info on tokens [EIS] Adding more info on tokens Oct 27, 2025
@seanhandley seanhandley merged commit 2d56513 into main Oct 27, 2025
7 checks passed
@seanhandley seanhandley deleted the add_token_exp branch October 27, 2025 16:29
rhr323 pushed a commit to rhr323/docs-content that referenced this pull request Oct 27, 2025
Co-authored-by: Liam Thompson <leemthompo@gmail.com>
naemono pushed a commit to naemono/docs-content that referenced this pull request Oct 28, 2025
Co-authored-by: Liam Thompson <leemthompo@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants