Skip to content

Latest commit

 

History

History
22 lines (19 loc) · 2.5 KB

CompletionPayload.md

File metadata and controls

22 lines (19 loc) · 2.5 KB

# CompletionPayload

Properties

Name Type Description Notes
model string ID of the model to use. [optional]
prompt string | array The prompt(s) to generate completions for, encoded as a string, a list of strings, or a list of token lists. [optional]
max_tokens int The maximum number of tokens to generate. Requests can use up to 2048 tokens shared between prompt and completion. [optional] [default to 16]
temperature float The maximum number of tokens to generate. Requests can use up to 2048 tokens shared between prompt and completion. [optional] [default to 1]
top_p float An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. [optional] [default to 1]
n int How many completions to generate for each prompt. [optional] [default to 1]
stream bool Whether to stream back partial progress. If set, tokens will be sent as data-only server-sent events as they become available, with the stream terminated by a data: [DONE] message. [optional] [default to false]
logprobs int Include the log probabilities on the logprobs most likely tokens, as well the chosen tokens. [optional]
echo bool Echo back the prompt in addition to the completion. [optional] [default to false]
stop string | array Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence. [optional]
presence_penalty float Number between 0 and 1 that penalizes new tokens based on whether they appear in the text so far. Increases the model's likelihood to talk about new topics. [optional] [default to 0]
frequency_penalty float Number between 0 and 1 that penalizes new tokens based on their existing frequency in the text so far. Decreases the model's likelihood to repeat the same line verbatim. [optional] [default to 0]
best_of int Generates best_of completions server-side and returns the "best" (the one with the lowest log probability per token). Results cannot be streamed. [optional] [default to 1]
logit_bias object Modify the likelihood of specified tokens appearing in the completion. [optional]

[Back to Model list] [Back to API list] [Back to README]