Bearer authentication header of the form Bearer <token>, where <token> is your auth token.
The unique identifier of the model to be used for generating the completion
The input text that serves as the starting point for the AI to generate a completion
The number of completion choices to generate and return. The model will generate multiple completions and return the best one(s) based on the specified criteria
When set to true, the API will include the original prompt in the completion response, effectively echoing it back
A value between -2.0 and 2.0 that penalizes new tokens based on their frequency in the text so far. Higher values decrease the model's likelihood to repeat the same lines verbatim
A dictionary that allows fine-tuning the likelihood of specified tokens appearing in the completion. Each key is a token, and the value is the bias (between -100 and 100)
The number of most likely tokens to return with their log probabilities. If specified, the API will return a list of the most likely tokens for each position
The maximum number of tokens to generate in the completion. The total length of input tokens and generated tokens is limited by the model's context length
The number of completions to generate for each prompt. Note that this may conflict with best_of if both are specified
A value between -2.0 and 2.0 that penalizes new tokens based on whether they appear in the text so far. Positive values increase the model's likelihood to talk about new topics
A seed for deterministic sampling. Using the same seed with the same parameters will generate the same completion
Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence
4If set to true, partial message deltas will be sent as data-only server-sent events. Tokens will be sent as they become available
Additional options to configure the behavior of streaming responses
A value between 0 and 2 that controls the randomness of the completion. Lower values make the output more focused and deterministic, while higher values make it more random
An alternative to temperature, called nucleus sampling. It considers the results of the tokens with top_p probability mass. 0.1 means only the tokens comprising the top 10% probability mass are considered
A unique identifier representing your end-user, which can help the API to monitor and detect abuse