Completions
Completion Request
Available options: llama3.1-8b
, llama3.1-70b
The prompt(s) to generate completions for, encoded as a string, array of strings, array of tokens, or array of token arrays.
The maximum number of tokens that can be generated in the chat completion. The total length of input tokens and generated tokens is limited by the model’s context length.
What sampling temperature to use, between 0 and 1.5. Higher values make the output more random, while lower values make it more focused and deterministic.
An alternative to sampling with temperature. Model considers tokens with top_p probability mass.
Whether to stream back partial progress.
Echo back the prompt in addition to the completion.
Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence.
A unique identifier representing your end-user, which can help Cerebras to monitor and detect abuse.
If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same seed
and parameters should return the same result. Determinism is not guaranteed.
Completion Response
The list of completion choices the model generated for the input prompt.
The Unix timestamp (in seconds) of when the completion was created.
A unique identifier for the completion.
The model used for completion.
The object type, which is always “text_completion”
This fingerprint represents the backend configuration that the model runs with.
Can be used in conjunction with the seed
request parameter to understand when backend changes have been made that might impact determinism.
Usage statistics for the completion request.
Was this page helpful?