Skip to main content
POST
/
v1
/
completions
{
  "id": "<string>",
  "created": 123,
  "model": "<string>",
  "choices": [
    {
      "index": 123,
      "text": "<string>",
      "logprobs": {
        "tokens": [
          "<string>"
        ],
        "token_logprobs": [
          123
        ],
        "top_logprobs": [
          {}
        ],
        "text_offset": [
          123
        ],
        "token_ids": [
          123
        ]
      },
      "finish_reason": "stop",
      "raw_output": {
        "prompt_fragments": [
          "<string>"
        ],
        "prompt_token_ids": [
          123
        ],
        "completion": "<string>",
        "completion_token_ids": [
          123
        ],
        "completion_logprobs": {
          "content": [
            {
              "token": "<string>",
              "logprob": 123,
              "sampling_logprob": 123,
              "bytes": [
                123
              ],
              "token_id": 123,
              "text_offset": 123,
              "top_logprobs": [
                {
                  "token": "<string>",
                  "logprob": 123,
                  "token_id": 123,
                  "bytes": [
                    123
                  ]
                }
              ],
              "last_activation": "<string>",
              "routing_matrix": "<string>",
              "extra_tokens": [
                123
              ],
              "extra_logprobs": [
                123
              ]
            }
          ]
        }
      },
      "prompt_token_ids": [
        123
      ],
      "token_ids": [
        123
      ]
    }
  ],
  "usage": {
    "prompt_tokens": 123,
    "total_tokens": 123,
    "completion_tokens": 123,
    "prompt_tokens_details": {
      "cached_tokens": 123
    }
  },
  "object": "text_completion",
  "perf_metrics": {}
}

Authorizations

Authorization
string
header
required

Bearer authentication using your Fireworks API key. Format: Bearer <API_KEY>

Body

application/json
model
string
required

The name of the model to use.

Example: "accounts/fireworks/models/kimi-k2-instruct-0905"

prompt
required

The prompt to generate completions for.

It can be a single string or an array of strings.

It can also be an array of integers or an array of integer arrays, which allows to pass already tokenized prompt.

If multiple prompts are specified, several choices with corresponding index will be returned in the output.

user
string | null

A unique identifier representing your end-user, which can help monitor and detect abuse.

prompt_cache_max_len
integer | null

Maximum length of the prompt to cache.

prompt_cache_isolation_key
string | null

Isolation key for prompt caching to separate cache entries.

raw_output
boolean | null
default:false

Return raw output from the model.

perf_metrics_in_response
boolean | null
default:false

Whether to include performance metrics in the response body.

Non-streaming requests: Performance metrics are always included in response headers (e.g., fireworks-prompt-tokens, fireworks-server-time-to-first-token). Setting this to true additionally includes the same metrics in the response body under the perf_metrics field.

Streaming requests: Performance metrics are only included in the response body under the perf_metrics field in the final chunk (when finish_reason is set). This is because headers may not be accessible during streaming.

The response body perf_metrics field contains the following metrics:

Basic Metrics (all deployments):

  • prompt-tokens: Number of tokens in the prompt
  • server-time-to-first-token: Time from request start to first token (in seconds)
  • server-processing-time: Total processing time (in seconds, only for completed requests)

Predicted Outputs Metrics:

  • speculation-prompt-tokens: Number of speculative prompt tokens
  • speculation-prompt-matched-tokens: Number of matched speculative prompt tokens (for completed requests)

Dedicated Deployment Only Metrics:

  • speculation-generated-tokens: Number of speculative generated tokens (for completed requests)
  • speculation-acceptance: Speculation acceptance rates by position
  • cached-prompt-tokens: Number of cached prompt tokens
  • backend-host: Hostname of the backend server
  • num-concurrent-requests: Number of concurrent requests
  • deployment: Deployment name
  • tokenizer-queue-duration: Time spent in tokenizer queue
  • tokenizer-duration: Time spent in tokenizer
  • prefill-queue-duration: Time spent in prefill queue
  • prefill-duration: Time spent in prefill
  • generation-queue-duration: Time spent in generation queue
stream
boolean | null
default:false

Whether to stream back partial progress. If set, tokens will be sent as data-only server-sent events as they become available, with the stream terminated by a data: [DONE] message.

n
integer
default:1

How many completions to generate for each prompt.

Note: Because this parameter generates many completions, it can quickly consume your token quota. Use carefully and ensure that you have reasonable settings for max_tokens and stop.

Required range: 1 <= x <= 128

Example: 1

stop

Up to 4 sequences where the API will stop generating further tokens. The returned text will NOT contain the stop sequence.

max_tokens
integer | null

The maximum number of tokens to generate in the completion. If the token count of your prompt plus max_tokens exceeds the model's context length, the behavior depends on context_length_exceeded_behavior. By default, max_tokens will be lowered to fit in the context window instead of returning an error.

max_completion_tokens
integer | null

Alias for max_tokens. Cannot be specified together with max_tokens.

temperature
number | null

What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic.

We generally recommend altering this or top_p but not both.

Required range: 0 <= x <= 2

Example: 1

top_k
integer | null

Top-k sampling is another sampling method where the k most probable next tokens are filtered and the probability mass is redistributed among only those k next tokens. The value of k controls the number of candidates for the next token at each step during text generation. Must be between 0 and 100.

Required range: 0 <= x <= 100

Example: 50

top_p
number | null

An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered.

We generally recommend altering this or temperature but not both.

Required range: 0 <= x <= 1

Example: 1

min_p
number | null

Minimum probability threshold for token selection. Only tokens with probability >= min_p are considered for selection. This is an alternative to top_p and top_k sampling.

Required range: 0 <= x <= 1

typical_p
number | null

Typical-p sampling is an alternative to nucleus sampling. It considers the most typical tokens whose cumulative probability is at most typical_p.

Required range: 0 <= x <= 1

frequency_penalty
number | null

Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim.

Reasonable value is around 0.1 to 1 if the aim is to just reduce repetitive samples somewhat. If the aim is to strongly suppress repetition, then one can increase the coefficients up to 2, but this can noticeably degrade the quality of samples. Negative values can be used to increase the likelihood of repetition.

See also presence_penalty for penalizing tokens that have at least one appearance at a fixed rate.

OpenAI compatible (follows OpenAI's conventions for handling token frequency and repetition penalties).

Required range: -2 <= x <= 2

presence_penalty
number | null

Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.

Reasonable value is around 0.1 to 1 if the aim is to just reduce repetitive samples somewhat. If the aim is to strongly suppress repetition, then one can increase the coefficients up to 2, but this can noticeably degrade the quality of samples. Negative values can be used to increase the likelihood of repetition.

See also frequency_penalty for penalizing tokens at an increasing rate depending on how often they appear.

OpenAI compatible (follows OpenAI's conventions for handling token frequency and repetition penalties).

Required range: -2 <= x <= 2

repetition_penalty
number | null

Applies a penalty to repeated tokens to discourage or encourage repetition. A value of 1.0 means no penalty, allowing free repetition. Values above 1.0 penalize repetition, reducing the likelihood of repeating tokens. Values between 0.0 and 1.0 reward repetition, increasing the chance of repeated tokens. For a good balance, a value of 1.2 is often recommended. Note that the penalty is applied to both the generated output and the prompt in decoder-only models.

Required range: 0 <= x <= 2

mirostat_target
number | null

Defines the target perplexity for the Mirostat algorithm. Perplexity measures the unpredictability of the generated text, with higher values encouraging more diverse and creative outputs, while lower values prioritize predictability and coherence. The algorithm dynamically adjusts the token selection to maintain this target during text generation.

If not specified, Mirostat sampling is disabled.

mirostat_lr
number | null

Specifies the learning rate for the Mirostat sampling algorithm, which controls how quickly the model adjusts its token distribution to maintain the target perplexity. A smaller value slows down the adjustments, leading to more stable but gradual shifts, while higher values speed up corrections at the cost of potential instability.

seed
integer | null

Random seed for deterministic sampling.

logprobs

Include log probabilities in the response. This accepts either a boolean or an integer:

If set to true, log probabilities are included and the number of alternatives can be controlled via top_logprobs (OpenAI-compatible behavior).

If set to an integer N (0-5), include log probabilities for up to N most likely tokens per position in the legacy format.

The API will always return the logprob of the sampled token, so there may be up to logprobs+1 elements in the response when an integer is used. The maximum value for the integer form is 5.

top_logprobs
integer | null

An integer between 0 and 5 specifying the number of most likely tokens to return at each token position, each with an associated log probability. The minimum value is 0 and the maximum value is 5.

When logprobs is set, top_logprobs can be used to modify how many top log probabilities are returned. If top_logprobs is not set, the API will return up to logprobs tokens per position.

Required range: 0 <= x <= 5

echo
boolean | null
default:false

Echo back the prompt in addition to the completion.

echo_last
integer | null

Echo back the last N tokens of the prompt in addition to the completion. This is useful for obtaining logprobs of the prompt suffix but without transferring too much data. Passing echo_last=len(prompt) is the same as echo=True

ignore_eos
boolean
default:false

This setting controls whether the model should ignore the End of Sequence (EOS) token. When set to True, the model will continue generating tokens even after the EOS token is produced. By default, it stops when the EOS token is reached.

context_length_exceeded_behavior
enum<string>
default:truncate

What to do if the token count of prompt plus max_tokens exceeds the model's context window.

Passing truncate limits the max_tokens to at most context_window_length - prompt_length. This is the default.

Passing error would trigger a request error.

The default of 'truncate' is selected as it allows to ask for high max_tokens value while respecting the context window length without having to do client-side prompt tokenization.

Note, that it differs from OpenAI's behavior that matches that of error.

Available options:
error,
truncate
response_format
ResponseFormat · object

Allows to force the model to produce specific output format.

Setting to { "type": "json_object" } enables JSON mode, which guarantees the message the model generates is valid JSON.

If "type" is "json_schema", a JSON schema must be provided. E.g., response_format = {"type": "json_schema", "json_schema": <json_schema>}.

Important: when using JSON mode, it's crucial to also instruct the model to produce JSON via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly "stuck" request.

Also note that the message content may be partially cut off if finish_reason="length", which indicates the generation exceeded max_tokens or the conversation exceeded the max context length. In this case the return value might not be a valid JSON.

logit_bias
Logit Bias · object

Modify the likelihood of specified tokens appearing in the completion. Accepts a json object that maps tokens (specified by their token ID in the tokenizer) to an associated bias value from -100 to 100. Mathematically, the bias is added to the logits generated by the model prior to sampling.

speculation

Speculative decoding prompt or token IDs to speed up generation.

prediction

OpenAI-compatible predicted output for speculative decoding. Can be a PredictedOutput object or a simple string. Automatically transformed to speculation.

metadata
Metadata · object

Additional metadata to store with the request for tracing/distillation.

reasoning_effort

Applicable to reasoning models only, this option controls the reasoning token length, and can be set to either 'none', 'low', 'medium', 'high' or an integer. 'low', 'medium' and 'high' correspond to progressively higher thinking effort and thus longer reasoning tokens. 'none' means disable thinking. You can alternatively set the option to an integer controlling the hard-cutoff for reasoning token length (this is not entirely OpenAI compatible, you might have to use fireworks.ai client library to bypass the schema check). Note: For OpenAI GPT OSS models, only the string values ('low', 'medium', 'high') are supported. Integer values will not work with these models.

Available options:
low,
medium,
high,
none
return_token_ids
boolean | null
default:false

Return token IDs alongside text to avoid retokenization drift.

images

The list of base64 encoded images for visual language completition generation.

They should be formatted as MIME_TYPE,

eg. data:image/jpeg;base64,

Additionally, the number of images provided should match the number of '' special token in the prompt

Response

Successful Response

The response message from a /v1/completions call.

id
string
required

A unique identifier of the response

created
integer
required

The Unix time in seconds when the response was generated

model
string
required

The model used for the completion

choices
Choice · object[]
required

The list of generated completion choices

usage
UsageInfo · object
required

Usage statistics for the completion

object
string
default:text_completion

The object type, which is always "text_completion"

perf_metrics
Perf Metrics · object