List Models
curl --request GET \
--url https://api.fireworks.ai/v1/accounts/{account_id}/models \
--header 'Authorization: Bearer <token>'
{
"models": [
{
"name": "<string>",
"displayName": "<string>",
"description": "<string>",
"createTime": "2023-11-07T05:31:56Z",
"createdBy": "<string>",
"state": "STATE_UNSPECIFIED",
"status": {
"code": "OK",
"message": "<string>"
},
"kind": "KIND_UNSPECIFIED",
"githubUrl": "<string>",
"huggingFaceUrl": "<string>",
"baseModelDetails": {
"worldSize": 123,
"checkpointFormat": "CHECKPOINT_FORMAT_UNSPECIFIED",
"parameterCount": "<string>",
"moe": true,
"tunable": true,
"modelType": "<string>"
},
"peftDetails": {
"baseModel": "<string>",
"r": 123,
"targetModules": [
"<string>"
],
"baseModelType": "<string>",
"mergeAddonModelName": "<string>"
},
"teftDetails": {},
"public": true,
"conversationConfig": {
"style": "<string>",
"system": "<string>",
"template": "<string>"
},
"contextLength": 123,
"supportsImageInput": true,
"supportsTools": true,
"importedFrom": "<string>",
"fineTuningJob": "<string>",
"defaultDraftModel": "<string>",
"defaultDraftTokenCount": 123,
"precisions": [
"PRECISION_UNSPECIFIED"
],
"deployedModelRefs": [
{
"name": "<string>",
"deployment": "<string>",
"state": "STATE_UNSPECIFIED",
"default": true,
"public": true
}
],
"cluster": "<string>",
"deprecationDate": {
"year": 123,
"month": 123,
"day": 123
},
"calibrated": true,
"tunable": true,
"supportsLora": true,
"useHfApplyChatTemplate": true
}
],
"nextPageToken": "<string>",
"totalSize": 123
}
Authorizations
Bearer authentication header of the form Bearer <token>
, where <token>
is your auth token.
Path Parameters
The Account Id
Query Parameters
The maximum number of models to return. The maximum page_size is 200, values above 200 will be coerced to 200. If unspecified, the default is 50.
A page token, received from a previous ListModels call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to ListModels must match the call that provided the page token.
Only model satisfying the provided filter (if specified) will be returned. See https://google.aip.dev/160 for the filter grammar.
A comma-separated list of fields to order by. e.g. "foo,bar" The default sort order is ascending. To specify a descending order for a field, append a " desc" suffix. e.g. "foo desc,bar" Subfields are specified with a "." character. e.g. "foo.bar" If not specified, the default order is by "name".
Response
Human-readable display name of the model. e.g. "My Model" Must be fewer than 64 characters long.
The description of the model. Must be fewer than 1000 characters long.
The creation time of the model.
The email address of the user who created this model.
The state of the model.
STATE_UNSPECIFIED
, UPLOADING
, READY
Contains detailed message when the last model operation fails.
The status code.
OK
, CANCELLED
, UNKNOWN
, INVALID_ARGUMENT
, DEADLINE_EXCEEDED
, NOT_FOUND
, ALREADY_EXISTS
, PERMISSION_DENIED
, UNAUTHENTICATED
, RESOURCE_EXHAUSTED
, FAILED_PRECONDITION
, ABORTED
, OUT_OF_RANGE
, UNIMPLEMENTED
, INTERNAL
, UNAVAILABLE
, DATA_LOSS
A developer-facing error message in English.
The kind of model. If not specified, the default is HF_PEFT_ADDON.
KIND_UNSPECIFIED
, HF_BASE_MODEL
, HF_PEFT_ADDON
, HF_TEFT_ADDON
, FLUMINA_BASE_MODEL
, FLUMINA_ADDON
, DRAFT_ADDON
, FIRE_AGENT
, LIVE_MERGE
The URL to GitHub repository of the model.
The URL to the Hugging Face model.
Base model details. Required if kind is HF_BASE_MODEL. Must not be set otherwise.
The default number of GPUs the model is served with. If not specified, the default is 1.
CHECKPOINT_FORMAT_UNSPECIFIED
, NATIVE
, HUGGINGFACE
The number of model parameters. For serverless models, this determines the price per token.
If true, this is a Mixture of Experts (MoE) model. For serverless models, this affects the price per token.
If true, this model is available for fine-tuning.
The type of the model.
PEFT addon details. Required if kind is HF_PEFT_ADDON or HF_TEFT_ADDON.
The rank of the update matrices. Must be between 4 and 64, inclusive.
The type of the model.
TEFT addon details. Required if kind is HF_TEFT_ADDON. Must not be set otherwise.
If true, the model will be publicly readable.
If set, the Chat Completions API will be enabled for this model.
The maximum context length supported by the model.
If set, images can be provided as input to the model.
If set, tools (i.e. functions) can be provided as input to the model, and the model may respond with one or more tool calls.
The name of the the model from which this was imported. This field is empty if the model was not imported.
If the model was created from a fine-tuning job, this is the fine-tuning job name.
The default draft model to use when creating a deployment. If empty, speculative decoding is disabled by default.
The default draft token count to use when creating a deployment. Must be specified if default_draft_model is specified.
PRECISION_UNSPECIFIED
, FP16
, FP8
, FP8_MM
, FP8_AR
, FP8_MM_KV_ATTN
, FP8_KV
, FP8_MM_V2
, FP8_V2
, FP8_MM_KV_ATTN_V2
, NF4
Populated from GetModel API call only.
The resource name of the base deployment the model is deployed to.
The state of the deployed model.
STATE_UNSPECIFIED
, UNDEPLOYING
, DEPLOYING
, DEPLOYED
, UPDATING
If true, this is the default target when querying this model without
the #<deployment>
suffix.
The first deployment a model is deployed to will have this field set to
true automatically.
If true, the deployed model will be publicly reachable.
The resource name of the BYOC cluster to which this model belongs. e.g. accounts/my-account/clusters/my-cluster. Empty if it belongs to a Fireworks cluster.
If specified, this is the date when the serverless deployment of the model will be taken down.
Year of the date. Must be from 1 to 9999, or 0 to specify a date without a year.
Month of a year. Must be from 1 to 12, or 0 to specify a year without a month and day.
Day of a month. Must be from 1 to 31 and valid for the year and month, or 0 to specify a year by itself or a year and month where the day isn't significant.
If true, the model is calibrated and can be deployed to non-FP16 precisions.
If true, the model can be fine-tuned. The value will be true if the tunable field is true, and the model is validated against the model_type field.
Whether this model supports LoRA.
If true, the model will use the Hugging Face apply_chat_template API to apply the chat template.
A token, which can be sent as page_token
to retrieve the next page.
If this field is omitted, there are no subsequent pages.
Was this page helpful?
curl --request GET \
--url https://api.fireworks.ai/v1/accounts/{account_id}/models \
--header 'Authorization: Bearer <token>'
{
"models": [
{
"name": "<string>",
"displayName": "<string>",
"description": "<string>",
"createTime": "2023-11-07T05:31:56Z",
"createdBy": "<string>",
"state": "STATE_UNSPECIFIED",
"status": {
"code": "OK",
"message": "<string>"
},
"kind": "KIND_UNSPECIFIED",
"githubUrl": "<string>",
"huggingFaceUrl": "<string>",
"baseModelDetails": {
"worldSize": 123,
"checkpointFormat": "CHECKPOINT_FORMAT_UNSPECIFIED",
"parameterCount": "<string>",
"moe": true,
"tunable": true,
"modelType": "<string>"
},
"peftDetails": {
"baseModel": "<string>",
"r": 123,
"targetModules": [
"<string>"
],
"baseModelType": "<string>",
"mergeAddonModelName": "<string>"
},
"teftDetails": {},
"public": true,
"conversationConfig": {
"style": "<string>",
"system": "<string>",
"template": "<string>"
},
"contextLength": 123,
"supportsImageInput": true,
"supportsTools": true,
"importedFrom": "<string>",
"fineTuningJob": "<string>",
"defaultDraftModel": "<string>",
"defaultDraftTokenCount": 123,
"precisions": [
"PRECISION_UNSPECIFIED"
],
"deployedModelRefs": [
{
"name": "<string>",
"deployment": "<string>",
"state": "STATE_UNSPECIFIED",
"default": true,
"public": true
}
],
"cluster": "<string>",
"deprecationDate": {
"year": 123,
"month": 123,
"day": 123
},
"calibrated": true,
"tunable": true,
"supportsLora": true,
"useHfApplyChatTemplate": true
}
],
"nextPageToken": "<string>",
"totalSize": 123
}