Fine-tuning

Does Fireworks offer a fine-tuning service?

Yes, we offer both a fine-tuning service and allow external fine-tuned models to be uploaded for inference. View docs for our fine-tuning service and docs for uploading external fine-tuned models. Fireworks lets you simultaneously deploy 100 fine tuned models for fast, serverless inference.

What models are supported for fine-tuning? Is Llama 3 supported for fine-tuning?

Yes, Llama 3 (8B and 70B) is supported for fine-tuning LoRA adapters and serving them via our serverless and on-demand deployments for inference. You can see the full list of models available for fine-tuning in our docs.