Methods
Supervised Fine Tuning - Text
Train text models with labeled examples of desired outputs
Supervised Fine Tuning - Vision
Train vision-language models with image and text pairs
Direct Preference Optimization
Align models with human preferences using pairwise comparisons
Reinforcement Fine Tuning
Train models using custom reward functions for complex reasoning tasks
Free tuning
When creating a fine-tuning job in the UI, look for the “Free tuning” filter in the model selection area:
Supported models
Fireworks supports fine-tuning for most major open source models, including DeepSeek, Qwen, Kimi, and Llama model families, and supports fine-tuning large state-of-the-art models like Kimi K2 0905 and DeepSeek V3.1. To see all models that support fine-tuning, visit the Model Library for text models or vision models.LoRA-based tuning
Managed fine-tuning uses Low-Rank Adaptation (LoRA) to fine-tune models efficiently. The fine-tuning process generates a LoRA addon — a small adapter that modifies the base model’s behavior without retraining all its weights. This approach is:- Faster and cheaper - Train models in hours, not days
- Easy to deploy - Deploy LoRA addons instantly on Fireworks
- Flexible - Run multiple LoRAs on a single base model deployment