What model precisions are available and how do I check them?
You can check the available precision types for any model using firectl. Models may support different numerical precisions like FP16, FP8, BF16, or INT8, which affect memory usage and inference speed.
# Check supported precisions for a modelfirectl get model accounts/fireworks/models/llama-v3p1-8b-instruct | grep -E "(Supported Precisions|Supported Precisions With Calibration)"