The harmonic mean of precision and recall, providing a single metric that balances both. Useful when you need good performance on both false positives and false negatives.
The F1 score combines precision and recall into a single metric. The harmonic mean punishes extreme values, requiring both metrics to be high for a good F1.
F1 formula: F1 = 2 × (Precision × Recall) / (Precision + Recall)
F1 characteristics:
When to use F1:
Variants:
F1 is the go-to metric for imbalanced classification problems where accuracy misleads. It forces you to care about both precision and recall.
We typically report F1 scores for Australian business classification projects, especially when dealing with imbalanced outcomes.
"Comparing fraud detection models: Model A has F1 of 0.82, Model B has 0.76. Model A better balances catching fraud while minimising false alerts."