Configuration settings that control the training process, such as learning rate, batch size, and number of epochs. Set before training begins.
Hyperparameters are configuration values that control how a machine learning model learns, as opposed to parameters which are learned during training. They must be set before training begins and significantly impact model performance.
Common hyperparameters:
For LLM fine-tuning:
Hyperparameter tuning approaches:
Hyperparameter tuning can significantly impact model performance and training costs. It's a key part of fine-tuning optimisation.
We handle hyperparameter optimisation for Australian businesses doing custom model training, finding configurations that balance performance and training cost.
"Adjusting learning rate to balance training speed and model quality - too high causes instability, too low wastes time and resources."