Configuration settings that control the training process, such as learning rate, batch size, and number of epochs. Set before training begins.
Hyperparameters are configuration values that control how a machine learning model learns, as opposed to parameters which are learned during training. They must be set before training begins and significantly impact model performance.
Common hyperparameters:
For LLM fine-tuning:
Hyperparameter tuning approaches:
Hyperparameter tuning can significantly impact model performance and training costs. It's a key part of fine-tuning optimisation.
We handle hyperparameter optimisation for businesses doing custom model training, finding configurations that balance performance and training cost.
"Adjusting learning rate to balance training speed and model quality - too high causes instability, too low wastes time and resources."