The optimisation algorithm used to train neural networks by iteratively adjusting weights to minimise the loss function.
Gradient descent is the core optimisation algorithm that trains neural networks. It iteratively adjusts model weights in the direction that reduces error, like finding the lowest point in a landscape by always walking downhill.
How gradient descent works:
Variants of gradient descent:
Key hyperparameters:
Challenges:
Understanding gradient descent helps explain why AI training requires significant compute and why learning rates matter.
We configure optimisation settings appropriately for fine-tuning projects, balancing training speed with model quality for Australian business clients.
"The algorithm adjusts model weights step by step to reduce prediction errors - like finding the bottom of a valley by always walking downhill."