The process of teaching an AI model by exposing it to data and adjusting its parameters to minimise errors.
Training is the process of teaching an AI model by showing it examples and adjusting its parameters to improve performance. For language models, this involves learning to predict text from massive datasets.
Training stages for LLMs:
What happens during training:
Training requirements for LLMs:
This is why most businesses use pre-trained models rather than training from scratch. Fine-tuning offers a middle ground - smaller data and compute requirements to specialise an existing model.
Training large models from scratch costs millions. Most businesses use pre-trained models, optionally with fine-tuning for specialisation.
We help Australian businesses leverage pre-trained models effectively. Full training is rarely needed - fine-tuning or RAG typically achieves business goals at a fraction of the cost.
"GPT-4 was trained on trillions of tokens of internet text, books, and code, requiring months of compute on thousands of GPUs."