Large AI models trained on broad data that can be adapted to many downstream tasks. GPT-4, Claude, and BERT are examples that serve as the foundation for specific applications.
Foundation models are large-scale AI models trained on diverse data that can be adapted to numerous downstream tasks. The term, coined by Stanford researchers, captures how these models serve as the base for many applications.
Characteristics of foundation models:
Types of foundation models:
Using foundation models:
Foundation models provide powerful capabilities out-of-the-box. Businesses build on them via APIs rather than training from scratch, dramatically reducing AI adoption costs.
We help Australian businesses leverage foundation models effectively - choosing the right model, implementing via APIs, and customising through fine-tuning or RAG when needed.
"Using Claude as the foundation for a customer service chatbot, customised with company knowledge via RAG without training from scratch."