The component of a transformer model that generates output sequences. GPT-style models are "decoder-only" architectures optimised for text generation.
In transformer architecture, the decoder is responsible for generating output sequences one token at a time. Modern generative AI models like GPT, Claude, and Llama are "decoder-only" architectures, meaning they consist entirely of decoder layers.
The decoder works through autoregressive generation:
Key characteristics of decoder models:
Decoder-only models contrast with encoder-only models (like BERT, for understanding) and encoder-decoder models (like T5, for translation and summarisation).
Decoder models excel at content generation, coding, and conversational AI - the most common business applications of generative AI.
We primarily work with decoder models for our Australian business clients, as they power the chatbots, content generators, and coding assistants most businesses need.
"GPT models use decoder architecture to generate text one token at a time, each token informed by all previous tokens in the sequence."