GPT-3
| Generative Pre-trained Transformer 3 (GPT-3) | |
|---|---|
| Original author(s) | OpenAI | 
| Initial release | May 29, 2020 (publication); June 11, 2020 (OA API beta) | 
| Repository | |
| Predecessor | GPT-2 | 
| Successor | GPT-3.5 GPT-4 | 
| Type | |
| License | proprietary | 
| Website | openai | 
| Part of a series on | 
| Machine learning and data mining | 
|---|
Generative Pre-trained Transformer 3 (GPT-3) is a large language model released by OpenAI in 2020.
Like its predecessor, GPT-2, it is a decoder-only transformer model of deep neural network, which supersedes recurrence and convolution-based architectures with a technique known as "attention". This attention mechanism allows the model to focus selectively on segments of input text it predicts to be most relevant. GPT-3 has 175 billion parameters, each with 16-bit precision, requiring 350GB of storage since each parameter occupies 2 bytes. It has a context window size of 2048 tokens, and has demonstrated strong "zero-shot" and "few-shot" learning abilities on many tasks.
On September 22, 2020, Microsoft announced that it had licensed GPT-3 exclusively. Others can still receive output from its public API, but only Microsoft has access to the underlying model.