Contents
OpenAI's series of decoder-only transformer language models
This article hasn't been written yet
This is a stub — a placeholder for an article that is referenced by other articles but hasn't been fully written. Contribute this article
The Generative Pre-trained Transformer (GPT) series is a family of decoder-only transformer language models developed by OpenAI. GPT-1 (2018) demonstrated that pre-training on large text corpora followed by task-specific fine-tuning could achieve strong performance on NLP benchmarks. GPT-2 (2019), GPT-3 (2020), and GPT-4 (2023) each scaled parameters and training data. ChatGPT, based on GPT-3.5, launched in November 2022 and reached 100 million users within two months.
The Generative Pre-trained Transformer (GPT) series is a family of decoder-only transformer language models developed by OpenAI. GPT-1 (2018) demonstrated that pre-training on large text corpora followed by task-specific fine-tuning could achieve strong performance on NLP benchmarks. GPT-2 (2019), GPT-3 (2020), and GPT-4 (2023) each scaled parameters and training data. ChatGPT, based on GPT-3.5, launched in November 2022 and reached 100 million users within two months.