Generative Pre-trained Transformer 3 (GPT-3)
- Ben Harris
- Oct 26, 2020
- 1 min read
Generative Pre-trained Transformer 3 is an autoregressive language model that uses deep learning to produce human-like text. It is the third-generation language prediction model in the GPT-n series created by OpenAI, a for-profit San Francisco-based artificial intelligence research laboratory. Wikipedia
Comments