GPT-3
Generative Pre-trained Transformer 3 is an autoregressive language model that uses deep learning to produce human-like text.
175B parameters (100x larger than GPT-2)
Interesting new phenomenon (appearing only when network is large enough)
Resources
To fact check: The embedding dimensions are typically larger than 2: GPT uses 12288 dimensions.