Slide 14
Slide 14 text
GPT
( ( ← DX ))
GPT
Alec Radford, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. 2018. “Improving Language
Understanding by Generative Pre-Training”. Available at:
https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf.
GPT-2
Alec Radford, Jeff Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. “Language Models
are Unsupervised Multitask Learners”. Available at:
https://paperswithcode.com/paper/language-models-are-unsupervised-multitask.
GPT-3
Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind
Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen
Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter,
Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark,
Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. “Language
Models are Few-Shot Learners”. Available at: https://doi.org/10.48550/arXiv.2005.14165.
GPT-4
OpenAI. 2023. “GPT-4 Technical Report”. Available at: https://doi.org/10.48550/arXiv.2303.08774.
AGI ( ) — 2025-02-06 – p.14/34