References
Brown, Tom, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan,
Prafulla Dhariwal, Arvind Neelakantan, et al. 2020. “Language
Models Are Few-Shot Learners.” Advances in Neural Information
Processing Systems 33: 1877–1901.
Devlin, Jacob, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018.
“BERT: Pre-Training of Deep Bidirectional Transformers for
Language Understanding.” arXiv Preprint
arXiv:1810.04805.
Hoffmann, Jordan, Sebastian Borgeaud, Arthur Mensch, Elena Buchatskaya,
Trevor Cai, Eliza Rutherford, Diego de Las Casas, et al. 2022.
“Training Compute-Optimal Large Language Models.” arXiv
Preprint arXiv:2203.15556.
Kaplan, Jared, Sam McCandlish, Tom Henighan, Tom B Brown, Benjamin
Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu, and Dario
Amodei. 2020. “Scaling Laws for Neural Language Models.”
arXiv Preprint arXiv:2001.08361.
Liu, Yinhan, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi
Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov.
2019. “RoBERTa: A Robustly Optimized BERT Pretraining
Approach.” arXiv Preprint arXiv:1907.11692.
Ouyang, Long, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright,
Pamela Mishkin, Chong Zhang, et al. 2022. “Training Language
Models to Follow Instructions with Human Feedback.” Advances
in Neural Information Processing Systems 35: 27730–44.
Radford, Alec, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever.
2018. “Improving Language Understanding by Generative
Pre-Training.”
Radford, Alec, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, and
Ilya Sutskever. 2019. “Language Models Are Unsupervised Multitask
Learners.”
Rafailov, Rafael, Archit Sharma, Eric Mitchell, Stefano Ermon,
Christopher D Manning, and Chelsea Finn. 2023. “Direct Preference
Optimization: Your Language Model Is Secretly a Reward Model.”
Advances in Neural Information Processing Systems 36.
Vaswani, Ashish, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion
Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017.
“Attention Is All You Need.” Advances in Neural
Information Processing Systems 30.
Wei, Jason, Yi Tay, Rishi Bommasani, Colin Raffel, Barret Zoph,
Sebastian Borgeaud, Dani Yogatama, et al. 2022. “Emergent
Abilities of Large Language Models.” arXiv Preprint
arXiv:2206.07682.