Image Description
Image by Teresa Berndtsson / Better Images of AI / Letter Word Text Taxonomy / Licensed by CC-BY 4.0.
Introduction
We have curated a collection of articles, opinion pieces, videos, and resources related to generative AI models. This list is periodically updated to include new and relevant resources. This article is the fourth in the series. You can find the previous versions here: v1, v2, and v3.
Tutorials, Explainers, and Courses on Generative Models
- What are Generative AI models? – Video by Kate Soule, IBM Technology.
- Introduction to Large Language Models – Video by John Ewald, Google Cloud Tech.
- What is GPT-4 and how does it differ from ChatGPT? – Alex Hern, The Guardian.
- What Is ChatGPT Doing … and Why Does It Work? – Stephen Wolfram.
- Understanding Large Language Models — A Transformative Reading List – Sebastian Raschka.
- How ChatGPT is Trained – Video by Ari Seff.
- ChatGPT – what is it? How does it work? Should we be excited? Or scared? – Deep Dhillon, The Radical AI podcast.
- Everything you need to know about ChatGPT – Joanna Dungate, Turing Institute Blog.
- Turing video lecture series on foundation models: Session 1 | Session 2 | Session 3 | Session 4.
- Bard: What is Google’s Bard and how is it different to ChatGPT? – BBC.
- Bard FAQs – Google.
- Large Language Models from scratch | Large Language Models: Part 2* – Videos from Graphics in 5 minutes*.
- What are Large Language Models (LLMs)? – Video from Google for Developers.
- Risks of Large Language Models (LLM) – Phaedra Boinodiris, video from IBM Technology.
- How ChatGPT and Other LLMs Work—and Where They Could Go Next – David Nield, Wired.
- What are Large Language Models – Machine Learning Mastery.
- How To Delete Your Data From ChatGPT – Matt Burgess, Wired.
- 5 Ways ChatGPT Can Improve, Not Replace, Your Writing – David Nield, Wired.
- AI prompt engineering: learn how not to ask a chatbot a silly question – Callum Bains, The Guardian.
- How to tell if an image is AI-generated – The Guardian.
- GPT-4 – How does it work, and how do I build apps with it? – CS50 Tech Talk.
- Developing an LLM: Building, Training, Finetuning – Tutorial video from Sebastian Raschka.
- Finetuning Open-Source LLMs – Tutorial video from Sebastian Raschka.
- Building a LLM from scratch – Tutorial video from Sebastian Raschka.
- Generative AI for Beginners – A Course – Microsoft.
- What is generative AI? – IBM.
- Using generative AI to write code: a guide for researchers – The Alan Turing Institute.
- Introduction to Generative AI – Google.
Journal, Conference, arXiv, and Other Articles
- Scientists’ Perspectives on the Potential for Generative AI in their Fields – Meredith Ringel Morris, arXiv.
- LaMDA: Language Models for Dialog Applications – Romal Thoppilan et al., arXiv.
- What Language Model to Train if You Have One Million GPU Hours? – Teven Le Scao et al., arXiv.
- Alpaca: A Strong, Replicable Instruction-Following Model – Rohan Taori et al.
- Process for Adapting Language Models to Society (PALMS) with Values-Targeted Datasets – Irene Solaiman, Christy Dennison, NeurIPS 2021.
- On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? – Emily Bender, Timnit Gebru, Angelina McMillan-Major, Shmargaret Shmitchell, FAccT 2021.
- A Survey of Large Language Models – Wayne Xin Zhao et al., arXiv.
- A Watermark for Large Language Models – John Kirchenbauer, Jonas Geiping, Yuxin Wen, Jonathan Katz, Ian Miers, Tom Goldstein, arXiv.
- Between Subjectivity and Imposition: Power Dynamics in Data Annotation for Computer Vision – Milagros Miceli, Martin Schuessler, Tianling Yang, Proceedings of the ACM on Human-Computer Interaction.
- AI classifier for indicating AI-written text – OpenAI.
- Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling – Stella Biderman et al., arXiv.
- GPT-4 Technical Report – OpenAI, arXiv.
- GPT-4 System Card – OpenAI.
- BloombergGPT: A Large Language Model for Finance – Shijie Wu et al., arXiv.
- Evading Watermark based Detection of AI-Generated Content – Zhengyuan Jiang, Jinghuai Zhang, Neil Zhenqiang Gong, arXiv.
- PaLM 2 Technical Report – Google.
- Large language models (LLM) and ChatGPT: what will the impact on nuclear medicine be? – Ian L. Alberts, Lorenzo Mercolli, Thomas Pyka, George Prenosil, Kuangyu Shi, Axel Rominger, and Ali Afshar-Oromieh, Eur J Nucl Med Mol Imaging.
- Ethics of large language models in medicine and medical research – Hanzhou Li, John T Moon, Saptarshi Purkayastha, Leo Anthony Celi, Hari Trivedi and Judy W Gichoya, The Lancet.
- Science in the age of large language models – Abeba Birhane, Atoosa Kasirzadeh, David Leslie & Sandra Wachter, Nature.
- Standardizing chemical compounds with language models – Miruna T Cretu, Alessandra Toniato, Amol Thakkar, Amin A Debabeche, Teodoro Laino and Alain C Vaucher, Machine Learning: Science and Technology.
- How to keep text private? A systematic review of deep learning methods for privacy-preserving natural language processing – Samuel Sousa & Roman Kern, Artificial Intelligence Review.
- Material transformers: deep learning language models for generative materials design – Nihang Fu, Lai Wei, Yuqi Song, Qinyang Li, Rui Xin, Sadman Sadeed Omee, Rongzhi Dong, Edirisuriya M Dilanga Siriwardane and Jianjun Hu, Machine Learning: Science and Technology.
- Large language models encode clinical knowledge – Karan Singhal et al., Nature.
- BLOOM: A 176B-Parameter Open-Access Multilingual Language Model – Teven Le Scao et al., arXiv.
- SELFormer: molecular representation learning via SELFIES language models – Atakan Yüksel, Erva Ulusoy, Atabey Ünlü and Tunca Doğan, Machine Learning: Science and Technology.
- Are Emergent Abilities of Large Language Models a Mirage? – Rylan Schaeffer, Brando Miranda, Sanmi Koyejo, NeurIPS 2023.
- Scaling Data-Constrained Language Models – Niklas Muennighoff, Alexander Rush, Boaz Barak, Teven Le Scao, Nouamane Tazi, Aleksandra Piktus, Sampo Pyysalo, Thomas Wolf, Colin Raffel, NeurIPS 2023.
- DecodingTrust: A Comprehensive Assessment of Trustworthiness in GPT Models – Boxin Wang, Weixin Chen, Hengzhi Pei, Chulin Xie, Mintong Kang, Chenhui Zhang, Chejian Xu, Zidi Xiong, Ritik Dutta, Rylan Schaeffer, Sang Truong, Simran Arora, Mantas Mazeika, Dan Hendrycks, Zinan Lin, Yu Cheng, Sanmi Koyejo, Dawn Song, Bo Li, NeurIPS 2023.
- A Watermark for Large Language Models – John Kirchenbauer, Jonas Geiping, Yuxin Wen, Jonathan Katz, Ian Miers, Tom Goldstein, ICML 2023.
- Foundation Models for Music: A Survey – Yinghao Ma et al., arXiv.
- Generative AI: A systematic review using topic modelling techniques – Priyanka Gupta, Bosheng Ding, Chong Guan, Ding Ding, Data and Information Management.
- Gemini: A Family of Highly Capable Multimodal Models – Rohan Anil et al., arXiv.
- Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context – Petko Georgiev et al., arXiv.
- The Llama 3 Herd of Models – Abhimanyu Dubey et al., arXiv.
- Monitoring AI-Modified Content at Scale: A Case Study on the Impact of ChatGPT on AI Conference Peer Reviews – Weixin Liang, Zachary Izzo, Yaohui Zhang, Haley Lepp, Hancheng Cao, Xuandong Zhao, Lingjiao Chen, Haotian Ye, Sheng Liu, Zhi Huang, Daniel A. McFarland, James Y. Zou, arXiv.
- Delving into ChatGPT usage in academic writing through excess vocabulary – Dmitry Kobak, Rita González-Márquez, Emőke-Ágnes Horvát, Jan Lause, arXiv.
Blog Posts and Interviews About Published Scientific Papers
- GPT-4 + Stable-Diffusion = ?: Enhancing Prompt Understanding of Text-to-Image Diffusion Models with Large Language Models – Long Lian, Boyi Li, Adam Yala, and Trevor Darrell, BAIR blog.
- Interview with Bo Li: A comprehensive assessment of trustworthiness in GPT models – AIhub.
- Interview with Changhoon Kim – enhancing the reliability of image generative AI – AIhub.
- Utilizing generative adversarial networks for stable structure generation in Angry Birds – Matthew Stephenson and Frederic Abraham, AIhub.
- Riemannian score-based generative modelling – Valentin De Bortoli, AIhub.
- **[Interview with Paula Feldman: generating 3d models of blood vessels](https://aihub.org/2023/12/06/interview-with-paula-feld