Code and slides to accompany the online series of webinars: https://data4sci.com/chatgpt by Data For Science.
Large Language Models (LLMs) are perhaps the largest step forward in natural language processing in recent years. LLMs combine almost inconceivable amounts of textual data, the latest developments in Transformer deep neural networks, and self-supervised machine learning approaches to learn billions of parameters. The end result is a class of systems that is unprecedented in its capability to generate text and interact with users in a way that feels natural.
Over the past 2 to 3 years, a large number of LLMs have been trained by different industry and academic teams and optimized for specific tasks and architectures, such as unidirectional and bidirectional transformers. In this live course, you will learn about the fundamental concepts underlying LLMs and the pros and cons of each approach, and analyze specific models in some detail. Our goal is to provide you with the conceptual framework necessary to understand the latest developments in this area and to quickly evaluate which model might be the right solution for your own specific problem. Practical examples using ChatGPT and the OpenAI API will be used to give attendees a hands-on understanding of the power and limitations of this class of systems.
- Basic Principles
- Statistical Models
- Encoder-Decoder
- Transformer Models
- ChatGPT Architecture
- BERT Architecture
- LLAMA Architecture
- Model Comparison
- Understanding Embeddings
- Question Answering
- Recommendations
- Long Texts
- CODEX Model
- DALL-E
- BloombergGPT
- BlockGPT