Generative AI is experiencing rapid growth, and this repository serves as a comprehensive hub for updates on generative AI research, interview materials, notebooks, and more!
Explore the following resources:
- Monthly Best GenAI Papers List
- GenAI Interview Resources
- Applied LLMs Mastery 2024 (created by Aishwarya Naresh Reganti) course material
- List of all GenAI-related free courses (over 30 already listed)
- List of code repositories/notebooks for developing generative AI applications
We'll be updating this repository regularly, so keep an eye out for the latest additions!
Happy Learning!
- 5-day roadmap to learn LLM foundations out now! (Click Here)
- 60 Common GenAI Interview Questions out now! (Click Here)
- [Week 4] Applied LLMs Mastery content has been released- RAG (Retrieval-Augmented Generation)
- ICLR 2024 paper summaries click here
- List of free GenAI courses click here
*Updated at the end of every month
Date | Name | Summary | Topics |
---|---|---|---|
31 Jan 2024 | Large Language Models for Mathematical Reasoning: Progresses and Challenges | This survey delves into the landscape of Large Language Models in mathematical problem-solving, addressing various problem types, datasets, techniques, factors, and challenges. It comprehensively explores the advancements and obstacles in this burgeoning field, providing insights into the current state and future directions. This survey offers a holistic perspective on LLMs' role in mathematical reasoning, aiming to guide future research in this rapidly evolving domain. | Task Specific LLMs |
29 Jan 2024 | Corrective Retrieval Augmented Generation | To address concerns about retrieval-augmented generation models' robustness, Corrective Retrieval Augmented Generation (CRAG) is proposed. CRAG incorporates a lightweight retrieval evaluator to assess document quality and triggers different retrieval actions based on confidence levels. It extends retrieval results through large-scale web searches and employs a decompose-then-recompose algorithm to focus on key information. Experiments demonstrate CRAG's effectiveness in enhancing RAG-based approaches across various generation tasks. | RAG |
29 Jan 2024 | MoE-LLaVA: Mixture of Experts for Large Vision-Language Models | This work introduces MoE-Tuning, a training strategy for Large Vision-Language Models that addresses the computational costs of existing scaling methods by constructing sparse models with constant computational overhead. It also presents MoE-LLaVA, a MoE-based sparse LVLM architecture that activates only the top-k experts during deployment. Experimental results demonstrate MoE-LLaVA's significant performance across various visual understanding and object hallucination benchmarks, providing insights for more efficient multi-modal learning systems. | MoE Models |
29 Jan 2024 | The Power of Noise: Redefining Retrieval for RAG Systems | This study examines the impact of Information Retrieval components on Retrieval-Augmented Generation systems, complementing previous research focused on LLMs' generative aspect within RAG systems. By analyzing characteristics such as document relevance, position, and context size, the study reveals unexpected insights, like the surprising performance boost from including irrelevant documents. These findings emphasize the importance of developing specialized strategies to integrate retrieval with language generation models, guiding future research in this area. | RAG |
24 Jan 2024 | MM-LLMs: Recent Advances in MultiModal Large Language Models | This paper presents a comprehensive survey of MultiModal Large Language Models (MM-LLMs), which augment off-the-shelf LLMs to support multimodal inputs or outputs. It outlines design formulations, introduces 26 existing MM-LLMs, reviews their performance on mainstream benchmarks, and summarizes key training recipes. Promising directions for MM-LLMs are explored, alongside a real-time tracking website for the latest developments, aiming to contribute to the ongoing advancement of the MM-LLMs domain. | Multimodal LLMs |
23 Jan 2024 | Red Teaming Visual Language Models | A novel red teaming dataset, RTVLM, is introduced to assess Vision-Language Models' (VLMs) performance in generating harmful or inaccurate content. It encompasses 10 subtasks across faithfulness, privacy, safety, and fairness aspects. Analysis reveals significant performance gaps among prominent open-source VLMs, prompting exploration of red teaming alignment techniques. Application of red teaming alignment to LLaVA-v1.5 bolsters model performance, indicating the need for further development in this area. | Red-Teaming |
23 Jan 2024 | Lumiere: A Space-Time Diffusion Model for Video Generation | Lumiere is introduced as a text-to-video diffusion model aimed at synthesizing realistic and coherent motion in videos. It employs a Space-Time U-Net architecture to generate entire video durations in a single pass, enabling global temporal consistency. Through spatial and temporal down- and up-sampling, and leveraging a pre-trained text-to-image diffusion model, Lumiere achieves state-of-the-art text-to-video generation results, facilitating various content creation and video editing tasks with ease. | Diffusion Models |
22 Jan 2024 | WARM: On the Benefits of Weight Averaged Reward Models | Reinforcement Learning with Human Feedback for large language models can lead to reward hacking. To address this, Weight Averaged Reward Models (WARM) are proposed, where multiple fine-tuned reward models are averaged in weight space. WARM improves efficiency and reliability under distribution shifts and preference inconsistencies, enhancing the quality and alignment of LLM predictions. Experiments on summarization tasks demonstrate WARM's effectiveness, with RL fine-tuned models using WARM outperforming single RM counterparts. | Instruction Tuning |
18 Jan 2024 | Self-Rewarding Language Models | This paper introduces Self-Rewarding Language Models, where the language model itself provides rewards during training via LLM-as-a-Judge prompting. Through iterative training, the model not only improves its instruction-following ability but also enhances its capacity to generate high-quality rewards. Fine-tuning Llama 2 70B using this approach yields a model that surpasses existing systems on the AlpacaEval 2.0 leaderboard, showcasing potential for continual improvement in both performance axes. | Prompt Engineering |
16 Jan 2024 | Code Generation with AlphaCodium: From Prompt Engineering to Flow Engineering | AlphaCodium is proposed as a new approach to code generation by Large Language Models, emphasizing a test-based, multi-stage, code-oriented iterative flow tailored for code tasks. Tested on the CodeContests dataset, AlphaCodium consistently improves LLM performance, significantly boosting accuracy compared to direct prompts. The principles and best practices derived from this approach are deemed broadly applicable to general code generation tasks. | Code Generation |
13 Jan 2024 | Leveraging Large Language Models for NLG Evaluation: A Survey | This survey delves into leveraging Large Language Models for evaluating Natural Language Generation, providing a comprehensive taxonomy for organizing existing evaluation metrics. It critically assesses LLM-based methodologies, highlighting their strengths, limitations, and unresolved challenges such as bias and domain-specificity. The survey aims to offer insights to researchers and advocate for fairer and more advanced NLG evaluation techniques. | Evaluation |
12 Jan 2024 | How Johnny Can Persuade LLMs to Jailbreak Them: Rethinking Persuasion to Challenge AI Safety by Humanizing LLMs | This paper explores a new perspective on AI safety by considering large language models as human-like communicators and studying how to jailbreak them through persuasion. It introduces a persuasion taxonomy and applies it to generate interpretable persuasive adversarial prompts (PAP), achieving high attack success rates on LLMs like GPT-3.5 and GPT-4. The study also highlights gaps in existing defenses against such attacks and advocates for more fundamental mitigation strategies for interactive LLMs. | Red-Teaming |
11 Jan 2024 | Seven Failure Points When Engineering a Retrieval Augmented Generation System | The paper explores the integration of semantic search capabilities into applications through Retrieval Augmented Generation (RAG) systems. It identifies seven failure points in RAG system design based on case studies across various domains. Key takeaways include the feasibility of validating RAG systems during operation and the evolving nature of system robustness. The paper concludes with suggestions for potential research directions to enhance RAG system effectiveness. | RAG |
10 Jan 2024 | TrustLLM: Trustworthiness in Large Language Models | The paper examines trustworthiness in large language models like ChatGPT, proposing principles and benchmarks. It evaluates 16 LLMs, finding a correlation between trustworthiness and effectiveness, but noting concerns about proprietary models outperforming open-source ones. It emphasizes the need for transparency in both models and underlying technologies for trustworthiness analysis. | Alignment |
9 Jan 2024 | Chain-of-Table: Evolving Tables in the Reasoning Chain for Table Understanding | The Chain-of-Table framework proposes leveraging tabular data explicitly in the reasoning chain to enhance table-based reasoning tasks. It guides large language models using in-context learning to iteratively generate operations and update the table, allowing for dynamic planning based on previous results. This approach achieves state-of-the-art performance on various table understanding benchmarks, showcasing its effectiveness in enhancing LLM-based reasoning. | Prompt Engineering, RAG |
8 Jan 2024 | MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts | The paper introduces MoE-Mamba, a model combining Mixture of Experts (MoE) with Sequential State Space Models (SSMs) to enhance scaling and performance. MoE-Mamba surpasses both Mamba and Transformer-MoE, achieving Transformer-like performance with fewer training steps while maintaining the inference gains of Mamba over Transformers. | MoE Models |
4 Jan 2024 | Blending Is All You Need: Cheaper, Better Alternative to Trillion-Parameters LLM | The paper investigates whether combining smaller chat AI models can match or exceed the performance of a single large model like ChatGPT, without requiring extensive computational resources. Through empirical evidence and A/B testing on the Chai research platform, the "blending" approach demonstrates potential to rival or surpass the capabilities of larger models. | Smaller Models |
Join 1000+ students on this 10-week adventure as we delve into the application of LLMs across a variety of use cases
Link to the course website
[Feb 2024] Registrations are still open click here to register
🗓️Week 1 [Jan 15 2024]: Practical Introduction to LLMs
- Applied LLM Foundations
- Real World LLM Use Cases
- Domain and Task Adaptation Methods
🗓️Week 2 [Jan 22 2024]: Prompting and Prompt Engineering
- Basic Prompting Principles
- Types of Prompting
- Applications, Risks and Advanced Prompting
🗓️Week 3 [Jan 29 2024]: LLM Fine-tuning
- Basics of Fine-Tuning
- Types of Fine-Tuning
- Fine-Tuning Challenges
🗓️Week 4 [Feb 5 2024]: RAG (Retrieval-Augmented Generation)
- Understanding the concept of RAG in LLMs
- Key components of RAG
- Advanced RAG Methods
🗓️Week 5 [ Feb 12 2024]: Tools for building LLM Apps
- Fine-tuning Tools
- RAG Tools
- Tools for observability, prompting, serving, vector search etc.
🗓️Week 6 [Feb 19 2024]: Evaluation Techniques(Coming Soon!)
- Types of Evaluation
- Common Evaluation Benchmarks
- Common Metrics
🗓️Week 7 [Feb 26 2024]: Building Your Own LLM Application(Coming Soon!)
- Components of LLM application
- Build your own LLM App end to end
🗓️Week 8 [March 4 2024]: Advanced Features and Deployment(Coming Soon!)
- LLM lifecycle and LLMOps
- LLM Monitoring and Observability
- Deployment strategies
🗓️Week 9 [March 11 2024]: Challenges with LLMs(Coming Soon!)
- Scaling Challenges
- Behavioral Challenges
- Future directions
🗓️Week 10 [March 18 2024]: Emerging Research Trends(Coming Soon!)
- Smaller and more performant models
- Multimodal models
- LLM Alignment
🗓️Week 11 Bonus [March 25 2024]: Foundations(Coming Soon!)
- Generative Models Foundations
- Self-Attention and Transformers
- Neural Networks for Language
-
Large Language Models by ETH Zurich
-
Understanding Large Language Models by Princeton
-
Transformers course by Huggingface
-
NLP course by Huggingface
-
CS324 - Large Language Models by Stanford
-
Generative AI with Large Language Models by Coursera
-
Introduction to Generative AI by Coursera
-
Generative AI Fundamentals by Google Cloud
-
Introduction to Large Language Models by Google Cloud
-
Introduction to Generative AI by Google Cloud
-
Generative AI Concepts by DataCamp (Daniel Tedesco Data Lead @ Google)
-
1 Hour Introduction to LLM (Large Language Models) by WeCloudData
-
LLMOps: Building Real-World Applications With Large Language Models by Udacity
-
Full Stack LLM Bootcamp by FSDL
-
Generative AI for beginners by Microsoft
-
Large Language Models: Application through Production by Databricks
-
Generative AI Foundations by AWS
-
LLM University by Cohere
-
LLM Learning Lab by Lightning AI
-
Functions, Tools and Agents with LangChain by Deeplearning.AI
-
LangChain for LLM Application Development by Deeplearning.AI
-
LLMOps by DeepLearning.AI
-
Automated Testing for LLMOps by DeepLearning.AI
-
LangChain & Vector Databases in Production by Activeloop
-
Reinforcement Learning from Human Feedback by DeepLearning.AI
-
Building Applications with Vector Databases by DeepLearning.AI
-
How Diffusion Models Work by DeepLearning.AI
-
Finetuning Large Language Models by Deeplearning.AI
-
LangChain: Chat with Your Data by Deeplearning.AI
-
Building Systems with the ChatGPT API by Deeplearning.AI
-
Building Applications with Vector Databases by Deeplearning.AI
-
ChatGPT Prompt Engineering for Developers by Deeplearning.AI
-
Advanced RAG Orchestration series by LlamaIndex
-
Building and Evaluating Advanced RAG Applications by DeepLearning.AI
-
Evaluating and Debugging Generative AI Models Using Weights and Biases by Deeplearning.AI
- Common GenAI Interview Questions
- Prompting and Prompt Engineering
- Model Fine-Tuning
- Model Evaluation
- MLOps for GenAI
- Generative Models Foundations
- Latest Research Trends
- Designing an LLM-Powered Search Engine
- Building a Customer Support Chatbot
- Building a system for natural language interaction with your data.
- Building an AI Co-pilot
- Designing a Custom Chatbot for Q/A on Multimodal Data (Text, Images, Tables, CSV Files)
- Building an Automated Product Description and Image Generation System for E-commerce
- AWS Bedrock Workshop Tutorials by Amazon Web Services
- Langchain Tutorials by gkamradt
- LLM Applications for production by ray-project
- LLM tutorials by Ollama
- LLM Hub by mallahyari
- LLM Fine-tuning tutorials by ashishpatel26
- PEFT example notebooks by Huggingface
- Free LLM Fine-Tuning Notebooks by Youssef Hosni
If you want to add to the repository or find any issues, please feel free to raise a PR and ensure correct placement within the relevant section or category.
To cite this guide, use the below format:
@article{areganti_generative_ai_guide,
author = {Reganti, Aishwarya Naresh},
journal = {https://github.com/aishwaryanr/awesome-generative-ai-resources},
month = {01},
title = {{Generative AI Guide}},
year = {2024}
}
[MIT License]