/LLM_and_ChatGPT

LLM and ChatGPT work in progress!

Primary LanguageJupyter Notebook

Skills and projects covered in LLM and ChatGPT folders:

Section 1: Large Language Models: Application through Production (EdX)

Use Hugging Face datasets and large models; set up tokenizer config

Build a knowledge-based question answering / search system

  • Convert a dataset to vectors and save them in a vector library (FAISS) or database (Chroma)
  • Vectorize a query (with filters in it) and saved the output as context
  • Combine context with original prompt as new prompts to generate search results

This module also has tutorials on Pinecone and Weaviate

Build tree LLM-chain-based models

  • LLM1 moderates the comments generated by LLM2
  • Use LLM with LangChain agents (Wiki, Google, Python REPL) to do automatic data analyses
  • An LLM agent that allows user to have free chat with documents (e.g., Shakespeare's books)

Fine-tune LLMs with Hugging Face, Tensorboard, and DeepSpeed (multiple GPU cluster support) on a traditional IMDB classification; evaluate summarization performance with NLTK and ROUGE

Hugging Face Disaggregator (for q quick demographic analysis) and evaluate (for toxicity), gender expression generation, and SHAP (for interpretability, i.e., token-level contribution for the final generated output)

MLOps of a sample model with MLFlow, focusing on model registeration, versioning, monitoring/performance Tracking, and pushing to the production

Section 2: my fine-tuning practices:


Section 3: Generative AI and Large Language Models (Coursera)


Section 4: openai_api_ and prompt dev


Section 5: Foundation Models from the Ground Up (EdX)


Section 6: Apply LLMOps using VetexAI(DeepLearningAI - GCP)

course notes


Section 7: Azure AI APIs

This section includes several cases of Azure AI services, including Azure OpenAI, AI Studio, Speech (TTS & STT), AI Search, Document Intelligence, Storage, Vision, Language Service (translation, sentiment analysis, lang detection, intent & entity), and Semantic Kernel (similar to LangChain)

Iterative development is crucial for LLM apps. I use a platform called Langfuse, similar to W&B and MLFlow. The code includes setting up Llama3 on Groq, logs, prompts, and datasets for LLM monitoring and improvement.

other resources