/awesome-llm

Awesome series for Large Language Model(LLM)s

Creative Commons Zero v1.0 UniversalCC0-1.0

Awesome LLM

Awesome

Awesome series for Large Language Model(LLM)s

Contents

Models

Overview

Name Parameter size Announcement date
BERT-Large (336M) 336 million 2018
T5 (11B) 11 billion 2020
Gopher (280B) 280 billion 2021
GPT-J (6B) 6 billion 2021
LaMDA (137B) 137 billion 2021
Megatron-Turing NLG (530B) 530 billion 2021
T0 (11B) 11 billion 2021
Macaw (11B) 11 billion 2021
GLaM (1.2T) 1.2 trillion 2021
T5 FLAN (540B) 540 billion 2022
OPT-175B (175B) 175 billion 2022
ChatGPT (175B) 175 billion 2022
GPT 3.5 (175B) 175 billion 2022
AlexaTM (20B) 20 billion 2022
Bloom (176B) 176 billion 2022
Bard Not yet announced 2023
GPT 4 Not yet announced 2023
AlphaCode (41.4B) 41.4 billion 2022
Chinchilla (70B) 70 billion 2022
Sparrow (70B) 70 billion 2022
PaLM (540B) 540 billion 2022
NLLB (54.5B) 54.5 billion 2022
Alexa TM (20B) 20 billion 2022
Galactica (120B) 120 billion 2022
UL2 (20B) 20 billion 2022
Jurassic-1 (178B) 178 billion 2022
LLaMA (65B) 65 billion 2023
Stanford Alpaca (7B) 7 billion 2023
GPT-NeoX 2.0 (20B) 20 billion 2023
BloombergGPT 50 billion 2023
Dolly 6 billion 2023
Jurassic-2 Not yet announced 2023
OpenAssistant LLaMa 30 billion 2023
Koala 13 billion 2023
Vicuna 13 billion 2023
PaLM2 Not yet announced, Smaller than PaLM1 2023
LIMA 65 billion 2023
MPT 7 billion 2023
Falcon 40 billion 2023
Llama 2 70 billion 2023
Google Gemini Not yet announced 2023
Microsoft Phi-2 2.7 billion 2023
Grok-0 33 billion 2023
Grok-1 314 billion 2023
Solar 10.7 billion 2024
Gemma 7 billion 2024
Grok-1.5 Not yet announced 2024
DBRX 132 billion 2024
Claude 3 Not yet announced 2024
Gemma 1.1 7 billion 2024
Llama 3 70 billion 2024

⬆️ Go to top

Open models

⬆️ Go to top

Projects

  • Visual ChatGPT - Announced by Microsoft / 2023
  • LMOps - Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities.

⬆️ Go to top

Commercial models

GPT

⬆️ Go to top

Gemini

  • Gemini - Announced by Google Deepmind / 2023

Bard

  • Bard - Announced by Google / 2023

⬆️ Go to top

Codex

⬆️ Go to top

Datasets

  • Sphere - Announced by Meta / 2022
    • 134M documents split into 906M passages as the web corpus.
  • Common Crawl
    • 3.15B pages and over than 380TiB size dataset, public, free to use.
  • SQuAD 2.0
    • 100,000+ question dataset for QA.
  • Pile
    • 825 GiB diverse, open source language modelling data set.
  • RACE
    • A large-scale reading comprehension dataset with more than 28,000 passages and nearly 100,000 questions.
  • Wikipedia
    • Wikipedia dataset containing cleaned articles of all languages.

⬆️ Go to top

Benchmarks

⬆️ Go to top

Materials

Papers

Posts

⬆️ Go to top

Projects

GitHub repositories

  • Stanford Alpaca - Repo stars of tatsu-lab/stanford_alpaca - A repository of Stanford Alpaca project, a model fine-tuned from the LLaMA 7B model on 52K instruction-following demonstrations.
  • Dolly - Repo stars of databrickslabs/dolly - A large language model trained on the Databricks Machine Learning Platform.
  • AutoGPT - Repo stars of Significant-Gravitas/Auto-GPT - An experimental open-source attempt to make GPT-4 fully autonomous.
  • dalai - Repo stars of cocktailpeanut/dalai - The cli tool to run LLaMA on the local machine.
  • LLaMA-Adapter - Repo stars of ZrrSkywalker/LLaMA-Adapter - Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters.
  • alpaca-lora - Repo stars of tloen/alpaca-lora - Instruct-tune LLaMA on consumer hardware.
  • llama_index - Repo stars of jerryjliu/llama_index - A project that provides a central interface to connect your LLM's with external data.
  • openai/evals - Repo stars of openai/evals - A curated list of reinforcement learning with human feedback resources.
  • trlx - Repo stars of promptslab/Promptify - A repo for distributed training of language models with Reinforcement Learning via Human Feedback. (RLHF)
  • pythia - Repo stars of EleutherAI/pythia - A suite of 16 LLMs all trained on public data seen in the exact same order and ranging in size from 70M to 12B parameters.
  • Embedchain - Repo stars of embedchain/embedchain - Framework to create ChatGPT like bots over your dataset.

⬆️ Go to top

HuggingFace repositories

  • OpenAssistant SFT 6 - 30 billion LLaMa-based model made by HuggingFace for the chatting conversation.
  • Vicuna Delta v0 - An open-source chatbot trained by fine-tuning LLaMA on user-shared conversations collected from ShareGPT.
  • MPT 7B - A decoder-style transformer pre-trained from scratch on 1T tokens of English text and code. This model was trained by MosaicML.
  • Falcon 7B - A 7B parameters causal decoder-only model built by TII and trained on 1,500B tokens of RefinedWeb enhanced with curated corpora.

⬆️ Go to top

Reading materials

⬆️ Go to top

Contributing

We welcome contributions to the Awesome LLMOps list! If you'd like to suggest an addition or make a correction, please follow these guidelines:

  1. Fork the repository and create a new branch for your contribution.
  2. Make your changes to the README.md file.
  3. Ensure that your contribution is relevant to the topic of LLM.
  4. Use the following format to add your contribution:
[Name of Resource](Link to Resource) - Description of resource
  1. Add your contribution in alphabetical order within its category.
  2. Make sure that your contribution is not already listed.
  3. Provide a brief description of the resource and explain why it is relevant to LLM.
  4. Create a pull request with a clear title and description of your changes.

We appreciate your contributions and thank you for helping to make the Awesome LLM list even more awesome!

⬆️ Go to top