DataLama's Stars
openai/openai-cookbook
Examples and guides for using the OpenAI API
stanfordnlp/dspy
DSPy: The framework for programming—not prompting—language models
microsoft/unilm
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
joonspk-research/generative_agents
Generative Agents: Interactive Simulacra of Human Behavior
BlinkDL/RWKV-LM
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, constant space (no kv-cache), fast training, infinite ctx_len, and free sentence embedding.
NVIDIA/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
FlagOpen/FlagEmbedding
Retrieval and Retrieval-augmented LLMs
EleutherAI/lm-evaluation-harness
A framework for few-shot evaluation of language models.
ijl/orjson
Fast, correct Python JSON library supporting dataclasses, datetimes, and numpy
paradedb/paradedb
Postgres for Search and Analytics
google/gemma_pytorch
The official PyTorch implementation of Google's Gemma models
microsoft/LLMLingua
[EMNLP'23, ACL'24] To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.
pytorch/torchtune
PyTorch native post-training library
openai/transformer-debugger
microsoft/LMOps
General technology for enabling AI capabilities w/ LLMs and MLLMs
pydantic/logfire
Uncomplicated Observability for Python and beyond! 🪵🔥
apple/axlearn
An Extensible Deep Learning Library
flexflow/FlexFlow
FlexFlow Serve: Low-Latency, High-Performance LLM Serving
huggingface/nanotron
Minimalistic large language model 3D-parallelism training
explosion/spacy-llm
🦙 Integrating LLMs into structured NLP pipelines
princeton-nlp/MeZO
[NeurIPS 2023] MeZO: Fine-Tuning Language Models with Just Forward Passes. https://arxiv.org/abs/2305.17333
songys/AwesomeKorean_Data
한국어 데이터 세트 링크
theodo-group/GenossGPT
One API for all LLMs either Private or Public (Anthropic, Llama V2, GPT 3.5/4, Vertex, GPT4ALL, HuggingFace ...) 🌈🐂 Replace OpenAI GPT with any LLMs in your app with one line.
triton-inference-server/tensorrtllm_backend
The Triton TensorRT-LLM Backend
kakaobrain/honeybee
Official implementation of project Honeybee (CVPR 2024)
HeegyuKim/open-korean-instructions
언어모델을 학습하기 위한 공개 한국어 instruction dataset들을 모아두었습니다.
Lilcob/-DL_PaperReadingMeeting
Deep Learning Paper Reading Meeting-Archive
UpstageAI/evalverse
The Universe of Evaluation. All about the evaluation for LLMs.
liner-engineering/llm-meetup
Liner LLM Meetup archive
openhackathons-org/End-to-End-LLM
This repository is an AI Bootcamp material that consist of a workflow for LLM