eah13's Stars
Lightning-AI/lit-llama
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
anthropics/hh-rlhf
Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"
gradio-app/gradio
Build and share delightful machine learning apps, all in Python. 🌟 Star to support our work!
delta-io/delta
An open-source storage framework that enables building a Lakehouse architecture with compute engines including Spark, PrestoDB, Flink, Trino, and Hive and APIs
delta-io/delta-sharing
An open protocol for secure data sharing
yizhongw/self-instruct
Aligning pretrained language models with instruction data generated by themselves.
tatsu-lab/stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
Cranial-XIX/llm-pddl
glassroom/heinsen_routing
Reference implementation of "An Algorithm for Routing Vectors in Sequences" (Heinsen, 2022) and "An Algorithm for Routing Capsules in All Domains" (Heinsen, 2019), for composing deep neural networks.
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
denisidoro/navi
An interactive cheatsheet tool for the command-line
atuinsh/atuin
✨ Magical shell history
replit/ReplitLM
Inference code and configs for the ReplitLM model family
aibasel/downward
The Fast Downward domain-independent classical planning system
IST-DASLab/sparsegpt
Code for the ICML 2023 paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot".
ardatan/graphql-mesh
🕸️ GraphQL Federation Framework for any API services such as REST, OpenAPI, Swagger, SOAP, gRPC and more...
bigscience-workshop/data-preparation
Code used for sourcing and cleaning the BigScience ROOTS corpus
bigscience-workshop/data_tooling
Tools for managing datasets for governance and training.
bigscience-workshop/metadata
Experiments on including metadata such as URLs, timestamps, website descriptions and HTML tags during pretraining.
bigscience-workshop/petals
🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
Unstructured-IO/unstructured
Open source libraries and APIs to build custom preprocessing pipelines for labeling, training, or production machine learning pipelines.
getcursor/cursor
The AI Code Editor
atfortes/Awesome-LLM-Reasoning
Reasoning in LLMs: Papers and Resources, including Chain-of-Thought, OpenAI o1, and DeepSeek-R1 🍓
crazyofapple/Reading_groups
A paper & resource list of large language models, including course, paper, demo, figures
deepset-ai/haystack-core-integrations
Additional packages (components, document stores and the likes) to extend the capabilities of Haystack version 2.0 and onwards
deepset-ai/prompthub
rolandtannous/haystack-memory
Basic Memory library for Haystack NLP agents
deepset-ai/haystack
AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.
eth-sri/lmql
A language for constraint-guided and efficient LLM programming.
deepspeedai/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.