OctoberChang's Stars
yangshun/tech-interview-handbook
💯 Curated coding interview preparation materials for busy software engineers
karpathy/nanoGPT
The simplest, fastest repository for training/finetuning medium-sized GPTs.
hiyouga/LLaMA-Factory
Efficiently Fine-Tune 100+ LLMs in WebUI (ACL 2024)
facebookresearch/faiss
A library for efficient similarity search and clustering of dense vectors.
openai/chatgpt-retrieval-plugin
The ChatGPT Retrieval Plugin lets you easily find personal or work documents by asking questions in natural language.
microsoft/unilm
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
microsoft/LightGBM
A fast, distributed, high performance gradient boosting (GBT, GBDT, GBRT, GBM or MART) framework based on decision tree algorithms, used for ranking, classification and many other machine learning tasks.
unslothai/unsloth
Finetune Llama 3.1, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory
NVIDIA/TensorRT
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
NVIDIA/Megatron-LM
Ongoing research training transformer models at scale
huggingface/accelerate
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
axolotl-ai-cloud/axolotl
Go ahead and axolotl questions
FlagOpen/FlagEmbedding
Retrieval and Retrieval-augmented LLMs
Lightning-AI/lit-llama
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
algorithmica-org/algorithmica
A computer science textbook
linkedin/Liger-Kernel
Efficient Triton Kernels for LLM Training
ycjuan/libffm
A Library for Field-aware Factorization Machines
ELS-RD/kernl
Kernl lets you run PyTorch transformer models several times faster on GPU with a single line of code, and is designed to be easily hackable.
awslabs/sockeye
Sequence-to-sequence framework with a focus on Neural Machine Translation based on PyTorch
aws/deep-learning-containers
AWS Deep Learning Containers (DLCs) are a set of Docker images for training and serving models in TensorFlow, TensorFlow 2, PyTorch, and MXNet.
naver/splade
SPLADE: sparse neural search (SIGIR21, SIGIR22)
gabriben/awesome-generative-information-retrieval
amzn/pecos
PECOS - Prediction for Enormous and Correlated Spaces
facebookresearch/atlas
Code repository for supporting the paper "Atlas Few-shot Learning with Retrieval Augmented Language Models",(https//arxiv.org/abs/2208.03299)
ing-bank/sparse_dot_topn
Python package to accelerate the sparse matrix multiplication and top-n similarity selection
Optimization-AI/LibAUC
LibAUC: A Deep Learning Library for X-Risk Optimization
neulab/knn-transformers
PyTorch + HuggingFace code for RetoMaton: "Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval" (ICML 2022), including an implementation of kNN-LM and kNN-MT
luyug/Condenser
EMNLP 2021 - Pre-training architectures for dense retrieval
qqaatw/pytorch-realm-orqa
PyTorch reimplementation of REALM and ORQA
OctoberChang/GAMLP
Code of GAMLP for Open Graph Benchmark