richardsun-voyager's Stars
karpathy/nanoGPT
The simplest, fastest repository for training/finetuning medium-sized GPTs.
tloen/alpaca-lora
Instruct-tune LLaMA on consumer hardware
NVIDIA/DeepLearningExamples
State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with reproducible accuracy and performance on enterprise-grade infrastructure.
BlinkDL/RWKV-LM
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
Lightning-AI/litgpt
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
huggingface/trl
Train transformer language models with reinforcement learning.
roboticcam/machine-learning-notes
My continuously updated Machine Learning, Probabilistic Models and Deep Learning notes and demos (2000+ slides) 我不间断更新的机器学习,概率模型和深度学习的讲义(2000+页)和视频链接
EleutherAI/lm-evaluation-harness
A framework for few-shot evaluation of language models.
OpenNMT/OpenNMT-py
Open Source Neural Machine Translation and (Large) Language Models in PyTorch
Lightning-AI/lit-llama
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
OpenGVLab/LLaMA-Adapter
[ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters
bentrevett/pytorch-sentiment-analysis
Tutorials on getting started with PyTorch and TorchText for sentiment analysis.
KoljaB/RealtimeTTS
Converts text to speech in realtime
KinWaiCheuk/nnAudio
Audio processing by using pytorch 1D convolution network
OpenLMLab/LOMO
LOMO: LOw-Memory Optimization
karpathy/pytorch-normalizing-flows
Normalizing flows in PyTorch. Current intended use is education not production.
alibaba/Megatron-LLaMA
Best practice for training LLaMA models in Megatron-LM
lupantech/ScienceQA
Data and code for NeurIPS 2022 Paper "Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering".
timbmg/Sentence-VAE
PyTorch Re-Implementation of "Generating Sentences from a Continuous Space" by Bowman et al 2015 https://arxiv.org/abs/1511.06349
ikostrikov/pytorch-flows
PyTorch implementations of algorithms for density estimation
XuezheMax/NeuroNLP2
Deep neural models for core NLP tasks (Pytorch version)
allanj/pytorch_neural_crf
Pytorch implementation of LSTM/BERT-CRF for named entity recognition
ypeleg/llama
User-friendly LLaMA: Train or Run the model using PyTorch. Nothing else.
Kyubyong/bert_ner
Ner with Bert
XuezheMax/flowseq
Generative Flow based Sequence-to-Sequence Toolkit written in Python.
serp-ai/LLaMA-8bit-LoRA
Repository for Chat LLaMA - training a LoRA for the LLaMA (1 or 2) models on HuggingFace with 8-bit or 4-bit quantization. Research only.
ProjectD-AI/LLaMA-Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
allanj/phd-thesis
Phd thesis
nlplab-best-team/diagnostic-reasoning
Explore the ability of large language models (LLMs) to perform history taking through diagnostic reasoning, and how to enhance such ability.
lingo-mit/context-ablations
🤗 Transformers: State-of-the-art Natural Language Processing for Pytorch, TensorFlow, and JAX.