Joshua-Ren's Stars
lm-sys/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
hiyouga/LLaMA-Factory
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
meta-llama/llama-recipes
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama for WhatsApp & Messenger.
voxel51/fiftyone
Refine high-quality datasets and visual AI models
thuml/Time-Series-Library
A Library for Advanced Deep Time Series Models.
WooooDyy/LLM-Agent-Paper-List
The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.
princeton-nlp/tree-of-thought-llm
[NeurIPS 2023] Tree of Thoughts: Deliberate Problem Solving with Large Language Models
huggingface/alignment-handbook
Robust recipes to align language models with human and AI preferences
EleutherAI/pythia
The hub for EleutherAI's work on interpretability and learning dynamics
eric-mitchell/direct-preference-optimization
Reference implementation for DPO (Direct Preference Optimization)
cure-lab/LTSF-Linear
[AAAI-23 Oral] Official implementation of the paper "Are Transformers Effective for Time Series Forecasting?"
thuml/Autoformer
About Code release for "Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting" (NeurIPS 2021), https://arxiv.org/abs/2106.13008
yuqinie98/PatchTST
An offical implementation of PatchTST: "A Time Series is Worth 64 Words: Long-term Forecasting with Transformers." (ICLR 2023) https://arxiv.org/abs/2211.14730
uclaml/SPIN
The official implementation of Self-Play Fine-Tuning (SPIN)
ContextualAI/HALOs
A library with extensible implementations of DPO, KTO, PPO, ORPO, and other human-aware loss functions (HALOs).
MAZiqing/FEDformer
mims-harvard/TFC-pretraining
Self-supervised contrastive learning for time series via time-frequency consistency
princeton-nlp/LESS
[ICML 2024] LESS: Selecting Influential Data for Targeted Instruction Tuning
djsutherland/arxiv-collector
A little Python script to collect LaTeX sources for upload to the arXiv.
LLaMafia/llamafia.github
scienceetonnante/grokking
KindXiaoming/Omnigrok
Omnigrok: Grokking Beyond Algorithmic Data
shuyhere/about-super-alignment
Feeling confused about super alignment? Here is a reading list
linlu-qiu/lm-inductive-reasoning
tommccoy1/embers-of-autoregression
carlguo866/circle-survival
Code repository for "Survival of the Fittest Representation: A Case Study with Modular Addition."
Shawn-Guo-CN/SFT_function_learning
Reference implementation for DPO (Direct Preference Optimization)
ServiceNow/THANOS
This ANOmaly is Synthetic - A Timeseries Recipe Data Generator
Joshua-Ren/Learning_dynamics_LLM
Joshua-Ren/iICL