TrueNobility303's Stars
yaircarmon/recapp
RECAPP: Crafting a More Efficient Catalyst for Convex Optimization
Ledzy/BAdam
hiyouga/LLaMA-Factory
Efficiently Fine-Tune 100+ LLMs in WebUI (ACL 2024)
gsmalinovsky/ProxSkip-Public
Camera ready version
alshedivat/al-folio
A beautiful, simple, clean, and responsive Jekyll theme for academics
x-zho14/MAPLE
Code for Model Agnostic Sample Reweighting for Out-of-Distribution Learning
Raymond30/CG-BiO
danielle-hausler/ms-optimal
tatsu-lab/stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
sowmaster/esjacobians
Implementations of the algorithms described in the paper: On the Convergence Theory for Hessian-Free Bilevel Algorithms.
2003pro/ScaleBiO
This is the official implementation of ScaleBiO: Scalable Bilevel Optimization for LLM Data Reweighting
phquang/Contextual-Transformation-Network
Chavdarova/LAGAN-Lookahead_Minimax
Source code for "Taming GANs with Lookahead–Minmax", ICLR 2021.
bojone/rerope
Rectified Rotary Position Embeddings
bojone/NBCE
Naive Bayes-based Context Extension
LIONS-EPFL/BiSAM
The public code for BiSAM paper.
ZhuiyiTechnology/roformer
Rotary Transformer
yang-song/score_sde
Official code for Score-Based Generative Modeling through Stochastic Differential Equations (ICLR 2021, Oral)
yang-song/score_sde_pytorch
PyTorch implementation for Score-Based Generative Modeling through Stochastic Differential Equations (ICLR 2021, Oral)
OPTML-Group/DeepZero
[ICLR'24] "DeepZero: Scaling up Zeroth-Order Optimization for Deep Model Training" by Aochuan Chen*, Yimeng Zhang*, Jinghan Jia, James Diffenderfer, Jiancheng Liu, Konstantinos Parasyris, Yihua Zhang, Zheng Zhang, Bhavya Kailkhura, Sijia Liu
weizeming/SAM_AT
datamllab/LongLM
[ICML'24 Spotlight] LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning
marlbenchmark/on-policy
This is the official implementation of Multi-Agent PPO (MAPPO).
gbaydin/hypergradient-descent
Hypergradient descent
allenbai01/transformers-as-statisticians
kach/gradient-descent-the-ultimate-optimizer
Code for our NeurIPS 2022 paper
QingruZhang/AdaLoRA
AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning (ICLR 2023).
osehmathias/lisa
LISA: Layerwise Importance Sampling for Memory-Efficient Large Language Model Fine-Tuning
IST-DASLab/spops
Outsider565/LoRA-GA