zhixuan-lin's Stars
OpenNLPLab/HGRN
[NeurIPS 2023 spotlight] Official implementation of HGRN in our NeurIPS 2023 paper - Hierarchically Gated Recurrent Neural Network for Sequence Modeling
corl-team/xland-minigrid
JAX-accelerated Meta-Reinforcement Learning Environments Inspired by XLand and MiniGrid 🏎️
NicolasZucchet/Online-learning-LR-dependencies
Implementation of the "Online learning of long-range dependencies" paper, NeurIPS 2023
BartoszJarocki/cv
Print-friendly, minimalist CV page
instadeepai/flashbax
⚡ Flashbax: Accelerated Replay Buffers in JAX
johnma2006/mamba-minimal
Simple, minimal implementation of the Mamba SSM in one file of PyTorch.
practical-tutorials/project-based-learning
Curated list of project-based tutorials
hristo-vrigazov/mmap.ninja
Memory mapped numpy arrays of varying shapes
jurgisp/pydreamer
PyTorch implementation of DreamerV2 model-based RL algorithm
state-spaces/mamba
Mamba SSM architecture
brett-daley/trajectory-aware-etraces
ICML 2023: Trajectory-Aware Eligibility Traces for Off-Policy Reinforcement Learning. https://arxiv.org/abs/2301.11321
facebookresearch/motif
Intrinsic Motivation from Artificial Intelligence Feedback
NM512/dreamerv3-torch
Implementation of Dreamer v3 in pytorch.
stas00/ml-engineering
Machine Learning Engineering Open Book
karpathy/nn-zero-to-hero
Neural Networks: Zero to Hero
UT-Austin-RPL/amago
a simple and scalable agent for training adaptive policies with sequence-based RL
01-ai/Yi
A series of large language models trained from scratch by developers @01-ai
luchris429/popjaxrl
Benchmarking RL for POMDPs in Pure JAX [Code for "Structured State Space Models for In-Context Reinforcement Learning" (NeurIPS 2023)]
state-spaces/s4
Structured state space sequence models
lindermanlab/S5
vwxyzjn/cleanba
CleanRL's implementation of DeepMind's Podracer Sebulba Architecture for Distributed DRL
eureka-research/Eureka
Official Repository for "Eureka: Human-Level Reward Design via Coding Large Language Models" (ICLR 2024)
cpacker/MemGPT
Letta (fka MemGPT) is a framework for creating stateful LLM services.
bstadie/krazyworld
krazy grid world
PWhiddy/PokemonRedExperiments
Playing Pokemon Red with Reinforcement Learning
NicolasZucchet/minimal-LRU
Non official implementation of the Linear Recurrent Unit (LRU, Orvieto et al. 2023)
BlinkDL/RWKV-LM
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
Hannibal046/RWKV-howto
possibly useful materials for learning RWKV language model.
RulinShao/LightSeq
Official repository for LightSeq: Sequence Level Parallelism for Distributed Training of Long Context Transformers
luchris429/purejaxrl
Really Fast End-to-End Jax RL Implementations