Pinned Repositories
hearts-gym
Multi-agent Hearts card game environment
accelerate
🚀 A simple way to train and use PyTorch models with multi-GPU, TPU, mixed-precision
DALLE-pytorch
Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, in Pytorch
dotfiles
Optimized workflow
immo-search
Make real estate information more easily available
PyTorch-VeLO
VeLO optimizer in PyTorch
rldurak
Bachelor's thesis on reinforcement learning for the card game Durak
sb3-ppg
Phasic policy gradient algorithm for stable-baselines3
SMWLevelGenerator
Master's thesis on generating Super Mario World levels using deep neural networks
janEbert's Repositories
janEbert/PyTorch-VeLO
VeLO optimizer in PyTorch
janEbert/dotfiles
Optimized workflow
janEbert/accelerate
🚀 A simple way to train and use PyTorch models with multi-GPU, TPU, mixed-precision
janEbert/Adan
Adan: Adaptive Nesterov Momentum Algorithm for Faster Optimizing Deep Models
janEbert/apex
A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch
janEbert/BitBLAS
BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment.
janEbert/BitMat
An efficent implementation of the method proposed in "The Era of 1-bit LLMs"
janEbert/datasets
🤗 The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools
janEbert/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
janEbert/flash-attention
Fast and memory-efficient exact attention
janEbert/fly
janEbert/gptel
A no-frills ChatGPT client for Emacs
janEbert/janEbert.github.io
Personal website
janEbert/lightning
Deep learning framework to train, deploy, and ship AI products Lightning fast.
janEbert/litgpt
Load, pretrain, finetune, deploy 20+ LLMs on your own data. Uses state-of-the-art techniques: flash attention, FSDP, 4-bit, LoRA, and more.
janEbert/lm-evaluation-harness
A framework for few-shot evaluation of autoregressive language models.
janEbert/localpilot
janEbert/megablocks
janEbert/Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
janEbert/Megatron-LM
Ongoing research training transformer models at scale
janEbert/mne-python
MNE: Magnetoencephalography (MEG) and Electroencephalography (EEG) in Python
janEbert/mup
maximal update parametrization (µP)
janEbert/NeMo
NeMo: a toolkit for conversational AI
janEbert/pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
janEbert/Quicksetup-ai
A flexible template, as a quick setup for deep learning projects in pytorch-lightning
janEbert/safari
Convolutions for Sequence Modeling
janEbert/TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
janEbert/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
janEbert/triton
Development repository for the Triton language and compiler
janEbert/utilities
Common Python utilities and GitHub Actions in Lightning Ecosystem