qwenzo
Data Engineering and Analytics Master's student at TUM. Interested in ML, DL, NLP, Data Science, and Front-End Development.
Fraunhofer IISMunich, Germany.
qwenzo's Stars
SchwinnL/LLM_Embedding_Attack
Code to conduct an embedding attack on LLMs
TransformerLensOrg/TransformerLens
A library for mechanistic interpretability of GPT-style language models
jbloomAus/SAELens
Training Sparse Autoencoders on Language Models
OpenRLHF/OpenRLHF
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)
openai/transformer-debugger
anthropics/hh-rlhf
Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"
PKU-Alignment/safe-rlhf
Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
callummcdougall/sae_vis
Create feature-centric and prompt-centric visualizations for sparse autoencoders (like those from Anthropic's published research).
Lightning-AI/litgpt
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
pytorch-labs/gpt-fast
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
IDSIA/sacred
Sacred is a tool to help you configure, organize, log and reproduce experiments developed at IDSIA.
TUM-DAML/seml
SEML: Slurm Experiment Management Library
jessevig/bertviz
BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)
tensorops/TransformerX
Flexible Python library providing building blocks (layers) for reproducible Transformers research (Tensorflow ✅, Pytorch 🔜, and Jax 🔜)
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
recommenders-team/recommenders
Best Practices on Recommendation Systems
langchain-ai/langchain
🦜🔗 Build context-aware reasoning applications
huggingface/peft
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
kermitt2/grobid
A machine learning software for extracting information from scholarly documents
archinetai/audio-diffusion-pytorch
Audio generation using diffusion models, in PyTorch.
microsoft/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
microsoft/torchscale
Foundation Architecture for (M)LLMs
thuml/Transfer-Learning-Library
Transfer Learning Library for Domain Adaptation, Task Adaptation, and Domain Generalization
erictzeng/adda
Valentin-Metz/tum_video_scraper
Download and jumpcut lecture videos from https://live.rbg.tum.de/ and https://tum.cloud.panopto.eu/
conventional-commits/conventionalcommits.org
The conventional commits specification
google/patrick
Parameterized testing in R is kind of cool!
r-lib/covr
Test coverage reports for R
tidyverse/dplyr
dplyr: A grammar of data manipulation
pop-os/pop
A project for managing all Pop!_OS sources