saahithjanapati's Stars
labmlai/annotated_deep_learning_paper_implementations
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
black-forest-labs/flux
Official inference repo for FLUX.1 models
tensorflow/tensor2tensor
Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.
naklecha/llama3-from-scratch
llama3 implementation one matrix multiplication at a time
princeton-nlp/SWE-agent
[NeurIPS 2024] SWE-agent takes a GitHub issue and tries to automatically fix it, using GPT-4, or your LM of choice. It can also be employed for offensive cybersecurity or competitive coding challenges.
gpu-mode/lectures
Material for gpu-mode lectures
mpoon/gpt-repository-loader
Convert code repos into an LLM prompt-friendly format. Mostly built by GPT-4.
johnma2006/mamba-minimal
Simple, minimal implementation of the Mamba SSM in one file of PyTorch.
lucidrains/perceiver-pytorch
Implementation of Perceiver, General Perception with Iterative Attention, in Pytorch
rohitgandikota/sliders
Concept Sliders for Precise Control of Diffusion Models
pmichel31415/are-16-heads-really-better-than-1
Code for the paper "Are Sixteen Heads Really Better than One?"
rafaljozefowicz/lm
lancopku/label-words-are-anchors
Repository for Label Words are Anchors: An Information Flow Perspective for Understanding In-Context Learning
openai/neural-gpu
Code for the Neural GPU model originally described in "Neural GPUs Learn Algorithms"
saprmarks/feature-circuits
jakegrigsby/deep_control
Deep Reinforcement Learning for Continuous Control in PyTorch
ansuini/IntrinsicDimDeep
Code for the intrinsic dimensionality estimate of data representations
jbloomAus/DecisionTransformerInterpretability
Interpreting how transformers simulate agents performing RL tasks
jxmorris12/gptzip
Losslessly encode text natively with arithmetic coding and HuggingFace Transformers
bacnguyencong/rbm-pytorch
An implementation of Restricted Boltzmann Machine in Pytorch
evandez/relations
How do transformer LMs encode relations?
csinva/interpretable-embeddings
Interpretable text embeddings by asking LLMs yes/no questions (NeurIPS 2024)
dannyallover/overthinking_the_truth
for-ai/llm-profiling-toolkit
ArthurConmy/MishformerLens
MishformerLens intends to be a drop-in replacement for TransformerLens that AST patches HuggingFace Transformers rather than implementing a custom, numerically inaccurate Transformer architecture.
rohitgandikota/erasing-llm
Erasing conceptual knowledge from language models through low-rank fine-tuning
tomfletcher/GeometryOfData
jmmanley/two-nn-dimensionality-estimator
Implementation of TWO-NN method for estimating intrinsic dimension (Facco et al., 2017, Scientific Reports).
diegodoimo/geometry_icl_finetuning
Fradenti/GRIDE_repo