Pinned Repositories
academic-kickstart
Easily create a beautiful website using Academic and Hugo
awesome-finetuning
A curated list of resources on fine-tuning language models.
cleverhans
An adversarial example library for constructing attacks, building defenses, and benchmarking both
interpretability-impact
Studying the impact of interpretability and analysis research in NLP
nnia-tutorial
Repository for my tutorial group which is part of the lecture Neural Networks: Implementation and Application
pre-training
Pre-Training Buys Better Robustness and Uncertainty
TF-NNLM-TK
A toolkit for neural language modeling using Tensorflow including basic models like RNNs and LSTMs as well as more advanced models.
mmarius's Repositories
mmarius/awesome-finetuning
A curated list of resources on fine-tuning language models.
mmarius/interpretability-impact
Studying the impact of interpretability and analysis research in NLP
mmarius/acr-memorization
mmarius/adapter-transformers
Huggingface Transformers + Adapters = ❤️
mmarius/BayesianTransferLearning
mmarius/composable-sft
mmarius/composer
Supercharge Your Model Training
mmarius/dynamic-lm-kb
LM with limited access to KB
mmarius/edge-of-stability
mmarius/evaluate
🤗 Evaluate: A library for easily evaluating machine learning models and datasets.
mmarius/examples
Fast and flexible reference benchmarks
mmarius/fairseq
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
mmarius/fine-tuning-connectivity
mmarius/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
mmarius/HF-Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
mmarius/llm2vec-blogpost
mmarius/Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
mmarius/mmarius.github.io
A beautiful, simple, clean, and responsive Jekyll theme for academics
mmarius/nanotron
Minimalistic large language model 3D-parallelism training
mmarius/OLMo
Modeling, training, eval, and inference code for OLMo
mmarius/OpenPrompt
An Open-Source Toolkit for Prompt-Learning.
mmarius/peft
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
mmarius/pet
This repository contains the code for "How many data points is a prompt worth?"
mmarius/promptsource
Toolkit for collecting and applying templates of prompting instances
mmarius/pythia
The hub for EleutherAI's work on interpretability and learning dynamics
mmarius/reproduce-finetuning
mmarius/shortformer
Code for the Shortformer model, from the paper by Ofir Press, Noah A. Smith and Mike Lewis.
mmarius/t0-analysis
mmarius/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
mmarius/verbatim-memorization
Demystifying Verbatim Memorization in Large Language Models