Pinned Repositories
adaptive-computation-time-pytorch
Alex Graves' Adaptive Computation Time in PyTorch
bert_on_stilts
Fork of huggingface/pytorch-pretrained-BERT for BERT on STILTs
extratrees_cuda
Optimizing Extremely Randomized Trees with GPUs
lrqa
minimal-gpt-neox-20b
minimal-llama
minimal-opt
saliency_investigation
Code for "Investigating and Simplifying Masking-based Saliency Methods for Model Interpretability" (https://arxiv.org/abs/2010.09750)
transformers
Code and models for BERT on STILTs
usc_dae
Repository for Unsupervised Sentence Compression using Denoising Auto-Encoders
zphang's Repositories
zphang/minimal-llama
zphang/minimal-gpt-neox-20b
zphang/bert_on_stilts
Fork of huggingface/pytorch-pretrained-BERT for BERT on STILTs
zphang/minimal-opt
zphang/transformers
Code and models for BERT on STILTs
zphang/usc_dae
Repository for Unsupervised Sentence Compression using Denoising Auto-Encoders
zphang/lrqa
zphang/hyperllama
zphang/llm_feedback
zphang/minimal-t5
zphang/my_pefty_llama
Minimal implementation of multiple PEFT methods for LLaMA fine-tuning
zphang/sndict
Structured Nested Dictionaries
zphang/hpt
zphang/llama_peft
zphang/zphang.github.io
Github Page
zphang/architecture-objective
zphang/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.
zphang/doc-chat-ui
zphang/FLAN
zphang/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
zphang/GPTQ-for-LLaMa
4 bits quantization of LLaMa using GPTQ
zphang/hf_benchmark_sample
zphang/jiant
The jiant toolkit for general-purpose text understanding models
zphang/lm_evaluation_harness
zphang/Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
zphang/mpi4py
Python bindings for MPI
zphang/peft
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
zphang/pegasus
zphang/summarization_experiments
zphang/t5x