Pinned Repositories
AC-SGD
Code associated with the paper **Fine-tuning Language Models over Slow Networks using Activation Compression with Guarantees**.
CocktailSGD
EventExtraction
LorrinWWW.github.io
Jue's blog
Pyramid
Code associated with the paper **Pyramid: A Layered Model for Nested Named Entity Recognition**, at ACL 2020
SkipBERT
Code associated with the paper **SkipBERT: Efficient Inference with Shallow Layer Skipping**, at ACL 2022
Snapshot
Code associated with the paper Effective Continual Learning for Text Classification with Lightweight Snapshots, at AAAI 2023
two-are-better-than-one
Code associated with the paper **Two are Better Than One: Joint Entity and Relation Extraction with Table-Sequence Encoders**, at EMNLP 2020
weakly-supervised-slot-filling
Code associated with the paper **Effective Slot Filling via Weakly-Supervised Dual-Model Learning**, at AAAI 2021
OpenChatKit
LorrinWWW's Repositories
LorrinWWW/SkipBERT
Code associated with the paper **SkipBERT: Efficient Inference with Shallow Layer Skipping**, at ACL 2022
LorrinWWW/LorrinWWW.github.io
Jue's blog
LorrinWWW/Snapshot
Code associated with the paper Effective Continual Learning for Text Classification with Lightweight Snapshots, at AAAI 2023
LorrinWWW/DeeperSpeed
DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.
LorrinWWW/GLM-130B
GLM-130B: An Open Bilingual Pre-Trained Model
LorrinWWW/advertorch
A Toolbox for Adversarial Robustness Research
LorrinWWW/ChatGPT-Next-Web
A cross-platform ChatGPT/Gemini UI (Web / PWA / Linux / Win / MacOS). 一键拥有你自己的跨平台 ChatGPT/Gemini 应用。
LorrinWWW/common-dockers
LorrinWWW/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and FastChat-T5.
LorrinWWW/FLASK
[ICLR 2024 Spotlight] FLASK: Fine-grained Language Model Evaluation based on Alignment Skill Sets
LorrinWWW/fmengine-torch
FMEngine [PyTorch version]
LorrinWWW/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
LorrinWWW/gpt-neox-llama
LorrinWWW/helm
Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models (https://arxiv.org/abs/2211.09110).
LorrinWWW/icetk
A unified tokenization tool for Images, Chinese and English.
LorrinWWW/json-stream
Simple streaming JSON parser and encoder.
LorrinWWW/LLM-Pruner
[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support LLaMA, Llama-2, BLOOM, Vicuna, Baichuan, etc.
LorrinWWW/lm-evaluation-harness
A framework for few-shot evaluation of autoregressive language models.
LorrinWWW/lm-format-enforcer
Enforce the output format (JSON Schema, Regex etc) of a language model
LorrinWWW/min-dalle
min(DALL·E) is a fast, minimal port of DALL·E Mega to PyTorch
LorrinWWW/naive-rank-server
LorrinWWW/natural-instructions
Expanding natural instructions
LorrinWWW/Open-Instruction-Generalist
Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks
LorrinWWW/open_clip
An open source implementation of CLIP.
LorrinWWW/OpenChatKit
LorrinWWW/Quick_Deployment_HELM
LorrinWWW/RedPajama-Data
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
LorrinWWW/scaled-rope
Lora
LorrinWWW/vBERT
LorrinWWW/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs