CZWin32768
PhD student at Beijing Institute of Technology. Interned at Microsoft Research and ByteDance AI Lab.
Canada
Pinned Repositories
SciTSR
Table structure recognition dataset of the paper: Complicated Table Structure Recognition
ACMICPC-Templates-in-Latex
A Simple ACM-ICPC (LaTeX) Templates Generator
gpu-job-manager
My GPU job manager
Memory-Monitor
Memory Monitor written by VB.NET
p2p-file-transfer
QiangChat-1.0
北京理工大学程序设计小学期 Linux下即时通讯文件传输聊天软件
seqmnist
unilm
UniLM AI - Unified "Language" Model Pre-training across Tasks, Languages, and Modalities
XLM-Align
XNLG
AAAI-20 paper: Cross-Lingual Natural Language Generation via Pre-Training
CZWin32768's Repositories
CZWin32768/XNLG
AAAI-20 paper: Cross-Lingual Natural Language Generation via Pre-Training
CZWin32768/XLM-Align
CZWin32768/bitsandbytes-aarch64
aarch64 for bitsandbytes
CZWin32768/CCLUE
古文语言理解测评基准 Classical Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard
CZWin32768/unilm
UniLM AI - Unified "Language" Model Pre-training across Tasks, Languages, and Modalities
CZWin32768/accelerate
🚀 A simple way to train and use PyTorch models with multi-GPU, TPU, mixed-precision
CZWin32768/adapter-transformers
Huggingface Transformers + Adapters = ❤️
CZWin32768/alpaca-lora
Instruct-tune LLaMA on consumer hardware
CZWin32768/awesome-chatgpt-prompts
This repo includes ChatGPT prompt curation to use ChatGPT better.
CZWin32768/cloudflare-scrape
A Python module to bypass Cloudflare's anti-bot page.
CZWin32768/crosslingual_winograd
"It's All in the Heads" (Findings of ACL 2021), official implementation and data
CZWin32768/DeepSpeedExamples
Example models using DeepSpeed
CZWin32768/DialoGPT
Large-scale pretraining for dialogue
CZWin32768/dockers
CZWin32768/flores
Facebook Low Resource (FLoRes) MT Benchmark
CZWin32768/improved-aesthetic-predictor
CLIP+MLP Aesthetic Score Predictor
CZWin32768/llama.cpp
Port of Facebook's LLaMA model in C/C++
CZWin32768/MAGIC
Language Models Can See: Plugging Visual Controls in Text Generation
CZWin32768/massive
Tools and Modeling Code for the MASSIVE dataset
CZWin32768/Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
CZWin32768/Neural-Collapse
[NeurIPS 2021] A Geometric Analysis of Neural Collapse with Unconstrained Features
CZWin32768/open_clip
An open source implementation of CLIP.
CZWin32768/opencompass
OpenCompass is an LLM evaluation platform, supporting a wide range of models (InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.
CZWin32768/ParlAI
A framework for training and evaluating AI models on a variety of openly available dialogue datasets.
CZWin32768/PrefixTuning
Prefix-Tuning: Optimizing Continuous Prompts for Generation
CZWin32768/RL4LMs
A modular RL library to fine-tune language models to human preferences
CZWin32768/RWKV-LM
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
CZWin32768/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
CZWin32768/trlx
A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)
CZWin32768/YaLM-100B
Pretrained language model with 100B parameters