Pinned Repositories
M4U
Code for the Paper M4U: Evaluating Multilingual Understanding and Reasoning for Large Multimodal Models.
BitBLAS
BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment.
fairseq
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
flash-attention
Fast and memory-efficient exact attention
gpt-fast
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
JARVIS
JARVIS, a system to connect LLMs with ML community
LMOps
General technology for enabling AI capabilities w/ LLMs and MLLMs
MT-Reading-List
A machine translation reading list maintained by Tsinghua Natural Language Processing Group
torchscale
Transformers at any scale
unilm
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
ustcwhy's Repositories
ustcwhy/unilm
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
ustcwhy/BitBLAS
BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment.
ustcwhy/fairseq
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
ustcwhy/flash-attention
Fast and memory-efficient exact attention
ustcwhy/gpt-fast
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
ustcwhy/JARVIS
JARVIS, a system to connect LLMs with ML community
ustcwhy/LMOps
General technology for enabling AI capabilities w/ LLMs and MLLMs
ustcwhy/MT-Reading-List
A machine translation reading list maintained by Tsinghua Natural Language Processing Group
ustcwhy/torchscale
Transformers at any scale
ustcwhy/OpenRLHF
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & Mixtral)
ustcwhy/TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
ustcwhy/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
ustcwhy/ustcwhy.github.io
Github Pages template for academic personal websites, forked from mmistakes/minimal-mistakes
ustcwhy/VILA
VILA - a multi-image visual language model with training, inference and evaluation recipe, deployable from cloud to edge (Jetson Orin and laptops)
ustcwhy/WorkingTime