Pinned Repositories
M4U
Code for the Paper M4U: Evaluating Multilingual Understanding and Reasoning for Large Multimodal Models.
awesome-detection-transformer
Collect some papers about transformer for detection and segmentation. Awesome Detection Transformer for Computer Vision (CV)
deep-learning-for-image-processing
deep learning for image processing including classification and object-detection etc.
detr
End-to-End Object Detection with Transformers
detrex
detrex is a research platform for Transformer-based Instance Recognition algorithms including DETR (ECCV 2020), Deformable-DETR (ICLR 2021), Conditional-DETR (ICCV 2021), DAB-DETR (ICLR 2022), DN-DETR (CVPR 2022), DINO (ICLR 2023), H-DETR (CVPR 2023), MaskDINO (CVPR 2023), etc.
fairseq
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
flash-attention
Fast and memory-efficient exact attention
gpt-fast
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
torchscale
Transformers at any scale
unilm
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
ustcwhy's Repositories
ustcwhy/unilm
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
ustcwhy/awesome-detection-transformer
Collect some papers about transformer for detection and segmentation. Awesome Detection Transformer for Computer Vision (CV)
ustcwhy/deep-learning-for-image-processing
deep learning for image processing including classification and object-detection etc.
ustcwhy/detr
End-to-End Object Detection with Transformers
ustcwhy/detrex
detrex is a research platform for Transformer-based Instance Recognition algorithms including DETR (ECCV 2020), Deformable-DETR (ICLR 2021), Conditional-DETR (ICCV 2021), DAB-DETR (ICLR 2022), DN-DETR (CVPR 2022), DINO (ICLR 2023), H-DETR (CVPR 2023), MaskDINO (CVPR 2023), etc.
ustcwhy/fairseq
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
ustcwhy/flash-attention
Fast and memory-efficient exact attention
ustcwhy/gpt-fast
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
ustcwhy/LAVIS
LAVIS - A One-stop Library for Language-Vision Intelligence
ustcwhy/mae
PyTorch implementation of MAE https//arxiv.org/abs/2111.06377
ustcwhy/MT-Reading-List
A machine translation reading list maintained by Tsinghua Natural Language Processing Group
ustcwhy/torchscale
Transformers at any scale
ustcwhy/JARVIS
JARVIS, a system to connect LLMs with ML community
ustcwhy/LMOps
General technology for enabling AI capabilities w/ LLMs and MLLMs
ustcwhy/Metaworld
Collections of robotics environments geared towards benchmarking multi-task and meta reinforcement learning
ustcwhy/OpenChatKit
ustcwhy/openvla
OpenVLA: An open-source vision-language-action model for robotic manipulation.
ustcwhy/TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
ustcwhy/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
ustcwhy/ustcwhy.github.io
Github Pages template for academic personal websites, forked from mmistakes/minimal-mistakes
ustcwhy/VideoMAE
[NeurIPS 2022 Spotlight] VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training
ustcwhy/WorkingTime