Pinned Repositories
Compact-Transformers
Escaping the Big Data Paradigm with Compact Transformers, 2021 (Train your Vision Transformers in 30 mins on CIFAR-10 with a single GPU!)
Cross-Scale-Non-Local-Attention
PyTorch code for our paper "Image Super-Resolution with Cross-Scale Non-Local Attention and Exhaustive Self-Exemplars Mining" (CVPR2020).
Matting-Anything
Matting Anything Model (MAM), an efficient and versatile framework for estimating the alpha matte of any instance in an image with flexible and interactive visual or linguistic user prompt guidance.
Neighborhood-Attention-Transformer
Neighborhood Attention Transformer, arxiv 2022 / CVPR 2023. Dilated Neighborhood Attention Transformer, arxiv 2022
OneFormer
OneFormer: One Transformer to Rule Universal Image Segmentation, arxiv 2022 / CVPR 2023
Prompt-Free-Diffusion
Prompt-Free Diffusion: Taking "Text" out of Text-to-Image Diffusion Models, arxiv 2023 / CVPR 2024
Pyramid-Attention-Networks
[IJCV] Pyramid Attention Networks for Image Restoration: new SOTA results on multiple image restoration tasks: denoising, demosaicing, compression artifact reduction, super-resolution
Smooth-Diffusion
Smooth Diffusion: Crafting Smooth Latent Spaces in Diffusion Models arXiv 2023 / CVPR 2024
VCoder
VCoder: Versatile Vision Encoders for Multimodal Large Language Models, arXiv 2023 / CVPR 2024
Versatile-Diffusion
Versatile Diffusion: Text, Images and Variations All in One Diffusion Model, arXiv 2022 / ICCV 2023
SHI Labs's Repositories
SHI-Labs/OneFormer
OneFormer: One Transformer to Rule Universal Image Segmentation, arxiv 2022 / CVPR 2023
SHI-Labs/Versatile-Diffusion
Versatile Diffusion: Text, Images and Variations All in One Diffusion Model, arXiv 2022 / ICCV 2023
SHI-Labs/Neighborhood-Attention-Transformer
Neighborhood Attention Transformer, arxiv 2022 / CVPR 2023. Dilated Neighborhood Attention Transformer, arxiv 2022
SHI-Labs/Prompt-Free-Diffusion
Prompt-Free Diffusion: Taking "Text" out of Text-to-Image Diffusion Models, arxiv 2023 / CVPR 2024
SHI-Labs/Matting-Anything
Matting Anything Model (MAM), an efficient and versatile framework for estimating the alpha matte of any instance in an image with flexible and interactive visual or linguistic user prompt guidance.
SHI-Labs/Compact-Transformers
Escaping the Big Data Paradigm with Compact Transformers, 2021 (Train your Vision Transformers in 30 mins on CIFAR-10 with a single GPU!)
SHI-Labs/NATTEN
Neighborhood Attention Extension. Bringing attention to a neighborhood near you!
SHI-Labs/Smooth-Diffusion
Smooth Diffusion: Crafting Smooth Latent Spaces in Diffusion Models arXiv 2023 / CVPR 2024
SHI-Labs/Rethinking-Text-Segmentation
[CVPR 2021] Rethinking Text Segmentation: A Novel Dataset and A Text-Specific Refinement Approach
SHI-Labs/VCoder
VCoder: Versatile Vision Encoders for Multimodal Large Language Models, arXiv 2023 / CVPR 2024
SHI-Labs/Agriculture-Vision
[CVPR 2020 & 2021 & 2022 & 2023] Agriculture-Vision Dataset, Prize Challenge and Workshop: A joint effort with many great collaborators to bring Agriculture and Computer Vision/AI communities together to benefit humanity!
SHI-Labs/FcF-Inpainting
[WACV 2023] Keys to Better Image Inpainting: Structure and Texture Go Hand in Hand
SHI-Labs/Convolutional-MLPs
[Preprint] ConvMLP: Hierarchical Convolutional MLPs for Vision, 2021
SHI-Labs/VMFormer
[Preprint] VMFormer: End-to-End Video Matting with Transformer
SHI-Labs/StyleNAT
New flexible and efficient image generation framework that sets new SOTA on FFHQ-256 with FID 2.05, 2022
SHI-Labs/Forget-Me-Not
Forget-Me-Not: Learning to Forget in Text-to-Image Diffusion Models, 2023
SHI-Labs/Unsupervised-Domain-Adaptation-with-Differential-Treatment
[CVPR 2020] Differential Treatment for Stuff and Things: A Simple Unsupervised Domain Adaptation Method for Semantic Segmentation
SHI-Labs/Text2Video-Zero-sd-webui
SHI-Labs/SH-GAN
[WACV 2023] Image Completion with Heterogeneously Filtered Spectral Hints
SHI-Labs/VIM
SHI-Labs/DiSparse-Multitask-Model-Compression
[CVPR 2022] DiSparse: Disentangled Sparsification for Multitask Model Compression
SHI-Labs/OneFormer-Colab
[Colab Demo Code] OneFormer: One Transformer to Rule Universal Image Segmentation.
SHI-Labs/Boosted-Dynamic-Networks
Boosted Dynamic Neural Networks, AAAI 2023
SHI-Labs/PAIR-Diffusion
PAIR-Diffusion: Object-Level Image Editing with Structure-and-Appearance Paired Diffusion Models, 2023
SHI-Labs/LIVE-Layerwise-Image-Vectorization
[CVPR 2022 Oral] Towards Layer-wise Image Vectorization
SHI-Labs/Text2Video-Zero
a copy of "Text-to-Image Diffusion Models are Zero-Shot Video Generators", ICCV 2023
SHI-Labs/SeMask-Segmentation
[Preprint] SeMask: Semantically Masked Transformers for Semantic Segmentation.
SHI-Labs/micromotion-styleGAN
SHI-Labs/SinNeRF
"SinNeRF: Training Neural Radiance Fields on Complex Scenes from a Single Image", Dejia Xu, Yifan Jiang, Peihao Wang, Zhiwen Fan, Humphrey Shi, Zhangyang Wang
SHI-Labs/VideoINR-Continuous-Space-Time-Super-Resolution
[CVPR 2022] VideoINR: Learning Video Implicit Neural Representation for Continuous Space-Time Super-Resolution