LoserSun's Stars
implus/mae_segmentation
reproduction of semantic segmentation using masked autoencoder (mae)
MCG-NJU/VideoMAE
[NeurIPS 2022 Spotlight] VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training
amazon-science/bigdetection
BigDetection: A Large-scale Benchmark for Improved Object Detector Pre-training
NVlabs/GroupViT
Official PyTorch implementation of GroupViT: Semantic Segmentation Emerges from Text Supervision, CVPR 2022.
rapidsai/cucim
cuCIM - RAPIDS GPU-accelerated image processing library
mzr1996/backbone-example
An example to implement a new backbone with OpenMMLab framework.
facebookresearch/mae
PyTorch implementation of MAE https//arxiv.org/abs/2111.06377
ZwwWayne/K-Net
[NeurIPS2021] Code Release of K-Net: Towards Unified Image Segmentation
282857341/nnFormer
junyuchen245/Transformer_for_medical_image_analysis
A collection of papers about Transformer in the field of medical image analysis.
DingXiaoH/RepLKNet-pytorch
Scaling Up Your Kernels to 31x31: Revisiting Large Kernel Design in CNNs (CVPR 2022)
zdou0830/METER
METER: A Multimodal End-to-end TransformER Framework
junyuchen245/TransMorph_Transformer_for_Medical_Image_Registration
TransMorph: Transformer for Unsupervised Medical Image Registration (PyTorch)
MegEngine/RepLKNet
Official MegEngine implementation of RepLKNet
Pang-Yatian/Point-MAE
[ECCV2022] Masked Autoencoders for Point Cloud Self-supervised Learning
whai362/PVTv2-Seg
showlab/all-in-one
[CVPR2023] All in One: Exploring Unified Video-Language Pre-training
IDEA-Research/awesome-detection-transformer
Collect some papers about transformer for detection and segmentation. Awesome Detection Transformer for Computer Vision (CV)
IDEA-Research/DINO
[ICLR 2023] Official implementation of the paper "DINO: DETR with Improved DeNoising Anchor Boxes for End-to-End Object Detection"
salesforce/BLIP
PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
lucidrains/vit-pytorch
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
graykode/nlp-tutorial
Natural Language Processing Tutorial for Deep Learning Researchers
open-mmlab/awesome-vit
microsoft/unilm
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
facebookresearch/Detic
Code release for "Detecting Twenty-thousand Classes using Image-level Supervision".
frgfm/torch-cam
Class activation maps for your PyTorch models (CAM, Grad-CAM, Grad-CAM++, Smooth Grad-CAM++, Score-CAM, SS-CAM, IS-CAM, XGrad-CAM, Layer-CAM)
mli/paper-reading
深度学习经典、新论文逐段精读
open-mmlab/mmdeploy
OpenMMLab Model Deployment Framework
IDEA-CCNL/Fengshenbang-LM
Fengshenbang-LM(封神榜大模型)是IDEA研究院认知计算与自然语言研究中心主导的大模型开源体系,成为中文AIGC和认知智能的基础设施。
facebookresearch/Mask2Former
Code release for "Masked-attention Mask Transformer for Universal Image Segmentation"