Mike-YANG-11's Stars
zhengli97/PromptKD
[CVPR 2024] Official PyTorch Code for "PromptKD: Unsupervised Prompt Distillation for Vision-Language Models"
ShuweiShao/AF-SfMLearner
[MedIA2022 & ICRA2021] Self-Supervised Monocular Depth and Ego-Motion Estimation in Endoscopy: Appearance Flow to the Rescue
EdisonLeeeee/Awesome-Masked-Autoencoders
A collection of literature after or concurrent with Masked Autoencoder (MAE) (Kaiming He el al.).
facebookresearch/dinov2
PyTorch code and models for the DINOv2 self-supervised learning method.
HSG-AIML/GDA
Code repository for "Parameter Efficient Self-supervised Geospatial Domain Adaptation", CVPR 2024
samar-khanna/ExPLoRA
Official code repository for paper: "ExPLoRA: Parameter-Efficient Extended Pre-training to Adapt Vision Transformers under Domain Shifts"
FreedomIntelligence/HuatuoGPT-Vision
Medical Multimodal LLMs
tntek/source-free-domain-adaptation
CapsuleEndoscope/EndoSLAM
EndoSLAM Dataset and an Unsupervised Monocular Visual Odometry and Depth Estimation Approach for Endoscopic Videos: Endo-SfMLearner
NVlabs/RADIO
Official repository for "AM-RADIO: Reduce All Domains Into One"
LeapLabTHU/EfficientTrain
1.5−3.0× lossless training or pre-training speedup. An off-the-shelf, easy-to-implement algorithm for the efficient training of foundation visual backbones.
naver/unic
PyTorch code and pretrained weights for the UNIC models.
ESandML/SSL4GIE
Official code repository for: A Study on Self-Supervised Pretraining for Vision Problems in Gastrointestinal Endoscopy
labmlai/annotated_deep_learning_paper_implementations
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
tianyu0207/weakly-polyp
[MICCAI'22] Contrastive Transformer-based Multiple Instance Learning for Weakly Supervised Polyp Frame Detection.
bfshi/scaling_on_scales
When do we not need larger vision models?
Lupin1998/Awesome-MIM
[Survey] Masked Modeling for Self-supervised Representation Learning on Vision and Beyond (https://arxiv.org/abs/2401.00897)
DepthAnything/Depth-Anything-V2
[NeurIPS 2024] Depth Anything V2. A More Capable Foundation Model for Monocular Depth Estimation
cmhungsteve/Awesome-Transformer-Attention
An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites
Malitha123/awesome-video-self-supervised-learning
A curated list of awesome self-supervised learning methods in videos
facebookresearch/sscd-copy-detection
Open source implementation of "A Self-Supervised Descriptor for Image Copy Detection" (SSCD).
zhaozh10/Awesome-CLIP-in-Medical-Imaging
A Survey on CLIP in Medical Imaging
facebookresearch/maws
Code and models for the paper "The effectiveness of MAE pre-pretraining for billion-scale pretraining" https://arxiv.org/abs/2303.13496
WeixiongLin/PMC-CLIP
The official codes for "PMC-CLIP: Contrastive Language-Image Pre-training using Biomedical Documents"
PKU-YuanGroup/MoE-LLaVA
Mixture-of-Experts for Large Vision-Language Models
NVlabs/FoundationPose
[CVPR 2024 Highlight] FoundationPose: Unified 6D Pose Estimation and Tracking of Novel Objects
jhairgallardo/awesome-continual-self-supervised-learning
List of papers that combine self-supervision and continual learning
uncbiag/Awesome-Foundation-Models
A curated list of foundation models for vision and language tasks
yeerwen/MedCoSS
CVPR 2024 (Highlight)
JiazuoYu/MoE-Adapters4CL
Code for paper "Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters" CVPR2024