masked-image-modeling

There are 36 repositories under masked-image-modeling topic.

  • open-mmlab/mmpretrain

    OpenMMLab Pre-training Toolbox and Benchmark

    Language:Python3.2k317561k
  • open-mmlab/mmselfsup

    OpenMMLab Self-Supervised Learning Toolbox and Benchmark

    Language:Python3.1k44273425
  • SparK

    keyu-tian/SparK

    [ICLR'23 Spotlight🔥] The first successful BERT/MAE-style pretraining on any convolutional network; Pytorch impl. of "Designing BERT for Convolutional Networks: Sparse and Hierarchical Masked Modeling"

    Language:Python1.4k268081
  • microsoft/SimMIM

    This is an official implementation for "SimMIM: A Simple Framework for Masked Image Modeling".

    Language:Python880224181
  • Westlake-AI/openmixup

    CAIRI Supervised, Semi- and Self-Supervised Visual Representation Learning Toolbox and Benchmark

    Language:Python581165359
  • Alpha-VL/ConvMAE

    ConvMAE: Masked Convolution Meets Masked Autoencoders

    Language:Python464113639
  • hustvl/MIMDet

    [ICCV 2023] You Only Look at One Partial Sequence

    Language:Python329102730
  • Lupin1998/Awesome-MIM

    [Survey] Masked Modeling for Self-supervised Representation Learning on Vision and Beyond (https://arxiv.org/abs/2401.00897)

    Language:Python2607614
  • implus/UM-MAE

    Official Codes for "Uniform Masking: Enabling MAE Pre-training for Pyramid-based Vision Transformers with Locality"

    Language:Jupyter Notebook23252220
  • lxtGH/CAE

    This is a PyTorch implementation of “Context AutoEncoder for Self-Supervised Representation Learning"

    Language:Python18351522
  • LayneH/GreenMIM

    [NeurIPS2022] Official implementation of the paper 'Green Hierarchical Vision Transformer for Masked Image Modeling'.

    Language:Python1653126
  • dvlab-research/MOOD

    Official PyTorch implementation of MOOD series: (1) MOODv1: Rethinking Out-of-distributionDetection: Masked Image Modeling Is All You Need. (2) MOODv2: Masked Image Modeling for Out-of-Distribution Detection.

    Language:Python1333114
  • Sense-X/MixMIM

    MixMIM: Mixed and Masked Image Modeling for Efficient Visual Representation Learning

    Language:Python1238236
  • salesforce/MUST

    PyTorch code for MUST

    Language:Python10361012
  • Haochen-Wang409/HPM

    [CVPR'23] Hard Patches Mining for Masked Image Modeling

    Language:Python813116
  • mae-scalable-vision-learners

    ariG23498/mae-scalable-vision-learners

    A TensorFlow 2.x implementation of Masked Autoencoders Are Scalable Vision Learners

    Language:Jupyter Notebook725915
  • Atten4Vis/CAE

    This is a PyTorch implementation of “Context AutoEncoder for Self-Supervised Representation Learning"

    Language:Python62225
  • liuxingbin/dbot

    [ICLR2024] Exploring Target Representations for Masked Autoencoders

    Language:Python50477
  • bwconrad/can

    PyTorch reimplementation of "A simple, efficient and scalable contrastive masked autoencoder for learning visual representations".

    Language:Python35216
  • aicip/Cross-Scale-MAE

    Official Code of the paper "Cross-Scale MAE: A Tale of Multi-Scale Exploitation in Remote Sensing"

    Language:Python28710
  • haofanwang/awesome-vision-language-modeling

    Recent Advances in Vision-Language Pre-training!

  • lixiaotong97/mc-BEiT

    [ECCV 2022] Official pytorch implementation of "mc-BEiT: Multi-choice Discretization for Image BERT Pre-training" in European Conference on Computer Vision (ECCV) 2022.

    Language:Python22202
  • Westlake-AI/A2MIM

    [ICML 2023] Architecture-Agnostic Masked Image Modeling -- From ViT back to CNN

    Language:Python18232
  • russellllaputa/MIRL

    [NeurIPS 2023] Masked Image Residual Learning for Scaling Deeper Vision Transformers

    Language:Python17303
  • JunlinHan/CropMix

    Code of CropMix: Sampling a Rich Input Distribution via Multi-Scale Cropping

    Language:Jupyter Notebook15112
  • AndreaCossu/continual-pretraining-nlp-vision

    Code to reproduce experiments from the paper "Continual Pre-Training Mitigates Forgetting in Language and Vision" https://arxiv.org/abs/2205.09357

    Language:Jupyter Notebook14201
  • AndyShih12/mac

    PyTorch implementation for "Training and Inference on Any-Order Autoregressive Models the Right Way", NeurIPS 2022 Oral, TPM 2023 Best Paper Honorable Mention

    Language:Python12213
  • stoneMo/DeepAVFusion

    Official codebase for "Unveiling the Power of Audio-Visual Early Fusion Transformers with Dense Interactions through Masked Modeling".

  • yifanzhang-pro/M-MAE

    Official implementation of Matrix Variational Masked Autoencoder (M-MAE) for ICML paper "Information Flow in Self-Supervised Learning" (https://arxiv.org/abs/2309.17281)

    Language:Python9203
  • LumenPallidium/energy_transformer

    Pytorch implementation of an energy transformer - an energy-based reccurrent variant of the transformer.

    Language:Python8200
  • bwconrad/masked-distillation

    Pytorch reimplementation of "A Unified View of Masked Image Modeling".

    Language:Python431
  • self-supervised-wafermaps

    faris-k/self-supervised-wafermaps

    Self-Supervised Representation Learning of Semiconductor Wafer Maps using PyTorch

    Language:Jupyter Notebook3250
  • maple-research-lab/AdPE

    code for "AdPE: Adversarial Positional Embeddings for Pretraining Vision Transformers via MAE+"

    Language:Python3300
  • chadHGY/CAM

    Learning Cortical Anomaly through Masked Encoding for Unsupervised Heterogeneity Mapping.

  • MohamedOmar2020/QuPath_scripts

    Custom groovy scripts for QuaPath

    Language:Groovy10
  • olibridge01/MaskedImageModelling

    Pre-training a VisionTransformer with Masked Image Modelling for semantic segmentation

    Language:Python