There are 8 repositories under masked-image-modeling topic.
OpenMMLab Pre-training Toolbox and Benchmark
OpenMMLab Self-Supervised Learning Toolbox and Benchmark
CAIRI Supervised, Semi- and Self-Supervised Visual Representation Learning Toolbox and Benchmark
[Survey] Masked Modeling for Self-supervised Representation Learning on Vision and Beyond (https://arxiv.org/abs/2401.00897)
Official PyTorch implementation of MOOD series: (1) MOODv1: Rethinking Out-of-distributionDetection: Masked Image Modeling Is All You Need. (2) MOODv2: Masked Image Modeling for Out-of-Distribution Detection.
PyTorch code for MUST
[CVPR'23] Hard Patches Mining for Masked Image Modeling
A TensorFlow 2.x implementation of Masked Autoencoders Are Scalable Vision Learners
[ICLR2024] Exploring Target Representations for Masked Autoencoders
Official Code of the paper "Cross-Scale MAE: A Tale of Multi-Scale Exploitation in Remote Sensing"
Recent Advances in Vision-Language Pre-training!
[ECCV 2022] Official pytorch implementation of "mc-BEiT: Multi-choice Discretization for Image BERT Pre-training" in European Conference on Computer Vision (ECCV) 2022.
[ICML 2023] Architecture-Agnostic Masked Image Modeling -- From ViT back to CNN
[NeurIPS 2023] Masked Image Residual Learning for Scaling Deeper Vision Transformers
Official codebase for "Unveiling the Power of Audio-Visual Early Fusion Transformers with Dense Interactions through Masked Modeling".
Code to reproduce experiments from the paper "Continual Pre-Training Mitigates Forgetting in Language and Vision" https://arxiv.org/abs/2205.09357
MambaMIM: Pre-training Mamba with State Space Token-interpolation
PyTorch implementation for "Training and Inference on Any-Order Autoregressive Models the Right Way", NeurIPS 2022 Oral, TPM 2023 Best Paper Honorable Mention
Official implementation of Matrix Variational Masked Autoencoder (M-MAE) for ICML paper "Information Flow in Self-Supervised Learning" (https://arxiv.org/abs/2309.17281)
Pytorch implementation of an energy transformer - an energy-based reccurrent variant of the transformer.
[MICCAI 2024] HySparK: Hybrid Sparse Masking for Large Scale Medical Image Pre-Training
Pytorch reimplementation of "A Unified View of Masked Image Modeling".
code for "AdPE: Adversarial Positional Embeddings for Pretraining Vision Transformers via MAE+"
Self-Supervised Representation Learning of Semiconductor Wafer Maps using PyTorch