moono / moco-tf2.x

MoCo: Momentum Contrast for Unsupervised Visual Representation Learning

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

moco-tf2.x

Note

  • Difference between official implementation
    • 8 GPUs vs 4 GPUs
    • 53 Hours vs 147 hours (Unsupervised training time) - much slower than official one
  • Batch normalization - tf
    • If one sets batch normalization layer as un-trainable, tf will normalize input with their moving mean & var, even though you use training=True
  • Lack of information about how to properly apply weight regularization within distributed environment

Result (Result lower than official one)

  • MoCo v1
    • Could not reproduce same accuracy (Linear classification protocol on Imagenet) result as official one.
InfoNCE (K+1) Accuracy
MoCo V1 InfoNCE Accuracy
Train Accuracy Validation Accuracy
lincls Accuracy Accuracy
  • Comparison with official result
ResNet-50 pre-train
epochs
pre-train
time
MoCo v1
top-1 acc.
Official
Result
200 53 hours 60.6
This repo
Result
200 147 hours 50.8

About

MoCo: Momentum Contrast for Unsupervised Visual Representation Learning


Languages

Language:Python 100.0%