SunTongtongtong / Transformer-paper-collection

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Vision-Transformer-paper-collection

Unsupervised Domain Adaptation:

  • Exploiting Both Domain-specific and Invariant Knowledge via a Win-win Transformer for Unsupervised Domain Adaptation,Tsinghua,DeepAI Arxiv Deit as baseline, two class token in deit were used on source/target domain respectively.
  • CDTrans: Cross-domain Transformer for Unsupervised Domain Adaptation: feature alignment based on the cross attention between source and target.

Normalisation

  • PowerNorm: Rethinking Batch Normalization in Transformers, UC berkeley calculate quadratic mean as normalisation
  • Test-Time Personalization with a Transformer for Human Pose Estimation, NIPS2021 Raise a new self-supervised task for Human pose estimation. Having two poses images of the same person, extract keypoint from one image and features from another images, regenerate the ground truth and compare the reconstruction loss. During test time, redo the self-supervised task and tune the model weight.
  • DYNAMIC TOKEN NORMALIZATION IMPROVES VISION TRANSFORMER To solve the problem, current layer normalisation in transformer cannot extract local discriminative information, the paper propose a Layer-Instance normalisation, the parameter to balance these two is the similar to IBN normalisation in resnet.

Generalization

  • Delving Deep into the Generalization of Vision Transformers under Distribution Shifts Define four kinds of shift for images and three generalization methods combined with vision transformer, including adversarial learning, information theory, and self-supervised learning. The key idea is based on entropy maximization and minimization.

About