SHI-Labs / Neighborhood-Attention-Transformer

Neighborhood Attention Transformer, arxiv 2022 / CVPR 2023. Dilated Neighborhood Attention Transformer, arxiv 2022

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

No position encoding? Could you explain some your thoughts?

laisimiao opened this issue · comments

No position encoding? Could you explain some your thoughts?

Thank you for your interest.
It's common practice in hierarchical vision transformers that use local attention with relative positional biases not to use absolute positional encoding (i.e. Swin), and we simply followed that idea.

Closing this due to inactivity. If you still have questions feel free to open it back up.