SHI-Labs / Neighborhood-Attention-Transformer

Neighborhood Attention Transformer, arxiv 2022 / CVPR 2023. Dilated Neighborhood Attention Transformer, arxiv 2022

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

abbreviation for rpb

qsh-zh opened this issue · comments

Thanks for your awesome work.

Can you provide some clues for what is rpb and apply_pb, which does not appear in standard attention?

Thanks

Thank you for your interest.
RPB is relative positional biases, they're learnable biases that are added to attention weights.
They're widely used especially in restricted attention patterns, such as Swin's WSA, and we adopted that into NA.

You can refer to these works for more information.
https://arxiv.org/abs/1711.11575
https://arxiv.org/abs/1904.11491
https://arxiv.org/abs/1910.10683
https://arxiv.org/abs/2002.12804

@alihassanijr thanks for your timely response.

Thanks for your awesome work!