thuml / Flowformer

About Code release for "Flowformer: Linearizing Transformers with Conservation Flows" (ICML 2022), https://arxiv.org/pdf/2202.06258.pdf

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

lra attention相关的代码里 没有使用任何mask来屏蔽padding token

zxgineng opened this issue · comments

感谢分享!
lra attention相关的代码里 好像没有任何mask来屏蔽padding token。你们的实验结果里,就是没有使用任何mask吗?

Hi, we do not mask the padding tokens. If you want, you can add a mask to the newly padded token (but I do not think this will affect the performance).