Lyken17 / pytorch-OpCounter

Count the MACs / FLOPs of your PyTorch model.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Problem in bert

zetaodu opened this issue · comments

I find thop will not calculate the parameters in BertEmbedding and if I define two self_attention blocks in one layer, it will only calculate one.

Second self_attention block should also be used in forward method