jayparks / transformer

A Pytorch Implementation of "Attention is All You Need" and "Weighted Transformer Network for Machine Translation"

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

How to keep constrains of sum(k)=1 and sum(α)=1?

sunzewei2715 opened this issue · comments

In the original paper(weighted transformer), the author mentioned that "all bounds are respected during each training step by projection."

I have no idea what "by project" means and don't know how to keep the constrains of sum(k)=1 and sum(α)=1.

It seems there is no particular processing in this repository except for initialization. Could you please explain?