ZhikangNiu / encodec-pytorch

unofficial implementation of the High Fidelity Neural Audio Compression

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

traing convergence

yangyyt opened this issue · comments

commented

The effect of multi-gpus training is not as good as that of single-card training, and it feels that multi-card training is quickly overfitted.

Maybe every gpu's codebook weight different?
you can try this code?

# distrib.broadcast_tensors(self.buffers()) # FIXME: this is not working for some reason