lliuz / ARFlow

The official PyTorch implementation of the paper "Learning by Analogy: Reliable Supervision from Transformations for Unsupervised Optical Flow Estimation".

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

scaling up the loss before calculating gradient

hanit92 opened this issue · comments

Hi,
first of all, great paper and great code, thank you for sharing it :)
I was wondering - why do you scale up the loss before the backward() call (multiplying by 1024.), and then dividing it again before the weights update?

It's a historical part from mixed precision training. I don't remember if it had any effect on the results in this project.
You can refer to this link for the purpose of this part.