apple / ml-mdm

Train high-quality text-to-image diffusion models in a data & compute efficient manner

Home Page:https://machinelearning.apple.com/research/matryoshka-diffusion-models

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

[Question] Training another layer

Sys-if opened this issue · comments

Hi,

First of all, thank you for sharing your code! I've been studying your paper and code, but I’m a bit unclear on how you implemented progressive training. Could you provide some insights into this? For example, do you freeze the inner U-Net for a certain number of iterations?

I'm replying to my own comment, as it might be useful for others. I have been experimenting with my own model and have implemented what I understood from the paper by progressively introducing the loss of the new high-resolution layer into the general loss, with a weight increasing from 0 to X. I also used a similar weight to progressively introduce the new layer features (as in Progressive Growing of GANs).