Zhendong-Wang / Prompt-Diffusion

Official PyTorch implementation of the paper "In-Context Learning Unlocked for Diffusion Models"

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

train

Tanghui2000 opened this issue · comments

Hello, thank you for your work, which is enlightening to me.
I have some questions to consult. I tried to train grayscale images and color images as a set of "examples", so as to achieve image coloring. However, the generated images are very different from the original ones.
Why does this happen? How to ensure consistency with the original image?

The consistency is maintained because the model is also given the query image as the inputs. As we show in the Figure 3. The training also depends on the dataset, you input and output image pairs should be consistent.

The consistency is maintained because the model is also given the query image as the inputs. As we show in the Figure 3. The training also depends on the dataset, you input and output image pairs should be consistent.

Thank you for helping me. It is now possible to keep the input consistent with the content of the generated image. But the resulting image's contrast and saturation seem to have changed, making it seem more noisy than the original, which bothers me. I'm using the coco dataset. Some of the images in the coco dataset are noisy. Is this misleading? Or is it something else?
屏幕截图 2023-11-29 091525

Yes, here I think image quality is important for finetuning. If the dataset is noisy, you may need longer training to converge and may converge to worse results.