More details about training
Sanster opened this issue · comments
Thanks for sharing the code and dataset. The encoder-only architecture makes DDCP faster and lighter than other methods, I really like the idea. I try to reimplement the paper, however, some training details are missing in the paper.
- What are
α
andβ
used for the pre-train model? In the utilsV4.py it's all equal1
- What are the total epochs for the pre-train model? In train.py, the default
epochs=300
Hi,
1、please see here.
2、We have printed the EPOCH of pre-train model. see here