You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The problem is about the fid-score on COCO2014CD. My fid is about 13, while the fid in paper is 5.8. The details for this part is not enough in paper. So it's really hard for reproducing.
-----原始邮件-----
发件人:"Jiahui Zhu" ***@***.***>
发送时间:2024-02-27 19:54:44 (星期二)
收件人: gligen/GLIGEN ***@***.***>
抄送: xiaobo123 ***@***.***>, Author ***@***.***>
主题: Re: [gligen/GLIGEN] Questions about details for reproducing, some of them are not noted in the paper. (Issue #71)
why gligen is always output 512512, how to output 640480
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you authored the thread.Message ID: ***@***.***>
The problem is about the fid-score on COCO2014CD. My fid is about 13, while the fid in paper is 5.8. The details for this part is not enough in paper. So it's really hard for reproducing.
So my questions can be concluded as following:
Is the pretrained model downloaded from here? https://ommer-lab.com/files/latent-diffusion/nitro/txt2img-f8-large/model.ckpt
Dropping probability for caption is 0.5 in github config file, and 0.1 in paper. Which one should i choose?
As we all know, ldm resolution is 256. Is gligen finetuned on 256 or 512 ?
While training, the warm-up is cosine or constant? Is 100,000 step enough for COCO?
In UNet, only gate-attention is not freezed, right?
The text was updated successfully, but these errors were encountered: