geekfolk
geekfolk t1_je3qyfr wrote
Reply to comment by Username912773 in [P] Consistency: Diffusion in a Single Forward Pass π by Beautiful-Gur-9456
I donβt know about this model, but GANs are typically smaller than diffusion models in terms of num of params. The image structure thing probably has something to do with the network architecture since GANs rarely use attention blocks and the network architecture of diffusion models is more hybrid (typically CNN + attention)
geekfolk t1_je3qiqw wrote
Reply to comment by huehue9812 in [P] Consistency: Diffusion in a Single Forward Pass π by Beautiful-Gur-9456
R1 is one form of 0-gp, itβs actually introduced in the paper that proposed 0-gp. See my link above
geekfolk t1_je3io3b wrote
Reply to comment by Beautiful-Gur-9456 in [P] Consistency: Diffusion in a Single Forward Pass π by Beautiful-Gur-9456
using pretrained models is kind of cheating, some GANs use this trick too (projected GANs). But as a standalone model, it does not seem to work as well as SOTA GANs (judged by the numbers in the paper)
​
>Still, it's a lot easier than trying to solve any kind of minimax problem.
This is true for GANs in the early days; however, modern GANs are proved to not have mode collapse and the training is proved to converge.
>It's actually reminiscent of GANs since it uses pre-trained networks
I assume you mean distilling a diffusion model in the paper. There have been some attempts to combine diffusion and GANs to get the best of both worlds but afaik none involved distillation, I'm curious if anyone has tried distilling diffusion models into GANs.
geekfolk t1_je23p7c wrote
How is it better than GANs though? or in other words, what's so bad about adversarial training? modern GANs (with zero centered gradient penalties) are pretty easy to train.
geekfolk t1_je59x39 wrote
Reply to comment by Beautiful-Gur-9456 in [P] Consistency: Diffusion in a Single Forward Pass π by Beautiful-Gur-9456
>I think it's worth a shot to replace LPIPS loss and adversarially train it as a discriminator
that would be very similar to this: https://openreview.net/forum?id=HZf7UbpWHuA