Submitted by SleekEagle t3_ye0iw7 in MachineLearning
Hey everyone!
Over the past couple years, we've all seen the incredible performance of Diffusion Models in Stable Diffusion, DALL-E 2, etc. A team at MIT recently published a new model that is inspired by physics much like diffusion models.
I wrote an introduction to Poisson Flow Generative Models as an explainer for the models, which use concepts from electrodynamics to generate data.
​
Some highlights:
- 10-20x faster than diffusion models on image generation with comparable performance
- Invertible mapping akin to normalizing flows allows for explicit likelihood estimate and a well-structured latent space
- Operates on a deterministic ODE and has no stochastic element unlike diffusion models and score-based SDE approaches
Here are some example CelebA images being generated with PFGMs:
https://i.redd.it/9f826fz726w91.gif
Note that the highest resolution images explored in the paper are LSUN bedroom 256 x 256 images
Looking forward to answering any questions!
Serverside t1_itw9tz2 wrote
Ok, I'll bite. It looks cool from what I see in the blog. How does the model being deterministic impact (or not impact) the generative capabilities? I would think that a deterministic mapping from original data to uniform angles would not perform as well when wanting to interpolate or extrapolate (for example, like VAEs vs normal autoencoders).