Submitted by CurrentlyJoblessFML t3_10g1tni in MachineLearning
Hi all,
I am trying to see if I can use DDPM (Denoising Diffusion Probabilistic Model) to denoise images using a supervised learning approach. However, I've learned that DDPM is only for unconditional image generation. Has anyone had experience using conditional DDPM and could help me out with some conceptual questions?
Here's what I'm trying to understand:
-
Say I have a pair of noisy and clean ground truth images.
-
Should I take my clean image and gradually corrupt it by adding gaussian noise in the forward diffusion (FD) process?
-
Could I get the network to learn the reverse diffusion process by giving it the noisy input, the FD noisy image, and positional embeddings? I was planning on concatenating the noisy input with the FD noisy image.
-
During training, the network learns to predict noise at t-1 given the image at t conditioned on the input noisy source image.
Here is an image showing you what I mean. Any thoughts or suggestions would be greatly appreciated. DDPM for image denoising
samb-t t1_j50gpn4 wrote
I think what you're looking for is palette which is for paired image-to-image translation with conditional diffusion models. I believe that approach is exactly what you're describing, concatenating down the channels dimension.