Submitted by midasp t3_y4elh1 in MachineLearning
In Computer Science, it is known that we are very close to the limits of compressing all the information found in an image. There is no way to losslessly compress images much farther. So we've resorted to lossy compression where some of the image's information is thrown away.
Instead of throwing away information, maybe there is another approach to getting smaller image files. What if a significant percentage of that information resides somewhere else?
Suppose we train a ML model (Resnet, Diffusion Models or whatever) on a wide and comprehensive set of images with two tasks. Task #1 is that the model can take an image, I, as input and outputs a smaller encoding, E. And the same model can be used for task #2, take the encoding E as input and give us the same image I as its output.
In this way, the ML model acts as a large external repository of image information that maps between I and E. Instead of transmitting I, we now just need encode I to encoding E, transmit the much smaller E. As long as both transmitter and receiver has the same ML model, the receiver applies the reverse and uses E to decode and get get back the original image, I.
Marvsdd01 t1_isdr7im wrote
take a look at auto-encoders for data compression ;)