Submitted by i_sanitize_my_hands t3_11wreix in MachineLearning
Hello ML sub,
How does one evaluate the quality of training images before actually training a model ? Training a model is surely expensive. What if one had a way of sort of ascertaining that the image quality of a training set for a particular task (say object detection or semantic segmentation etc) ? It doesn't have to be perfect but some kind of hint...
Could you please point me to some papers or studies or discussions on this ?
There are some objective metrics like PSNR or SSIM but they need a reference image
Joel_Duncan t1_jczkbc0 wrote
I keep seeing this getting asked like people are expecting a magic bullet solution.
​
In general you can only get out something within the realm of what you put in.
There are intelligent ways to structure training and models, but you can't fill in expected gaps without training with a reference or a close approximation of what those gaps are.
My best suggestion is to limit your input data or muxed model to specific high resolution subsets.
ex. You can train a LoRa on a small focused subset of data.