Submitted by ThoughtOk5558 t3_xvcman in deeplearning
I generated CIFAR10 images using energy based models from the joint distribution of an "airplane: 0" and "bird: 2" classes. As can be see below, the generated images can't be visually classified as any of the CIFAR10 classes, i.e., the prediction should roughly be uniform distribution.
Sampled from the joint distribution of CIFAR10 \"airplane\" and \"bird\" classes.
However, when I make inference using a pre-trained CIFAR10 model link the confidence scores of the predicted classes are very high.
​
I am aware of adversarial attacks and this is kind of adversarial attack.
So, here is my opinion (question). I believe CNNs or any network should consider the visual quality when making a prediction.
Should / can CNNs be improved to act this way?
Thank you.
XecutionStyle t1_ir0dv73 wrote
How do you propose we define quality?