Submitted by MazenAmria t3_zhvwvl in deeplearning
sqweeeeeeeeeeeeeeeps t1_izob3yb wrote
MNIST and Imagenrt is a huge range. Try something in between, preferably multiple. For example CIFAR-10 and CIFAR-100. I would expect it to perform more similarly to the full SWIN model on citar-10 because less data complexity.
MazenAmria OP t1_izon556 wrote
> I would expect it to perform more similarly to the full SWIN model on citar-10 because less data complexity.
And that's the problem. If I got say 98% accuracy on CIFAR-10 using SWIN-Tiny and then got the same 98% with a smaller model then I'm not proving anything. There are many simple models that can get 98% on CIFAR-10 so what improvement did I introduce to the SWIN-Tiny? But doing the same thing with ImageNet would be different.
sqweeeeeeeeeeeeeeeps t1_izphlmd wrote
? You are proving your SWIN model is overparameterized for CIFAR. Make an EVEN simpler model than those, you prob won’t be able to with off the shelf distillation. Doing this just for ImageNet literally doesn’t change anything. It’s just a different more complex dataset.
What’s your end goal? To come up with a distillation technique to make NN’s more efficient and smaller?
MazenAmria OP t1_izpii1s wrote
To examine SWIN itself whether it's overparameterized or not.
sqweeeeeeeeeeeeeeeps t1_izspv5o wrote
Showing you can create a smaller model with the same performance means SWIN is overparameterized for that given task. Give it datasets with varying complexity, not just one single one.
sqweeeeeeeeeeeeeeeps t1_izq6vbc wrote
It is.
pr0d_ t1_izqjmmk wrote
yeah as per my comment, the DEiT papers explored knowledge distillation based off Vision Transformers. What you want to do here is probably similar, and the resources needed to prove it is huge to say the list. Any chance you've discussed this with your advisor?
MazenAmria OP t1_izrgnco wrote
I remember reading it, I'll read it again and discuss it. Thanks.
Viewing a single comment thread. View all comments