Submitted by shahaff32 t3_y22rk0 in MachineLearning
shahaff32 OP t1_is0ths5 wrote
Reply to comment by londons_explorer in [R] Wavelet Feature Maps Compression for Image-to-Image CNNs by shahaff32
This is aimed mostly at edge devices, where an accelerator is not available (e.g. mobile phones), or you want to design a cheaper chip for a product that requires running such networks (e.g. autonomous vehicles)
This work was, in fact, partially supported by AVATAR consortium, aimed at smart vehicles. https://avatar.org.il/
londons_explorer t1_is11x8p wrote
Sure this work was aimed at that, but these same techniques can be used to make a datacenter-scale inference machine into an even more powerful one.
And presumably if a way can be found to do backpropagation in 'wavelet domain', then training could be done like this too.
shahaff32 OP t1_is13c2c wrote
We are in fact doing the backpropagation in the wavelet domain :)
The gradient simply goes through the inverse wavelet transform
​
See WCC/util/wavelet.py in our GitHub repo, lines 52-83 define the forward/backward of WT and IWT.
NeverCast t1_is2f1jt wrote
The immediate use case for me was on autonomous flight vehicles where weight and battery usage matters
Viewing a single comment thread. View all comments