Submitted by Open-Dragonfly6825 t3_10s3u1s in deeplearning
suflaj t1_j731s6u wrote
Reply to comment by Open-Dragonfly6825 in Why are FPGAs better than GPUs for deep learning? by Open-Dragonfly6825
I mean kernels in the sense of functions.
> Why wouldn't GPU parallelization make inference faster?
Because most DL models are deep, and not exactly wide. I've explained already, deep means a long serial chain. Not parallelizable outside of data parallelism, which doesn't speed up inference, and model parallelism (generally not implemented, and has heavy IO costs).
Wide models and how they become equivalent to deep ones are unexplored, although they are theoretically just as expressive.
Open-Dragonfly6825 OP t1_j73258d wrote
Ok, that makes sense. Just wanted to confirm I understood it well.
Thank you.
Viewing a single comment thread. View all comments