Submitted by Shardsmp t3_zil35t in MachineLearning
herokocho t1_izxnzhd wrote
Reply to comment by Shardsmp in [D] Does Google TPU v4 compete with GPUs in price/performance? by Shardsmp
not aware of any good comparisons out there, this is all anecdata from looking at profiler traces when training diffusion models and noticing that I was communication bottlenecked even on TPUs, so on GPUs it would be much worse.
it's usually better to use TPU as soon as you'd have to use multiple GPU nodes, and basically always better at v4-128 scale and above (v4-128 has 2x faster interconnect than anything smaller).
Viewing a single comment thread. View all comments