pommedeterresautee OP t1_ittsubj wrote
Reply to comment by ptillet in [P] Up to 12X faster GPU inference on Bert, T5 and other transformers with OpenAI Triton kernels by pommedeterresautee
Thank you a lot for *your* work and your message :-)
Regarding the bugs, for now they have been mostly workable, we follow with lots of excitement the MLIR rewriting and try to prepare ourselves.
I am really wondering what will happen to the ML community when Pytorch will release TorchDynamo / Inductor and so many people will start using Triton in their day to day work. Then tens of thousands of people or more with different backgrounds may start writing kernels...
As they say, what a time to be alive!
fakesoicansayshit t1_iufevix wrote
Can this be used in models like Stable Diffusion?
pommedeterresautee OP t1_iuh0y8u wrote
I think so but not tried. Requires to write search / replace patterns
Viewing a single comment thread. View all comments