Submitted by super_deap t3_11tmpc5 in MachineLearning
super_deap OP t1_jcl3whl wrote
Reply to comment by lucidraisin in [D] PyTorch 2.0 Native Flash Attention 32k Context Window by super_deap
That is understandable. I am working with that assumption as well. (I have failed too many of such experiments to have a blind faith 🙈)
lucidraisin t1_jcl6ecd wrote
no worries, thanks for running the experiments and sharing your results 🙏
[deleted] t1_jcl50r7 wrote
[deleted]
Viewing a single comment thread. View all comments