Ding Qiang
Ding Qiang
flash attention 已经被整合到 torch 2 了 https://pytorch.org/docs/stable/generated/torch.nn.functional.scaled_dot_product_attention.html
很荣幸我的建议被采纳。我还想问问您对之前`save_model`出错的具体原因有什么看法吗?我还没想通。 It's my pleasure to see my advice being accepted. Moreover, may I ask you if you have any comment on why `save_model` went wrong previously? I have not figured...
> 很荣幸我的建议被采纳。我还想问问您对之前`save_model`出错的具体原因有什么看法吗?我还没想通。 It's my pleasure to see my advice being accepted. Moreover, may I ask you if you have any comment on why `save_model` went wrong previously? I have not...