flash-attention
flash-attention copied to clipboard
There is no cu123 but cu124 for PyTorch
Is there a mistake for the released whl?
No mistake
No mistake
Can we get a cuda 124 version. As pytorch supports 12.4 but 12.3.
12.3 will work
It doesnt. Neither would 12.1 https://github.com/Dao-AILab/flash-attention/issues/1093
cuda minor version are compatible
II've compiled a version that works with CUDA 12.4 and PyTorch 2.4 for Windows. You can find the wheel at this address: https://huggingface.co/windzmf/flash_attn-2_6_3_cu124torch2_4cxx11_win_amd64.whl/tree/main This wheel is specifically compiled for:
- Flash Attention version 2.6.3
- CUDA 12.4
- PyTorch 2.4
- Windows AMD64 platform
- Python 3.11 ============================================================= 我编译了一个适用于CUDA 12.4和PyTorch 2.4的Windows版本。 您可以在以下地址找到: https://huggingface.co/windzmf/flash_attn-2_6_3_cu124torch2_4cxx11_win_amd64.whl/tree/main -Flash attention 2.6.3 -CUDA 12.4 -PyTorch 2.4 -Windows AMD64平台 -Python 3.11