flash-attention
flash-attention copied to clipboard
install the flash-attention
hello, i run the pip install flash-attn, but it fails. the gpu i used is tesla m40, can the flash attn be installed on m40?
No, m40 is of Maxwell generation (sm_52). FlashAttention right now supports Ampere (sm_80) and after.
@tridao
Can you explain why it does not support maxwell? Are there any technical limitations?
@tridao
Can you explain why it does not support maxwell? Are there any technical limitations?
Because it takes work. Maxwell does not have tensor cores. Someone would need to implement a version that doesn't use tensor cores.