Yong Shan
Yong Shan
Hi, I have verified that the system installed cuda version matches the torch cuda, both cu118. I haven't set `DS_SKIP_CUDA_CHECK`.
Sorry for the late reply. There is no other cuda installations, no prints from your referenced function. I still don't know what causes this warning. However, my code runs successfully.
@blefaudeux Hi, how to implement a blocksparse attention supporting attention mask (i.e. shape SxS)? I want to implemnt a sparse attention with specific layout. However, current blocksparse attention only use...
The same question. @TonyNemo
@TonyNemo
The same question. @TonyNemo
Hi, I met the same problem. Do you have any answers? @Helicqin
@bliu3650 Can you share the command you used?
Does anyone want to implement Lion for apex?