David Dahan
Results
2
comments of
David Dahan
does someone know if longT5 and all T5 models are blocked by bias support in flash attention ? https://github.com/Dao-AILab/flash-attention/pull/617
@ENate I was trying to understand the status and have an estimation of the code change to see if I can contribute.