David Dahan

Results 2 comments of David Dahan

does someone know if longT5 and all T5 models are blocked by bias support in flash attention ? https://github.com/Dao-AILab/flash-attention/pull/617

@ENate I was trying to understand the status and have an estimation of the code change to see if I can contribute.