Caspar23-Wang
Results
1
comments of
Caspar23-Wang
We have the same problem when using FlashAttention. (We’re using a 4090 GPU.) The detail Error below: python example_mla_decode.py Traceback (most recent call last): File "/root/autodl-tmp/tilelang/example_mla_decode.py", line 318, in main(batch,...