Caspar23-Wang

Results 1 comments of Caspar23-Wang

We have the same problem when using FlashAttention. (We’re using a 4090 GPU.) The detail Error below: python example_mla_decode.py Traceback (most recent call last): File "/root/autodl-tmp/tilelang/example_mla_decode.py", line 318, in main(batch,...