TensorRT
TensorRT copied to clipboard
how to support FlashAttention2?
Hi, does the currently transferred trt engine support flash focus2 by default? If it is not supported by default, how should I use fa2 in the output engine? Thanks,