cutlass topic
List
cutlass repositories
flash_attention_inference
20
Stars
2
Forks
Watchers
Performance of the C++ interface of flash attention and flash attention v2 in large language model (LLM) inference scenarios.
flux
650
Stars
42
Forks
Watchers
A fast communication-overlapping library for tensor/expert parallelism on GPUs.