Jayaditya Shah
Results
1
comments of
Jayaditya Shah
It doesn't work with GPT-OSS 20/120B even though vLLM is now supported, Flash Attention 3 is not.