Jayaditya Shah

Results 1 comments of Jayaditya Shah

It doesn't work with GPT-OSS 20/120B even though vLLM is now supported, Flash Attention 3 is not.