halexan

Results 3 issues of halexan

Could your please coded up a wrapper that removes a lot of the manual work in writing up a generic SinkhornTransformer encoder / decoder architecture. Thanks a lot! halexan

### Motivation VLLM has announced their support for running llama3.1-405b-fp8 on 8xA100. This is the [blog](https://blog.vllm.ai/2024/07/23/llama31.html) Does sglang support running DeepSeek-Coder-V2-Instruct-FP8 on 8xA100? ### Related resources _No response_

### Describe the bug ChatTTS部署成功,界面可以看到。 ![ChatTTS部署成功界面图](https://github.com/xorbitsai/inference/assets/17266912/b4bc826a-7c08-44ee-aec0-59c7e8a643c6) 日志: 2024-07-04 00:42:25,572 xinference.model.utils 97 INFO Use model cache from a different hub. 2024-07-04 00:42:26,514 xinference.thirdparty.ChatTTS.core 16784 INFO Load from local: /root/.xinference/cache/ChatTTS 2024-07-04 00:42:27,649...

gpu