Storm

Results 11 issues of Storm

When I conducted experiments using Baichuan 2-13B Chat, I found that the CPU utilization rate of each core was only 20%. I believe it is because BigDL LM did not...

user issue

### Describe the issue ![image](https://github.com/intel/intel-extension-for-pytorch/assets/57557769/5bc5cd5f-4433-479d-9bd2-77144afa6f54) hi, my machine is like this, I want to run benchmark, which command can get the best performance? The order I tried before was this,...

### Describe the issue May I ask if this library supports Qwen-14B Chat? If not supported, are there any adaptation plans in the future?

Query
LLM

When i use python_api_example or streaming_llm python scripts to inference Qwen-14B-Chat,the first two questions were outputted normally, but the third question has been repeating itself since then. I find it...

### Is there an existing issue for this? - [X] I have searched the existing issues ### Environment ```markdown - Milvus version:v2.3.0-1977-g4ae7cabb0-dev - Deployment mode(standalone or cluster):standalone - MQ type(rocksmq,...

kind/bug
triage/needs-information

### Search before asking - [X] I had searched in the [issues](https://github.com/eosphoros-ai/DB-GPT/issues?q=is%3Aissue) and found no similar feature requirement. ### Description Have you considered adding user management and session management functions,...

enhancement
Waiting for reply

### Search before asking - [X] I had searched in the [issues](https://github.com/eosphoros-ai/DB-GPT/issues?q=is%3Aissue) and found no similar issues. ### Operating system information Linux ### Python version information >=3.11 ### DB-GPT version...

bug
Waiting for reply

机器A800,vLLM 0.5.0,prompt是开始,输出max tokens=2048,temperature设0.7 vLLM加载Qwen2-72B-Instruct-gptq-int4,使用vLLM的benchmark脚本来做并发测试,无论是1个并发限制还是10个并发限制,输出均会重复。 https://github.com/vllm-project/vllm/blob/main/benchmarks/benchmark_serving.py ![企业微信截图_1718760902864](https://github.com/QwenLM/Qwen2/assets/57557769/380b10a9-77bf-429c-9e04-1e528fd5e43c) ![企业微信截图_1718760893772](https://github.com/QwenLM/Qwen2/assets/57557769/b1b261da-464f-4d86-b5ae-08eb54ce806f) 当然我也测试了无限制并发的情况下,也会生成重复 ![image](https://github.com/QwenLM/Qwen2/assets/57557769/5da613c0-f9e1-4b2f-bf4e-85d28c720bc8)

enhancement

### Is there an existing issue for this bug? - [X] I have searched the existing issues ### 🐛 Describe the bug 当我启动训练的时候出现了下面的错误,怎么办?Ubuntu 22.04 ``` llama) root@autodl-container-ad594b8360-ad0d4c6e:~/autodl-tmp/ColossalAI/applications/ColossalChat/examples/training_scripts# bash train_sft.sh GPU...

bug