pengxin233

Results 9 issues of pengxin233

### 📚 The doc issue When I call curl http://127.0.0.1:8082/metrics, it always returns empty results, even if it is called after model inference. But there is clearly a corresponding log...

### 📚 The doc issue I want to use grpc in the java service to call torchserve's model, but I don't seem to have found any relevant documentation. ### Suggest...

Can I convert the torchscript model to tensorrt format through torch_tensorrt? Is there any corresponding script that you can give me for reference?

question

### 📚 The doc issue I set the batchSize of the registered model to 10, and then set the micro_batch_size to 1. So for model inference, will it wait for...

### 🐛 Describe the bug The first time I requested torchserve, the score when preheating was inconsistent with the score after preheating was completed. But any score after the warm-up...

triaged

How to serialize the converted engine after using onnx-tensorrt? I didn’t see any relevant content in the document.

I found that the video memory will increase every once in a while during operation. I want to confirm whether there is a risk of video memory leakage in cvcuda...

question
need more info

Hello, can I ask if qdrant can be run on GPU? Will the performance be faster than CPU?

I would like to ask qdrant when will the data be merged into the index when it is inserted?