TensorRT icon indicating copy to clipboard operation
TensorRT copied to clipboard

Multiple stream share memory allocation

Open Joeyzhouqihui opened this issue 3 years ago • 2 comments

Hi, sorry for bothering. I am wondering if it is possible to let multiple models running on different cuda streams to share one single memory allocator?

Joeyzhouqihui avatar Sep 01 '22 14:09 Joeyzhouqihui

I think it's feasible, see https://docs.nvidia.com/deeplearning/tensorrt/api/c_api/classnvinfer1_1_1_i_gpu_allocator.html @nvpohanh correct me if I'm wrong.

zerollzeng avatar Sep 01 '22 15:09 zerollzeng

From what I see TF-TRT uses a custom GPU allocator and there may be multiple streams running in parallel.

zerollzeng avatar Sep 01 '22 15:09 zerollzeng

Yes, you can use it across multiple streams

nvpohanh avatar Dec 02 '22 10:12 nvpohanh

closed

zerollzeng avatar Dec 04 '22 15:12 zerollzeng