server icon indicating copy to clipboard operation
server copied to clipboard

Yolo in the Cloud

Open alexkehoe opened this issue 2 years ago • 1 comments

Hi we have a yolov model that we've converted to tensorRT for inference on the edge. We'd like to port this model to the cloud so we can process +100 streams. Can triton be used to run such a model in the cloud? Ie we would send an RTSP or WebRTC video stream to triton and then send just the bounding box results back to our local server for display to the user? FPS would be say 10hz

If Triton is not appropriate for this use case, can you recommend another way of trying to scale deployment of a yolo model for +100 video streams?

alexkehoe avatar Sep 22 '22 08:09 alexkehoe

Hi, please read the quickstart guide to see details for the Tritonserver.

I believe there are users that have used yolo models with Triton. We require the data to be sent from Triton clients. Please read up about the Triton Architecture.

https://github.com/triton-inference-server/client#triton-client-libraries-and-examples has some more examples you can refer to

jbkyang-nvi avatar Sep 22 '22 21:09 jbkyang-nvi

@jbkyang-nvi all those examples are using image files or image folders. Are there any examples with video streams?

OctaM avatar Sep 29 '22 13:09 OctaM

Similar question: https://github.com/triton-inference-server/server/issues/4487

tanmayv25 avatar Sep 30 '22 19:09 tanmayv25

https://developer.nvidia.com/deepstream-sdk to learn more about deepstream

jbkyang-nvi avatar Sep 30 '22 19:09 jbkyang-nvi

Closing issue due to lack of activity. Please re-open the issue if you would like to follow up with this issue

jbkyang-nvi avatar Nov 22 '22 03:11 jbkyang-nvi