Iman Tabrizian
Iman Tabrizian
I think this should work fine when using S3 or local model repository. I have filed a ticket to look into this (DLIS-3870).
I tried a small example locally and it did return an error if there wasn't enough shared memory. @rmccorm4 Could you please file a ticket for this issue so that...
@gandharv-kapoor Your issue looks like a different problem since I don't see logs from Python backend. It looks like you are using an ONNX model. Could you please open a...
We don't have any plans for supporting http2 right now. Can you share the use-cases that you have in mind for http2?
I don't think there is a way to get around this. The backend shared library needs to be reloaded to pick up the new changes. Currently, if a backend library...
Hi @shekhardw, we do have an example for using Triton C-API in the [simple.cc](https://github.com/triton-inference-server/server/blob/a3e96fa1d46b7a0c98280558b133f613a3c09d1a/src/simple.cc) file. This example reused some of the structs like `TRITONSERVER_InferenceRequest` object which you might be interested...
Closing due to in-activity.
cc @jbkyang-nvi @CoderHam
You can use the [model repository agents](https://github.com/triton-inference-server/server/blob/main/docs/repository_agents.md) to add custom operations that will be called when the model is loading/unloading. > How to deploy triton inference server on premise locally,...
@sourabh-burnwal I think you can encrypt the model and store the encrypted models in the model repository and use the model repository agent to decrypt the model when the model...