server
server copied to clipboard
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
Thanks for submitting a PR to Triton! Please go the the `Preview` tab above this description box and select the appropriate sub-template: * [PR description template for Triton Engineers](?expand=1&template=pull_request_template_internal_contrib.md) *...
**Description** I am getting memory corruption issues with stateful bls model, It seems like triton is trying to free some memory which is still in use **Triton Information** 24.07 Are...
We are running `tritonclient[http]=2.41.0` with server running `24.06-py3`. When there are O(600) requests reaching the server we intermittently receive the following error from triton: ``` Traceback (most recent call last):...
**Is your feature request related to a problem? Please describe.** When writing the `model.py` file for a Python backend model, it is very difficult to correctly use `triton_python_backend_utils` (aka `pb_utils`)....
**Description** I am testing sending data received as output from one model as input to my python backend to post process (I will eventually do an ensemble later) The problem...
**Description** I am trying to setup and and build ONNX runtime natively on Windows 10, without docker following the instructions that are mentioned in the [readme ](https://github.com/triton-inference-server/onnxruntime_backend/blob/main/README.md )file of the...
**Description** Hi, I have setup Triton version 2.47 for Windows, along with ONNX runtime backend, based on the assets for Triton 2.47 that are mentioned in this URL : https://github.com/triton-inference-server/server/releases/...
Description of problem: I did some experiments to measure timing performance to compare standalone inference based on a TensorRT model vs Triton serving the TensorRT model using identical input on...
**Is your feature request related to a problem? Please describe.** I have python components that I would like to use in multiple ensembles (both within a container but also in...
**Description** A clear and concise description of what the bug is. Timeout value defined in config.pbtxt is not triggered on defined value, but after the model has finished its current...