ort
ort copied to clipboard
Does it support TensorRT backend?
Hello, great job. In the README it seems we just support CUDA backend and openVINO backend but how about TensorRT backend which is used in ONNXruntime by default on Nvidia GPUs? Do we have any benchmark about speedup? Do we aim at providing a solution for deploying transformers? How about the roadmap?