TensorRT icon indicating copy to clipboard operation
TensorRT copied to clipboard

NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.

Results 628 TensorRT issues
Sort by recently updated
recently updated
newest added

[Functions](https://github.com/gramalingam/onnx/blob/main/docs/IR.md#functions) is a feature in onnx, which can be thought of as an operator combined with an implementation of the operator using other, more primitive, ops, referred to as the...

ONNX
triaged

When I use FMHA_v2, I found it does not support my scenes. So i wonder is there any way to use fmha_v2 except changing model. Thx a lot.

Plugins
triaged

## Description I get this this issue [08/31/2024-21:29:05] [TRT] [I] [MemUsageChange] TensorRT-managed allocation in building engine: CPU +6, GPU +64, now: ## Environment **TensorRT Version**:8.5.2.2 **NVIDIA GPU**:Orin NX **NVIDIA Driver...

needs-info
triaged

## Description For the following onnx model, ![Image](https://github.com/user-attachments/assets/9cc84911-32af-4b8a-870f-379583576e52) it can be imported by the onnx frontend in TensorRT. However, it failes to build the serialized network. The following error message...

Module:ONNX
triaged
stale
waiting for feedback

## Description I encountered the following error when trying to build trt engine from onnx model. ``` 110947: reshape: /Reshape_32 _ /Transpose_17_reshape_output.1-(f16[__mye111018_proxy.1,1,1025,2,801,2][]so[], mem_prop=0) | /Concat_47_output_0' castIn.1-(f16[__mye111018_proxy.1,1,2050,801,2][]so[], mem_prop=0), stream = 0...

Module:Engine Build
triaged

I had an export issue for attention layer in a video transformer architecture in tensorrt 10.7 and seemed to be fixed in 10.11. However when I exported using 10.11 on...

Module:ONNX
triaged
stale
waiting for feedback

I am trying to generate the model engine from the "ESS DNN stereo disparity" onnx file on my jetson orin nano using trtexec (and tensorrt10.3.0). I am using the model...

Module:ONNX
triaged

## Description I am encountering an accuracy discrepancy between ONNX inference and TensorRT FP32 inference. ## Environment **TensorRT Version**: 10.8.0.43 **NVIDIA GPU**: RTX 3060 **NVIDIA Driver Version**: 560.35.05 **CUDA Version**:...

**System Information** * **OS**: Ubuntu 22.04 * **GPU**: NVIDIA RTX 4090 * **TensorRT Version**: 10.11.0.33 * **PyTorch Version**: 2.7.0 * **ONNX Opset**: 14 --- ### 🧠 Problem Summary I converted...

Module:Performance
triaged
internal-bug-tracked