TensorRT icon indicating copy to clipboard operation
TensorRT copied to clipboard

NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.

Results 628 TensorRT issues
Sort by recently updated
recently updated
newest added

The results of tensorRT10.3 are different from those of TensorRT 10.8. The results of 10.8 are as expected, so the results of 10.3 are wrong. I want to use 10.3...

wontfix
triaged

Hi, i try to create tensorrt engine from an onnx model. I tried a few things and here are the inference latencies. Why is 3. and 4. performing worse than...

Module:Performance
triaged

## Description After upgrading tensorrt to 10.8, the model accuracy decreased. After setting all nodes of the model to output, the model accuracy was aligned. It was suspected that the...

triaged
Module:Accuracy
internal-bug-tracked

## Description I tried to convert model [CO-DETR](https://github.com/marcoslucianops/DeepStream-Yolo/blob/master/docs/CODETR.md) to TRT, but it fails with error below ```bash [12/12/2024-02:17:38] [E] Error[10]: Could not find any implementation for node {ForeignNode[/0/Cast_3.../0/backbone/Reshape_3 + /0/backbone/Transpose_3]}....

Module:Engine Build
triaged

--- name: Report a TensorRT issue about: Failed to export ONNX Model (Transformer) to TensorRT title: 'Assertion failed of TensorRT 10.9 when export ONNX Model to TensorRT (polygraphy and trtexec)'...

Module:ONNX
triaged

## Description As I had reflected in this [Skipping tactic 0x0000000000000000 due to Myelin error" degrade performance.](https://github.com/NVIDIA/TensorRT/issues/2838),set layer precision may failed in TensorRT 8.4.3 due to the ConstShuffleFusion. In these...

triaged

## Description It seems that TensorRT cannot eliminate a concatenation layer if there is an upstream custom layer. In a simple model that uses all standard operators, TensorRT engine building...

Feature Request
Module:Documentation
triaged

## Description Since implicit quantization is deprecated, I started migrating my model pipeline to explicit quantization. However, I encountered some issues: 1. Different behaviour with concat: With implicit quantization the...

triaged
Module:Quantization

During quantization (using pytorch_quantization), the qparams (scale and zero_point) of old Conv is computed using Calibrator. However, when the Conv and Batch Normalization (BN) layers are fused, the weights and...

triaged
Module:Quantization

When I use `trtexec` and I mix TensorRT engine plan files across different GPU models, I can get a warning: ``` Using an engine plan file across different models of...

Feature Request
triaged