Michal Guzek

Results 93 comments of Michal Guzek

I don't see any images for this older cuda version: https://hub.docker.com/r/nvidia/cuda/tags?name=devel-ubuntu18.04

> We've identified that the output from the layer '/bm_decoder/output_hypernetworks_mlps.0/layers.0/Gemm' in the TensorRT model is incorrect, while the output from the same layer in the ONNX model is correct. The...

Sorry, but can you also specify your OS?

@shingjan , I will be closing this ticket due to our policy to close tickets with no activity for more than 21 days after a reply had been posted. Please...

Does your project require the use of the specific calibrator from our samples or have you tried looking into Model Optimizer (https://github.com/NVIDIA/TensorRT-Model-Optimizer) for the calibration purposes (it should ensure your...

> So, will the implicit quantization still be supported or you, guys, are completely switching to the explicit one only ? I would like to avoid changing things in my...

@CoinCheung , yes, if you're using PyTorch then Model Optimizer can automate more parts of that flow: https://nvidia.github.io/TensorRT-Model-Optimizer/guides/_pytorch_quantization.html in contrast to manually configuring the optimization and quantization steps within TRT...

Hi @adaber, 1. I'm not up to speed with that sample, I'd have to check how to translate that calibrator: https://github.com/NVIDIA/TensorRT/blob/release/10.0/samples/python/efficientdet/build_engine.py#L37 to Model Optimizer's config. If this is an immediate...

> I mean we use ModelOpt to quantize the pytorch model into int8, and then export it to onnx from pytorch, then use tensorrt to build it into tensorrt engine?...

> [@moraxu](https://github.com/moraxu) Hi, I packed up the code and the associated onnx file, it is accessable here: > > https://github.com/CoinCheung/eewee/releases/download/0.0.0/code.zip > > There is a readme file in the zip...