Michal Guzek
Michal Guzek
@adaber , I don't see Windows in https://nvidia.github.io/TensorRT-Model-Optimizer/getting_started/2_installation.html so will let @riyadshairi979 to confirm
@ckolluru , in theory, INT8 inference should generally offer better performance than FP16 due to lower precision calculations, which use less computational power and memory bandwidth. However, I believe for...
@maisa32 sorry for the late reply, I've just checked with the PM team and: - Long term solution: DLA3 will support both EQ and IQ - Short/medium term solution: -...
CC @akhilg-nv or @asfiyab-nvidia who are more up to speed with demos
@ChuRuaNh0 , I will be closing this ticket due to our policy to close tickets with no activity for more than 21 days after a reply had been posted. Please...
@kimdwkimdw can you send the model to [email protected], I can instance an internal bug for this
@kylechang523 , I will be closing this ticket due to our policy to close tickets with no activity for more than 21 days after a reply had been posted. Please...
Have you serialized the NumPy arrays properly using JSON-compatible formats, i.e. like: ``` input_data = np.random.rand(1, 3, 384, 640).astype(np.float32) input_data_list = input_data.tolist() with open("custom.json", "w") as f: json.dump({"input": input_data_list}, f)...
@yikox , I will be closing this ticket due to our policy to close tickets with no activity for more than 21 days after a reply had been posted. Please...
According to the issue, the problem seems to be with the node that was offloaded to one of our backend DL graph compilers so we can investigate it internally. Can...