Ted Themistokleous
Ted Themistokleous
### Problem Description Seeing GPU fault when running the onnxruntime-inference-examples script using reduced layer bert models during benchmarking. It appears quantization/calibration steps work and the issue arises during inference. ```...
Fuse average pooling with convolution ``` @77 = gpu::code_object[code_object=9344,symbol_name=pad_kernel,global=262848,local=1024,](@57,@76) -> float_type, {1, 192, 37, 37}, {262848, 1369, 37, 1} @78 = load[offset=705600,end=1646400](@1) -> float_type, {1, 192, 35, 35}, {235200, 1225,...
Add additional flags to the MIGraphX Driver perf to allow for different timing methodologies to match how we run a model through onnxruntime. Handling things this way allows us to...
Viewing code in Onnxruntime for IsUnsupportedOpMode() before we compile in a model Seeing cases which are handled in MIGraphX correctly but are still marked as not supported in Onnxruntime -...
Sort out build pipeline or RHEL builds - [x] Get MIGraphX Build - [x] Get Onnxruntime build with wheels for MIGraphX EP - [ ] Get Onnxruntime build with wheels...
Similar to softmax cross entropy loss #3008 , thus able to leverage all the existing logic of this and test method. Added verify tests for the 3 reduction cases in...
Changes to MatMul parser to handle the Microsoft Contrib operator MatMulintegarToFloat Since we have the scale and zero points in our operands we can just perform a multiplied after int8...
Needed to be commented out by #3340 - [ ] Uncomment and investigate - [ ] Fix code that causes this if able - [ ] determine if test stops...
Disabled tidy check in #3340 - [ ] Uncomment and investigate - [ ] Fix code if able - [ ] reneable check