Use fill_argument for literals that have the same value
Codecov Report
Attention: Patch coverage is 0% with 11 lines in your changes missing coverage. Please review.
| Files with missing lines | Patch % | Lines |
|---|---|---|
| src/module.cpp | 0.00% | 11 Missing :warning: |
Additional details and impacted files
@@ Coverage Diff @@
## develop #3815 +/- ##
===========================================
- Coverage 92.04% 92.00% -0.03%
===========================================
Files 531 531
Lines 24526 24535 +9
===========================================
Hits 22573 22573
- Misses 1953 1962 +9
| Files with missing lines | Coverage Δ | |
|---|---|---|
| src/module.cpp | 85.70% <0.00%> (-1.02%) |
:arrow_down: |
:rocket: New features to boost your workflow:
- :snowflake: Test Analytics: Detect flaky tests, report on failures, and find test suite problems.
/AzurePipelines run
Azure Pipelines successfully started running 1 pipeline(s).
Clang tidy thinks things are too complex @pfultz2
Clang tidy thinks things are too complex @pfultz2
Yea I need to do some refactoring here.
| Test | Batch | Rate new a68707 |
Rate old e90b37 |
Diff | Compare |
|---|---|---|---|---|---|
| torchvision-resnet50 | 64 | 3,238.28 | 3,237.18 | 0.03% | :white_check_mark: |
| torchvision-resnet50_fp16 | 64 | 6,907.17 | 6,900.67 | 0.09% | :white_check_mark: |
| torchvision-densenet121 | 32 | 2,447.72 | 2,446.02 | 0.07% | :white_check_mark: |
| torchvision-densenet121_fp16 | 32 | 4,215.57 | 4,218.17 | -0.06% | :white_check_mark: |
| torchvision-inceptionv3 | 32 | 1,618.99 | 1,618.27 | 0.04% | :white_check_mark: |
| torchvision-inceptionv3_fp16 | 32 | 2,706.13 | 2,706.57 | -0.02% | :white_check_mark: |
| cadene-inceptionv4 | 16 | 756.23 | 756.25 | -0.00% | :white_check_mark: |
| cadene-resnext64x4 | 16 | 814.25 | 814.61 | -0.04% | :white_check_mark: |
| slim-mobilenet | 64 | 7,433.22 | 7,432.36 | 0.01% | :white_check_mark: |
| slim-nasnetalarge | 64 | 216.88 | 216.81 | 0.03% | :white_check_mark: |
| slim-resnet50v2 | 64 | 3,442.58 | 3,442.40 | 0.01% | :white_check_mark: |
| bert-mrpc-onnx | 8 | 1,143.33 | 1,144.28 | -0.08% | :white_check_mark: |
| bert-mrpc-tf | 1 | 456.98 | 457.20 | -0.05% | :white_check_mark: |
| pytorch-examples-wlang-gru | 1 | 492.50 | 495.12 | -0.53% | :white_check_mark: |
| pytorch-examples-wlang-lstm | 1 | 450.21 | 458.39 | -1.78% | :white_check_mark: |
| torchvision-resnet50_1 | 1 | 820.61 | 811.71 | 1.10% | :white_check_mark: |
| cadene-dpn92_1 | 1 | 427.26 | 429.64 | -0.55% | :white_check_mark: |
| cadene-resnext101_1 | 1 | 391.26 | 391.62 | -0.09% | :white_check_mark: |
| onnx-taau-downsample | 1 | 395.09 | 394.28 | 0.20% | :white_check_mark: |
| dlrm-criteoterabyte | 1 | 32.21 | 32.20 | 0.03% | :white_check_mark: |
| dlrm-criteoterabyte_fp16 | 1 | 51.11 | 51.26 | -0.28% | :white_check_mark: |
| agentmodel | 1 | 10,475.77 | 10,217.13 | 2.53% | :white_check_mark: |
| unet_fp16 | 2 | 59.31 | 59.41 | -0.17% | :white_check_mark: |
| resnet50v1_fp16 | 1 | 1,088.67 | 1,083.69 | 0.46% | :white_check_mark: |
| resnet50v1_int8 | 1 | 1,051.41 | 1,063.44 | -1.13% | :white_check_mark: |
| bert_base_cased_fp16 | 64 | 1,162.83 | 1,163.13 | -0.03% | :white_check_mark: |
| bert_large_uncased_fp16 | 32 | 356.18 | 356.06 | 0.03% | :white_check_mark: |
| bert_large_fp16 | 1 | 199.82 | 200.23 | -0.20% | :white_check_mark: |
| distilgpt2_fp16 | 16 | 2,226.65 | 2,225.52 | 0.05% | :white_check_mark: |
| yolov5s | 1 | 543.71 | 544.04 | -0.06% | :white_check_mark: |
| tinyllama | 1 | 43.69 | 43.63 | 0.14% | :white_check_mark: |
| vicuna-fastchat | 1 | 44.78 | 43.88 | 2.03% | :white_check_mark: |
| whisper-tiny-encoder | 1 | 419.35 | 419.66 | -0.08% | :white_check_mark: |
| whisper-tiny-decoder | 1 | 412.40 | 410.82 | 0.38% | :white_check_mark: |
| llama2_7b | 1 | nan | nan | nan% | :x: |
| qwen1.5-7b | 1 | 23.43 | 23.43 | 0.03% | :white_check_mark: |
| phi3-3.8b | 1 | nan | nan | nan% | :x: |
| mask-rcnn | 1 | 18.72 | 18.93 | -1.14% | :white_check_mark: |
| llama3-8b | 1 | 21.61 | 21.64 | -0.11% | :white_check_mark: |
| whisper-large-encoder | 1 | 10.17 | 10.18 | -0.01% | :white_check_mark: |
| whisper-large-decoder | 1 | 98.98 | 98.99 | -0.01% | :white_check_mark: |
| mistral-7b | 1 | 23.68 | 23.65 | 0.11% | :white_check_mark: |
| FLUX.1-schnell | 1 | 914.14 | 912.01 | 0.23% | :white_check_mark: |
| nan | nan | nan | nan | nan% | :x: |
This build is not recommended to merge :red_circle:
:x:bert-mrpc-tf: ERROR - check error output
2025-04-22 16:36:34.331801: I tensorflow/core/platform/cpu_feature_guard.cc:210] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: SSE3 SSE4.1 SSE4.2 AVX AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
WARNING: All log messages before absl::InitializeLog() is called are written to STDERR
I0000 00:00:1745357799.967252 163472 gpu_device.cc:2022] Created device /job:localhost/replica:0/task:0/device:GPU:0 with 62973 MB memory: -> device: 0, name: AMD Instinct MI250X/MI250, pci bus id: 0000:32:00.0
WARNING: All log messages before absl::InitializeLog() is called are written to STDERR
I0000 00:00:1745357800.890619 163472 mlir_graph_optimization_pass.cc:401] MLIR V1 optimization pass is not enabled
2025-04-22 16:36:50.983362: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
2025-04-22 16:36:50.983403: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
2025-04-22 16:36:50.983440: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
2025-04-22 16:36:50.983481: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
2025-04-22 16:36:50.983516: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
2025-04-22 16:36:50.983694: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
2025-04-22 16:36:50.983723: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
2025-04-22 16:36:50.983768: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
error: Failure when generating HSACO
error: Failure when generating HSACO
error: Failure when generating HSACO
error: Failure when generating HSACO
error: Failure when generating HSACO
error: Failure when generating HSACO
error: Failure when generating HSACO
error: Failure when generating HSACO
2025-04-22 16:36:50.984935: E tensorflow/compiler/mlir/tools/kernel_gen/tf_framework_c_interface.cc:228] INTERNAL: Generating device code failed.
2025-04-22 16:36:50.986148: W tensorflow/core/framework/op_kernel.cc:1829] UNKNOWN: JIT compilation failed.
2025-04-22 16:36:50.986166: I tensorflow/core/framework/local_rendezvous.cc:405] Local rendezvous is aborting with status: UNKNOWN: JIT compilation failed.
[[{{node import/bert/embeddings/LayerNorm/moments/SquaredDifference}}]]
2025-04-22 16:36:50.986177: I tensorflow/core/framework/local_rendezvous.cc:405] Local rendezvous is aborting with status: UNKNOWN: JIT compilation failed.
[[{{node import/bert/embeddings/LayerNorm/moments/SquaredDifference}}]]
[[import/loss/output/_21]]
2025-04-22 16:36:50.986191: I tensorflow/core/framework/local_rendezvous.cc:424] Local rendezvous recv item cancelled. Key hash: 11217777527359497193
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/client/session.py", line 1407, in _do_call
return fn(*args)
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/client/session.py", line 1390, in _run_fn
return self._call_tf_sessionrun(options, feed_dict, fetch_list,
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/client/session.py", line 1483, in _call_tf_sessionrun
return tf_session.TF_SessionRun_wrapper(self._session, options, feed_dict,
tensorflow.python.framework.errors_impl.UnknownError: 2 root error(s) found.
(0) UNKNOWN: JIT compilation failed.
[[{{node import/bert/embeddings/LayerNorm/moments/SquaredDifference}}]]
[[import/loss/output/_21]]
(1) UNKNOWN: JIT compilation failed.
[[{{node import/bert/embeddings/LayerNorm/moments/SquaredDifference}}]]
0 successful operations.
0 derived errors ignored.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/src/AMDMIGraphX/tools/accuracy/accuracy_checker.py", line 340, in
main()
File "/src/AMDMIGraphX/tools/accuracy/accuracy_checker.py", line 324, in main
y_out = sess.run(y, feed_dict=tf_dict)
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/client/session.py", line 977, in run
result = self._run(None, fetches, feed_dict, options_ptr,
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/client/session.py", line 1220, in _run
results = self._do_run(handle, final_targets, final_fetches,
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/client/session.py", line 1400, in _do_run
return self._do_call(_run_fn, feeds, fetches, targets, options,
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/client/session.py", line 1426, in _do_call
raise type(e)(node_def, op, message) # pylint: disable=no-value-for-parameter
tensorflow.python.framework.errors_impl.UnknownError: Graph execution error:
Detected at node 'import/bert/embeddings/LayerNorm/moments/SquaredDifference' defined at (most recent call last):
Node: 'import/bert/embeddings/LayerNorm/moments/SquaredDifference'
Detected at node 'import/bert/embeddings/LayerNorm/moments/SquaredDifference' defined at (most recent call last):
Node: 'import/bert/embeddings/LayerNorm/moments/SquaredDifference'
2 root error(s) found.
(0) UNKNOWN: JIT compilation failed.
[[{{node import/bert/embeddings/LayerNorm/moments/SquaredDifference}}]]
[[import/loss/output/_21]]
(1) UNKNOWN: JIT compilation failed.
[[{{node import/bert/embeddings/LayerNorm/moments/SquaredDifference}}]]
0 successful operations.
0 derived errors ignored.
Original stack trace for 'import/bert/embeddings/LayerNorm/moments/SquaredDifference'::red_circle:unet: FAILED: MIGraphX is not within tolerance - check verbose output
:red_circle:bert_large_uncased_fp16: FAILED: MIGraphX is not within tolerance - check verbose output
:x:llama2_7b: ERROR - check error output
Traceback (most recent call last):
File "/src/AMDMIGraphX/tools/accuracy/accuracy_checker.py", line 340, in
main()
File "/src/AMDMIGraphX/tools/accuracy/accuracy_checker.py", line 205, in main
model = migraphx.parse_onnx(model_name, default_dim_value=batch)
RuntimeError: /src/AMDMIGraphX/src/onnx/onnx_parser.cpp:265: parse_from: PARSE_FROM: Failed reading onnx file: /new-saved-models/llama2_7b/decoder_model.onnx:x:qwen1.5-7b: ERROR - check error output
usage: accuracy_checker.py [-h] [--onnx ONNX] [--tf TF] [--provider PROVIDER]
[--batch BATCH] [--fill1] [--fill0] [--fp16]
[--argmax] [--verbose] [--tolerance TOLERANCE]
[--input-dim INPUT_DIM] [--target TARGET]
[--ort-run] [--ort-logging]
[--disable-offload-copy] [--disable-fast-math]
[--exhaustive_tune]
accuracy_checker.py: error: unrecognized arguments: input_ids attention_mask position_ids 1 256 @attention_mask 1 256 @position_ids 1 256:x:phi3-3.8b: ERROR - check error output
usage: accuracy_checker.py [-h] [--onnx ONNX] [--tf TF] [--provider PROVIDER]
[--batch BATCH] [--fill1] [--fill0] [--fp16]
[--argmax] [--verbose] [--tolerance TOLERANCE]
[--input-dim INPUT_DIM] [--target TARGET]
[--ort-run] [--ort-logging]
[--disable-offload-copy] [--disable-fast-math]
[--exhaustive_tune]
accuracy_checker.py: error: unrecognized arguments: input_ids attention_mask position_ids 1 256 @attention_mask 1 256 @position_ids 1 256:red_circle:mask-rcnn: FAILED: MIGraphX is not within tolerance - check verbose output
:x:#whisper-large-encoder: ERROR - check error output
Traceback (most recent call last):
File "/src/AMDMIGraphX/tools/accuracy/accuracy_checker.py", line 340, in
main()
File "/src/AMDMIGraphX/tools/accuracy/accuracy_checker.py", line 205, in main
model = migraphx.parse_onnx(model_name, default_dim_value=batch)
RuntimeError: /src/AMDMIGraphX/src/include/migraphx/op/convolution.hpp:100: normalize_compute_shape: CONVOLUTION: mismatched channel numbers
Two reviewers on this, curious why this hasn't made it in. Just CI?