AMDMIGraphX icon indicating copy to clipboard operation
AMDMIGraphX copied to clipboard

Use fill_argument for literals that have the same value

Open pfultz2 opened this issue 10 months ago • 8 comments

pfultz2 avatar Feb 14 '25 17:02 pfultz2

Codecov Report

Attention: Patch coverage is 0% with 11 lines in your changes missing coverage. Please review.

Files with missing lines Patch % Lines
src/module.cpp 0.00% 11 Missing :warning:
Additional details and impacted files
@@             Coverage Diff             @@
##           develop    #3815      +/-   ##
===========================================
- Coverage    92.04%   92.00%   -0.03%     
===========================================
  Files          531      531              
  Lines        24526    24535       +9     
===========================================
  Hits         22573    22573              
- Misses        1953     1962       +9     
Files with missing lines Coverage Δ
src/module.cpp 85.70% <0.00%> (-1.02%) :arrow_down:
:rocket: New features to boost your workflow:
  • :snowflake: Test Analytics: Detect flaky tests, report on failures, and find test suite problems.

codecov[bot] avatar Feb 14 '25 18:02 codecov[bot]

/AzurePipelines run

jayhawk-commits avatar Mar 07 '25 22:03 jayhawk-commits

Azure Pipelines successfully started running 1 pipeline(s).

azure-pipelines[bot] avatar Mar 07 '25 22:03 azure-pipelines[bot]

Clang tidy thinks things are too complex @pfultz2

causten avatar Apr 16 '25 03:04 causten

Clang tidy thinks things are too complex @pfultz2

Yea I need to do some refactoring here.

pfultz2 avatar Apr 22 '25 17:04 pfultz2

Test Batch Rate new
a68707
Rate old
e90b37
Diff Compare
torchvision-resnet50 64 3,238.28 3,237.18 0.03% :white_check_mark:
torchvision-resnet50_fp16 64 6,907.17 6,900.67 0.09% :white_check_mark:
torchvision-densenet121 32 2,447.72 2,446.02 0.07% :white_check_mark:
torchvision-densenet121_fp16 32 4,215.57 4,218.17 -0.06% :white_check_mark:
torchvision-inceptionv3 32 1,618.99 1,618.27 0.04% :white_check_mark:
torchvision-inceptionv3_fp16 32 2,706.13 2,706.57 -0.02% :white_check_mark:
cadene-inceptionv4 16 756.23 756.25 -0.00% :white_check_mark:
cadene-resnext64x4 16 814.25 814.61 -0.04% :white_check_mark:
slim-mobilenet 64 7,433.22 7,432.36 0.01% :white_check_mark:
slim-nasnetalarge 64 216.88 216.81 0.03% :white_check_mark:
slim-resnet50v2 64 3,442.58 3,442.40 0.01% :white_check_mark:
bert-mrpc-onnx 8 1,143.33 1,144.28 -0.08% :white_check_mark:
bert-mrpc-tf 1 456.98 457.20 -0.05% :white_check_mark:
pytorch-examples-wlang-gru 1 492.50 495.12 -0.53% :white_check_mark:
pytorch-examples-wlang-lstm 1 450.21 458.39 -1.78% :white_check_mark:
torchvision-resnet50_1 1 820.61 811.71 1.10% :white_check_mark:
cadene-dpn92_1 1 427.26 429.64 -0.55% :white_check_mark:
cadene-resnext101_1 1 391.26 391.62 -0.09% :white_check_mark:
onnx-taau-downsample 1 395.09 394.28 0.20% :white_check_mark:
dlrm-criteoterabyte 1 32.21 32.20 0.03% :white_check_mark:
dlrm-criteoterabyte_fp16 1 51.11 51.26 -0.28% :white_check_mark:
agentmodel 1 10,475.77 10,217.13 2.53% :white_check_mark:
unet_fp16 2 59.31 59.41 -0.17% :white_check_mark:
resnet50v1_fp16 1 1,088.67 1,083.69 0.46% :white_check_mark:
resnet50v1_int8 1 1,051.41 1,063.44 -1.13% :white_check_mark:
bert_base_cased_fp16 64 1,162.83 1,163.13 -0.03% :white_check_mark:
bert_large_uncased_fp16 32 356.18 356.06 0.03% :white_check_mark:
bert_large_fp16 1 199.82 200.23 -0.20% :white_check_mark:
distilgpt2_fp16 16 2,226.65 2,225.52 0.05% :white_check_mark:
yolov5s 1 543.71 544.04 -0.06% :white_check_mark:
tinyllama 1 43.69 43.63 0.14% :white_check_mark:
vicuna-fastchat 1 44.78 43.88 2.03% :white_check_mark:
whisper-tiny-encoder 1 419.35 419.66 -0.08% :white_check_mark:
whisper-tiny-decoder 1 412.40 410.82 0.38% :white_check_mark:
llama2_7b 1 nan nan nan% :x:
qwen1.5-7b 1 23.43 23.43 0.03% :white_check_mark:
phi3-3.8b 1 nan nan nan% :x:
mask-rcnn 1 18.72 18.93 -1.14% :white_check_mark:
llama3-8b 1 21.61 21.64 -0.11% :white_check_mark:
whisper-large-encoder 1 10.17 10.18 -0.01% :white_check_mark:
whisper-large-decoder 1 98.98 98.99 -0.01% :white_check_mark:
mistral-7b 1 23.68 23.65 0.11% :white_check_mark:
FLUX.1-schnell 1 914.14 912.01 0.23% :white_check_mark:
nan nan nan nan nan% :x:

This build is not recommended to merge :red_circle:

migraphx-bot avatar Apr 22 '25 23:04 migraphx-bot


     :white_check_mark: bert-mrpc-onnx: PASSED: MIGraphX meets tolerance
:x:bert-mrpc-tf: ERROR - check error output2025-04-22 16:36:34.331801: I tensorflow/core/platform/cpu_feature_guard.cc:210] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: SSE3 SSE4.1 SSE4.2 AVX AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
WARNING: All log messages before absl::InitializeLog() is called are written to STDERR
I0000 00:00:1745357799.967252 163472 gpu_device.cc:2022] Created device /job:localhost/replica:0/task:0/device:GPU:0 with 62973 MB memory: -> device: 0, name: AMD Instinct MI250X/MI250, pci bus id: 0000:32:00.0
WARNING: All log messages before absl::InitializeLog() is called are written to STDERR
I0000 00:00:1745357800.890619 163472 mlir_graph_optimization_pass.cc:401] MLIR V1 optimization pass is not enabled
2025-04-22 16:36:50.983362: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
2025-04-22 16:36:50.983403: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
2025-04-22 16:36:50.983440: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
2025-04-22 16:36:50.983481: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
2025-04-22 16:36:50.983516: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
2025-04-22 16:36:50.983694: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
2025-04-22 16:36:50.983723: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
2025-04-22 16:36:50.983768: E external/local_xla/xla/service/gpu/llvm_gpu_backend/gpu_backend_lib.cc:250] bitcode module is required by this HLO module but was not found at ./opencl.bc
error: Failure when generating HSACO
error: Failure when generating HSACO
error: Failure when generating HSACO
error: Failure when generating HSACO
error: Failure when generating HSACO
error: Failure when generating HSACO
error: Failure when generating HSACO
error: Failure when generating HSACO
2025-04-22 16:36:50.984935: E tensorflow/compiler/mlir/tools/kernel_gen/tf_framework_c_interface.cc:228] INTERNAL: Generating device code failed.
2025-04-22 16:36:50.986148: W tensorflow/core/framework/op_kernel.cc:1829] UNKNOWN: JIT compilation failed.
2025-04-22 16:36:50.986166: I tensorflow/core/framework/local_rendezvous.cc:405] Local rendezvous is aborting with status: UNKNOWN: JIT compilation failed.
[[{{node import/bert/embeddings/LayerNorm/moments/SquaredDifference}}]]
2025-04-22 16:36:50.986177: I tensorflow/core/framework/local_rendezvous.cc:405] Local rendezvous is aborting with status: UNKNOWN: JIT compilation failed.
[[{{node import/bert/embeddings/LayerNorm/moments/SquaredDifference}}]]
[[import/loss/output/_21]]
2025-04-22 16:36:50.986191: I tensorflow/core/framework/local_rendezvous.cc:424] Local rendezvous recv item cancelled. Key hash: 11217777527359497193
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/client/session.py", line 1407, in _do_call
return fn(*args)
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/client/session.py", line 1390, in _run_fn
return self._call_tf_sessionrun(options, feed_dict, fetch_list,
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/client/session.py", line 1483, in _call_tf_sessionrun
return tf_session.TF_SessionRun_wrapper(self._session, options, feed_dict,
tensorflow.python.framework.errors_impl.UnknownError: 2 root error(s) found.
(0) UNKNOWN: JIT compilation failed.
[[{{node import/bert/embeddings/LayerNorm/moments/SquaredDifference}}]]
[[import/loss/output/_21]]
(1) UNKNOWN: JIT compilation failed.
[[{{node import/bert/embeddings/LayerNorm/moments/SquaredDifference}}]]
0 successful operations.
0 derived errors ignored.

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
File "/src/AMDMIGraphX/tools/accuracy/accuracy_checker.py", line 340, in
main()
File "/src/AMDMIGraphX/tools/accuracy/accuracy_checker.py", line 324, in main
y_out = sess.run(y, feed_dict=tf_dict)
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/client/session.py", line 977, in run
result = self._run(None, fetches, feed_dict, options_ptr,
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/client/session.py", line 1220, in _run
results = self._do_run(handle, final_targets, final_fetches,
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/client/session.py", line 1400, in _do_run
return self._do_call(_run_fn, feeds, fetches, targets, options,
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/client/session.py", line 1426, in _do_call
raise type(e)(node_def, op, message) # pylint: disable=no-value-for-parameter
tensorflow.python.framework.errors_impl.UnknownError: Graph execution error:

Detected at node 'import/bert/embeddings/LayerNorm/moments/SquaredDifference' defined at (most recent call last):
Node: 'import/bert/embeddings/LayerNorm/moments/SquaredDifference'
Detected at node 'import/bert/embeddings/LayerNorm/moments/SquaredDifference' defined at (most recent call last):
Node: 'import/bert/embeddings/LayerNorm/moments/SquaredDifference'
2 root error(s) found.
(0) UNKNOWN: JIT compilation failed.
[[{{node import/bert/embeddings/LayerNorm/moments/SquaredDifference}}]]
[[import/loss/output/_21]]
(1) UNKNOWN: JIT compilation failed.
[[{{node import/bert/embeddings/LayerNorm/moments/SquaredDifference}}]]
0 successful operations.
0 derived errors ignored.

Original stack trace for 'import/bert/embeddings/LayerNorm/moments/SquaredDifference':

     :white_check_mark: pytorch-examples-wlang-gru: PASSED: MIGraphX meets tolerance
     :white_check_mark: pytorch-examples-wlang-lstm: PASSED: MIGraphX meets tolerance
     :white_check_mark: dlrm-criteoterabyte: PASSED: MIGraphX meets tolerance
     :white_check_mark: agentmodel: PASSED: MIGraphX meets tolerance
:red_circle:unet: FAILED: MIGraphX is not within tolerance - check verbose output

     :white_check_mark: resnet50v1: PASSED: MIGraphX meets tolerance
     :white_check_mark: bert_base_cased_fp16: PASSED: MIGraphX meets tolerance
:red_circle:bert_large_uncased_fp16: FAILED: MIGraphX is not within tolerance - check verbose output

     :white_check_mark: bert_large: PASSED: MIGraphX meets tolerance
     :white_check_mark: yolov5s: PASSED: MIGraphX meets tolerance
     :white_check_mark: tinyllama: PASSED: MIGraphX meets tolerance
     :white_check_mark: vicuna-fastchat: PASSED: MIGraphX meets tolerance
     :white_check_mark: whisper-tiny-encoder: PASSED: MIGraphX meets tolerance
     :white_check_mark: whisper-tiny-decoder: PASSED: MIGraphX meets tolerance
     :white_check_mark: distilgpt2_fp16: PASSED: MIGraphX meets tolerance
:x:llama2_7b: ERROR - check error outputTraceback (most recent call last):
File "/src/AMDMIGraphX/tools/accuracy/accuracy_checker.py", line 340, in
main()
File "/src/AMDMIGraphX/tools/accuracy/accuracy_checker.py", line 205, in main
model = migraphx.parse_onnx(model_name, default_dim_value=batch)
RuntimeError: /src/AMDMIGraphX/src/onnx/onnx_parser.cpp:265: parse_from: PARSE_FROM: Failed reading onnx file: /new-saved-models/llama2_7b/decoder_model.onnx

:x:qwen1.5-7b: ERROR - check error outputusage: accuracy_checker.py [-h] [--onnx ONNX] [--tf TF] [--provider PROVIDER]
[--batch BATCH] [--fill1] [--fill0] [--fp16]
[--argmax] [--verbose] [--tolerance TOLERANCE]
[--input-dim INPUT_DIM] [--target TARGET]
[--ort-run] [--ort-logging]
[--disable-offload-copy] [--disable-fast-math]
[--exhaustive_tune]
accuracy_checker.py: error: unrecognized arguments: input_ids attention_mask position_ids 1 256 @attention_mask 1 256 @position_ids 1 256

:x:phi3-3.8b: ERROR - check error outputusage: accuracy_checker.py [-h] [--onnx ONNX] [--tf TF] [--provider PROVIDER]
[--batch BATCH] [--fill1] [--fill0] [--fp16]
[--argmax] [--verbose] [--tolerance TOLERANCE]
[--input-dim INPUT_DIM] [--target TARGET]
[--ort-run] [--ort-logging]
[--disable-offload-copy] [--disable-fast-math]
[--exhaustive_tune]
accuracy_checker.py: error: unrecognized arguments: input_ids attention_mask position_ids 1 256 @attention_mask 1 256 @position_ids 1 256

:red_circle:mask-rcnn: FAILED: MIGraphX is not within tolerance - check verbose output

     :white_check_mark: llama3-8b: PASSED: MIGraphX meets tolerance
:x:#whisper-large-encoder: ERROR - check error outputTraceback (most recent call last):
File "/src/AMDMIGraphX/tools/accuracy/accuracy_checker.py", line 340, in
main()
File "/src/AMDMIGraphX/tools/accuracy/accuracy_checker.py", line 205, in main
model = migraphx.parse_onnx(model_name, default_dim_value=batch)
RuntimeError: /src/AMDMIGraphX/src/include/migraphx/op/convolution.hpp:100: normalize_compute_shape: CONVOLUTION: mismatched channel numbers

     :white_check_mark: whisper-large-decoder: PASSED: MIGraphX meets tolerance
     :white_check_mark: mistral-7b: PASSED: MIGraphX meets tolerance
     :white_check_mark: FLUX.1-schnell: PASSED: MIGraphX meets tolerance

migraphx-bot avatar Apr 22 '25 23:04 migraphx-bot

Two reviewers on this, curious why this hasn't made it in. Just CI?

TedThemistokleous avatar Sep 15 '25 03:09 TedThemistokleous