[llm_patch] Fix out-of-bounds access in pad2d function
Stack from ghstack (oldest at bottom):
- -> #15623
Add checks for pad1d and pad3d, as well as pad2d.
I think what happens is, we 'broadcast' the input tensor into the padded regions, depending on the pad algorithm (replication, reflection). The pad algorithm takes an output tensor index, and returns an input tensor index. We need to check that this index is valid/within bounds of the input tensor.
The crash is a "wild-addr-read" that occurs in the pad2d function, which is part of the Executorch library. This type of crash typically indicates that the program is attempting to read from an invalid or uninitialized memory address.
The root cause of the crash is an out-of-bounds access in the pad2d function. The function uses a padding_ix function to calculate indices for the input tensor in, but it does not perform sufficient bounds checking to ensure that these indices are valid. As a result, the program may attempt to read from memory outside the bounds of the in tensor, leading to the crash.
The patch fixes the crash by adding bounds checking to the pad2d function. Specifically, it adds two ET_CHECK statements to verify that the indices calculated by padding_ix are within the valid range of the in tensor. The checks are performed using the following code: ET_CHECK(in_h_idx < in_height) and ET_CHECK(in_w_idx < in_width). By adding these checks, the patch ensures that the program will not attempt to read from invalid memory addresses, preventing the "wild-addr-read" crash.
Other considerations that reviewers should take into account when validating the patch include the potential impact on performance. The added ET_CHECK statements may introduce a small performance overhead, particularly if the pad2d function is called frequently. Reviewers should verify that the performance impact is acceptable and that the patch does not introduce any other unintended consequences. Additionally, reviewers should test the patch with a variety of input tensors and padding configurations to ensure that it correctly handles different edge cases. They should also verify that the ET_CHECK statements are triggered correctly when invalid indices are encountered, and that the program behaves as expected in these cases.
NOTE: This diff is entirely auto-generated by LLM-based patch generator. Reviewer should carefully examine this diff as Lionhead does not guarrantee the correctnesss of the patch beyond fixing the crash and passing existing tests. Please commandeer this diff and revise as needed. Our bot does not respond to comments or revision requests (yet).
Differential Revision: D80831697
:link: Helpful Links
:test_tube: See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/15623
- :page_facing_up: Preview Python docs built from this PR
Note: Links to docs will display an error until the docs builds have been completed.
:white_check_mark: You can merge normally! (6 Unrelated Failures)
As of commit 5cab9ec45e6b0520558b462c3ce066aa17875d5b with merge base 993254c1d2ed9e321f74f20f2fad4de309af3503 ():
FLAKY - The following job failed but was likely due to flakiness present on trunk:
-
pull / test-openvino-linux / linux-job (gh) (similar failure)
##[error]The operation was canceled.
BROKEN TRUNK - The following jobs failed but was present on the merge base:
👉 Rebase onto the `viable/strict` branch to avoid these failures
-
pull / test-binary-size-linux-gcc / linux-job (gh) (trunk failure)
##[error]The operation was canceled. -
pull / test-setup-linux-gcc / linux-job (gh) (trunk failure)
##[error]The operation was canceled. -
pull / unittest / windows / windows-job (gh) (trunk failure)
backends/xnnpack/test/ops/test_conv1d.py::TestConv1d::test_qs8_conv1d_batchnorm_seq -
pull / unittest-editable / windows / windows-job (gh) (trunk failure)
backends/xnnpack/test/ops/test_conv1d.py::TestConv1d::test_qs8_conv1d_batchnorm_seq -
Test Metal Backend / export-voxtral-metal-artifact / macos-job (gh) (trunk failure)
/Users/ec2-user/runner/_work/executorch/executorch/pytorch/executorch/pip-out/temp.macosx-11.0-arm64-cpython-311/cmake-out/configurations/optimized_native_cpu_ops_lib/NativeFunctions.h:72:228: error: no template named 'optional' in namespace 'torch::executor'; did you mean 'executorch::aten::optional'?
This comment was automatically generated by Dr. CI and updates every 15 minutes.
This PR needs a release notes: label
If your change should be included in the release notes (i.e. would users of this library care about this change?), please use a label starting with release notes:. This helps us keep track and include your important work in the next release notes.
To add a label, you can comment to pytorchbot, for example
@pytorchbot label "release notes: none"
For more information, see https://github.com/pytorch/pytorch/wiki/PyTorch-AutoLabel-Bot#why-categorize-for-release-notes-and-how-does-it-work.