[AUTOGENERATED] release/2.11_IFU_20260224#3000
Open
pragupta wants to merge 5 commits intorelease/2.11from
Open
[AUTOGENERATED] release/2.11_IFU_20260224#3000pragupta wants to merge 5 commits intorelease/2.11from
pragupta wants to merge 5 commits intorelease/2.11from
Conversation
|
Jenkins build for 9455a8e1be8f1f0656bfb16d463691767794d389 commit finished as FAILURE |
This PR fixes the unit test,
test/test_cuda.py::TestCuda::test_set_per_process_memory_fraction FAILED
[0.1163s]
```
Traceback (most recent call last):
File "/var/lib/jenkins/pytorch/test/test_cuda.py", line 471, in test_set_per_process_memory_fraction
tmp_tensor = torch.empty(application, dtype=torch.int8, device="cuda")
RuntimeError: Trying to create tensor with negative dimension -5681285432: [-5681285432]
```
This error occurs only on gfx1101 arch.
This error is coming from an integer overflow when another unit test,
test/test_cuda.py::TestCuda::test_randint_generation_for_large_numel
creates a tensor with a huge numel, which overflows into a higher
torch.cuda.max_memory_reserved() when you call
test/test_cuda.py::TestCuda::test_set_per_process_memory_fraction
afterward. To avoid this we introduced torch.cuda.empty_cache() and
torch.cuda.reset_peak_memory_stats() to clean up CUDA states.
JIRA: https://ontrack-internal.amd.com/browse/SWDEV-535295
(cherry picked from commit f86d184)
(cherry picked from commit 1b44228)
9455a8e to
f652e51
Compare
|
Jenkins build for f652e51cb476cb3a1393a48928ab9d318a741979 commit finished as FAILURE |
…d_memory_with_allocator (#2811) Use try/finally block. This follows a similar pattern elsewhere in test_cuda.py. Fixes #ROCm/TheRock#2118.
…ersistent reduction and no_x_dim removal (#2454) Cherry-pick of #2417 Need to resolve conflicts --------- Co-authored-by: Jack Taylor <108682042+jataylo@users.noreply.github.com> (cherry picked from commit eb47158) [release/2.9][ROCm][inductor] Add ROCm specific persistent reduction config. (#2861) In support of [SWDEV-566103](https://ontrack-internal.amd.com/browse/SWDEV-566103) [release/2.10] Fix Inductor Triton Heuristics (#2931) The ROCm release/2.10 branch was created by applying 15 commits to upstream release/2.10 branch. (See pytorch/pytorch@release/2.10...ROCm:pytorch:release/2.10) This PR fixes the issue with the missing disable_pointwise_autotuning function. There are three commits in this PR: First commit is a revert: 1c96f23 - Autotuning support for persistent reduction since it is already available in upstream release/2.10 and is not needed. (It reintroduced disable_pointwise_autotuning function.) The second commit (b9facd0) is needed for provenance, so I can apply the third commit: e5eee74 - Heuristics improvements for reduction kernels which was reverted last minute before the release/2.10 cutoff and then re-landed shortly afterwards the cutoff date but with a minor change. --------- Co-authored-by: Pandya, Vivek Vasudevbhai <vpandya@qti.qualcomm.com>
f652e51 to
5ea4c12
Compare
|
Jenkins build for 5ea4c12f76e8deb00fa19583c5e35129bb0f1dcf commit finished as FAILURE |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
rocm_base: 5decbe0