Skip to content

[AUTOGENERATED] release/2.11_IFU_20260224#3000

Open
pragupta wants to merge 5 commits intorelease/2.11from
release/2.11_IFU_20260224
Open

[AUTOGENERATED] release/2.11_IFU_20260224#3000
pragupta wants to merge 5 commits intorelease/2.11from
release/2.11_IFU_20260224

Conversation

@pragupta
Copy link
Collaborator

rocm_base: 5decbe0

@rocm-repo-management-api
Copy link

rocm-repo-management-api bot commented Feb 24, 2026

Jenkins build for 9455a8e1be8f1f0656bfb16d463691767794d389 commit finished as FAILURE
Links: Pipeline Overview / Build artifacts / Test Results

jithunnair-amd and others added 2 commits February 26, 2026 22:34
…for py3.9;

upgrade tensorboard compatible with numpy 2

Co-authored-by: Ethan Wee <Ethan.Wee@amd.com>
(cherry picked from commit e867a3d)
(cherry picked from commit c7a1e32)
(cherry picked from commit 2a215e4)
(cherry picked from commit 866cc1d)
(cherry picked from commit 4b46310)
This PR fixes the unit test,

test/test_cuda.py::TestCuda::test_set_per_process_memory_fraction FAILED
[0.1163s]

```
Traceback (most recent call last):
  File "/var/lib/jenkins/pytorch/test/test_cuda.py", line 471, in test_set_per_process_memory_fraction
    tmp_tensor = torch.empty(application, dtype=torch.int8, device="cuda")
RuntimeError: Trying to create tensor with negative dimension -5681285432: [-5681285432]
```
This error occurs only on gfx1101 arch.

This error is coming from an integer overflow when another unit test,
test/test_cuda.py::TestCuda::test_randint_generation_for_large_numel
creates a tensor with a huge numel, which overflows into a higher
torch.cuda.max_memory_reserved() when you call
test/test_cuda.py::TestCuda::test_set_per_process_memory_fraction
afterward. To avoid this we introduced torch.cuda.empty_cache() and
torch.cuda.reset_peak_memory_stats() to clean up CUDA states.

JIRA: https://ontrack-internal.amd.com/browse/SWDEV-535295
(cherry picked from commit f86d184)
(cherry picked from commit 1b44228)
@jithunnair-amd jithunnair-amd force-pushed the release/2.11_IFU_20260224 branch from 9455a8e to f652e51 Compare February 26, 2026 22:34
@rocm-repo-management-api
Copy link

rocm-repo-management-api bot commented Feb 26, 2026

Jenkins build for f652e51cb476cb3a1393a48928ab9d318a741979 commit finished as FAILURE
Links: Pipeline Overview / Build artifacts / Test Results

jeffdaily and others added 3 commits February 27, 2026 06:24
…d_memory_with_allocator (#2811)

Use try/finally block. This follows a similar pattern elsewhere in
test_cuda.py.

Fixes #ROCm/TheRock#2118.
…ersistent reduction and no_x_dim removal (#2454)

Cherry-pick of #2417
Need to resolve conflicts

---------

Co-authored-by: Jack Taylor <108682042+jataylo@users.noreply.github.com>
(cherry picked from commit eb47158)

[release/2.9][ROCm][inductor] Add ROCm specific persistent reduction config. (#2861)

In support of
[SWDEV-566103](https://ontrack-internal.amd.com/browse/SWDEV-566103)

[release/2.10] Fix Inductor Triton Heuristics (#2931)

The ROCm release/2.10 branch was created by applying 15 commits to
upstream release/2.10 branch.
(See
pytorch/pytorch@release/2.10...ROCm:pytorch:release/2.10)

This PR fixes the issue with the missing disable_pointwise_autotuning
function.

There are three commits in this PR:

First commit is a revert:
1c96f23 - Autotuning support for
persistent reduction

since it is already available in upstream release/2.10 and is not
needed. (It reintroduced disable_pointwise_autotuning function.)

The second commit (b9facd0) is needed
for provenance, so I can apply the third commit:
e5eee74 - Heuristics improvements for
reduction kernels

which was reverted last minute before the release/2.10 cutoff and then
re-landed shortly afterwards the cutoff date but with a minor change.

---------

Co-authored-by: Pandya, Vivek Vasudevbhai <vpandya@qti.qualcomm.com>
@jithunnair-amd jithunnair-amd force-pushed the release/2.11_IFU_20260224 branch from f652e51 to 5ea4c12 Compare February 27, 2026 06:27
@rocm-repo-management-api
Copy link

rocm-repo-management-api bot commented Feb 27, 2026

Jenkins build for 5ea4c12f76e8deb00fa19583c5e35129bb0f1dcf commit finished as FAILURE
Links: Pipeline Overview / Build artifacts / Test Results

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

6 participants