Skip to content

Commit c5d30d6

Browse files
committed
skip some tests
1 parent 39ac3a0 commit c5d30d6

5 files changed

Lines changed: 13 additions & 1 deletion

File tree

test/distributed/test_cupy_as_tensor.py

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,10 @@
88
import torch
99
from torch.multiprocessing.reductions import reduce_tensor
1010
from torch.testing._internal.common_cuda import SM100OrLater
11-
from torch.testing._internal.common_distributed import MultiProcContinuousTest
11+
from torch.testing._internal.common_distributed import (
12+
MultiProcContinuousTest,
13+
skip_if_rocm_multiprocess,
14+
)
1215
from torch.testing._internal.common_utils import (
1316
requires_cuda_p2p_access,
1417
run_tests,
@@ -67,6 +70,7 @@ def _init_device(self) -> None:
6770
def device(self) -> torch.device:
6871
return torch.device(device_type, self.rank)
6972

73+
@skip_if_rocm_multiprocess # RuntimeError: pidfd_getfd Operation not permitted"
7074
@skip_but_pass_in_sandcastle_if(
7175
SM100OrLater,
7276
"Fails if ran in docker environment without privileged access (https://github.com/pytorch/pytorch/issues/165170)",

test/test_cuda.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -5060,6 +5060,8 @@ def cb(device, alloc, device_alloc, device_free):
50605060

50615061
def test_allocator_fuzz(self):
50625062
# fuzz
5063+
if torch.version.hip and "expandable_segments:True" in torch._C._accelerator_getAllocatorSettings():
5064+
raise unittest.SkipTest("ROCm needs https://github.com/ROCm/rocm-systems/pull/3023")
50635065
state = random.getstate()
50645066
random.seed(123)
50655067
N = 10000

torch/_C/__init__.pyi.in

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2608,6 +2608,7 @@ def _accelerator_getDeviceStats(device_index: _int) -> dict[str, Any]: ...
26082608
def _accelerator_resetAccumulatedStats(device_index: _int) -> None: ...
26092609
def _accelerator_resetPeakStats(device_index: _int) -> None: ...
26102610
def _accelerator_getMemoryInfo(device_index: _int) -> tuple[_int, _int]: ...
2611+
def _accelerator_getAllocatorSettings() -> str: ...
26112612
def _accelerator_setAllocatorSettings(env: str) -> None: ...
26122613

26132614
# Defined in torch/csrc/jit/python/python_tracer.cpp

torch/_dynamo/trace_rules.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -468,6 +468,7 @@
468468
"torch._C._accelerator_getAccelerator",
469469
"torch._C._accelerator_getDeviceIndex",
470470
"torch._C._accelerator_getStream",
471+
"torch._C._accelerator_getAllocatorSettings",
471472
"torch._C._accelerator_setAllocatorSettings",
472473
"torch._C._accelerator_setStream",
473474
"torch._C._accelerator_synchronizeDevice",

torch/csrc/DeviceAccelerator.cpp

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -164,6 +164,10 @@ void initModule(PyObject* module) {
164164
return at::accelerator::getMemoryInfo(device_index);
165165
});
166166

167+
m.def("_accelerator_getAllocatorSettings", []() {
168+
return c10::CachingAllocator::getAllocatorSettings();
169+
});
170+
167171
m.def("_accelerator_setAllocatorSettings", [](std::string env) {
168172
c10::CachingAllocator::setAllocatorSettings(env);
169173
});

0 commit comments

Comments
 (0)