|
12 | 12 | load_tests, TEST_NUMPY, TEST_SCIPY, IS_WINDOWS, gradcheck, coalescedonoff, \ |
13 | 13 | DeterministicGuard, first_sample, TEST_WITH_CROSSREF, TEST_WITH_ROCM, skipIfTorchDynamo, \ |
14 | 14 | parametrize, subtest, is_coalesced_indices, suppress_warnings, instantiate_parametrized_tests, \ |
15 | | - skipIfCrossRef |
| 15 | + skipIfCrossRef, slowTest |
16 | 16 | from torch.testing._internal.common_cuda import TEST_CUDA |
17 | 17 | from torch.testing._internal.common_mps import mps_ops_modifier |
18 | 18 | from numbers import Number |
@@ -4934,6 +4934,7 @@ def test_generate_simple_inputs(self): |
4934 | 4934 | f' contiguous_indices{contiguous_indices}, contiguous_values={contiguous_values}') |
4935 | 4935 | assert not untested_combinations, untested_combinations |
4936 | 4936 |
|
| 4937 | + @slowTest |
4937 | 4938 | @all_sparse_layouts('layout', include_strided=False) |
4938 | 4939 | def test_constructor_autograd(self, device, layout): |
4939 | 4940 |
|
@@ -5490,6 +5491,7 @@ def test_sparse_mask(self, mask_layout, device, dtype): |
5490 | 5491 | result = mask.to_dense().sparse_mask(mask) |
5491 | 5492 | self.assertEqual(result, mask) |
5492 | 5493 |
|
| 5494 | + @slowTest |
5493 | 5495 | @all_sparse_layouts('layout', include_strided=False) |
5494 | 5496 | @parametrize("masked", [subtest(False, name='nonmasked'), subtest(True, name='masked')]) |
5495 | 5497 | @parametrize("fast_mode", [subtest(False, name='slow'), subtest(True, name='fast')]) |
|
0 commit comments