Skip to content

Commit 1641b00

Browse files
committed
Unskip Navi inductor UTs
(cherry picked from commit 7c3b97c)
1 parent 12b4a67 commit 1641b00

File tree

1 file changed

+0
-34
lines changed

1 file changed

+0
-34
lines changed

test/inductor/test_torchinductor.py

Lines changed: 0 additions & 34 deletions
Original file line numberDiff line numberDiff line change
@@ -1804,7 +1804,6 @@ def fn(x):
18041804
# make sure things also work if they aren't unrolled
18051805
self.common(fn, (torch.randn(8, 3),))
18061806

1807-
@skipIfRocmArch(NAVI_ARCH)
18081807
def test_multilayer_sum_low_prec(self):
18091808
# fp16 nyi for cpu
18101809
if self.device == "cpu":
@@ -1815,7 +1814,6 @@ def fn(a):
18151814

18161815
self.common(fn, ((torch.rand((10, 3, 352, 352), dtype=torch.float16),)))
18171816

1818-
@skipIfRocmArch(NAVI_ARCH)
18191817
def test_multilayer_prime_size(self):
18201818
def fn(a):
18211819
return torch.max(a), torch.sum(a)
@@ -1827,7 +1825,6 @@ def fn(a):
18271825

18281826
@skip_if_gpu_halide
18291827
@skipCPUIf(IS_MACOS, "fails on macos")
1830-
@skipIfRocmArch(NAVI_ARCH)
18311828
def test_multilayer_var(self):
18321829
def fn(a):
18331830
return torch.var(a)
@@ -2939,8 +2936,6 @@ def fn(a, b):
29392936

29402937
self.common(fn, (torch.randn(8, 8), torch.randn(8, 8)))
29412938

2942-
@skip_if_halide # only 32-bit indexing
2943-
@skipIfRocmArch(NAVI_ARCH)
29442939
def test_large_tensor_reduction(self):
29452940
if not _has_sufficient_memory(self.device, 4.5 * 1024**3): # 4.5 GiB
29462941
raise unittest.SkipTest("insufficient memory")
@@ -2961,8 +2956,6 @@ def fn(a):
29612956
expect = torch.tensor(2, dtype=torch.int8, device=self.device)
29622957
self.assertEqual(actual, expect)
29632958

2964-
@skip_if_gpu_halide # only 32-bit indexing
2965-
@skipIfRocmArch(NAVI_ARCH)
29662959
def test_large_broadcast_reduction(self):
29672960
if self.device == "cpu":
29682961
raise unittest.SkipTest("Fails on CPU")
@@ -4113,7 +4106,6 @@ def test_conv2d_channels_last(self):
41134106
check_lowp=False,
41144107
)
41154108

4116-
@skipIfRocmArch(NAVI_ARCH)
41174109
def test_conv2d_backward_channels_last(self):
41184110
def fn(grad_output, inp, weight):
41194111
convolution_backward_8 = torch.ops.aten.convolution_backward.default(
@@ -4898,8 +4890,6 @@ def fn(x, y):
48984890
self.assertEqual(a.stride(), c.stride())
48994891
self.assertEqual(c.stride()[2], 1)
49004892

4901-
@skip_if_gpu_halide
4902-
@skipIfRocmArch(NAVI_ARCH)
49034893
def test_std(self):
49044894
def fn(x):
49054895
return (
@@ -4942,7 +4932,6 @@ def test_batch_norm_2d(self):
49424932

49434933
# From yolov3
49444934
@with_tf32_off
4945-
@skipIfRocmArch(NAVI_ARCH)
49464935
def test_batch_norm_2d_2(self):
49474936
if self.device == "cpu":
49484937
raise unittest.SkipTest(f"requires {GPU_TYPE}")
@@ -5079,18 +5068,6 @@ def fn(x):
50795068

50805069
self.common(fn, (x,))
50815070

5082-
def test_polar(self):
5083-
def fn(dist, angle):
5084-
return torch.polar(dist, angle)
5085-
5086-
inp = (
5087-
torch.tensor([1, 2], dtype=torch.float64),
5088-
torch.tensor([np.pi / 2, 5 * np.pi / 4], dtype=torch.float64),
5089-
)
5090-
self.common(fn, (*inp,))
5091-
5092-
@skip_if_gpu_halide # incorrect result on CUDA
5093-
@skipIfRocmArch(NAVI_ARCH)
50945071
def test_cauchy(self):
50955072
def fn(x, y):
50965073
return torch.sum(1 / (torch.unsqueeze(x, -1) - y))
@@ -6491,7 +6468,6 @@ def fn(a):
64916468
y = fn_compiled(x)
64926469
self.assertTrue(y is not x)
64936470

6494-
@skipIfRocmArch(NAVI_ARCH)
64956471
def test_l1_loss(self):
64966472
def fn(a, b):
64976473
return torch.nn.functional.l1_loss(a, b), torch.nn.functional.mse_loss(a, b)
@@ -6899,7 +6875,6 @@ def fn(x):
68996875
fn, (torch.tensor([1, float("inf"), 2, float("-inf"), float("nan")]),)
69006876
)
69016877

6902-
@skipIfRocmArch(NAVI_ARCH)
69036878
def test_any(self):
69046879
def fn(x):
69056880
return (
@@ -7652,7 +7627,6 @@ def fn(a, dim, index, b, reduce):
76527627

76537628
@skip_if_gpu_halide
76547629
# issue #1150
7655-
@skipIfRocmArch(NAVI_ARCH)
76567630
def test_dense_mask_index(self):
76577631
r"""
76587632
There will be a little difference for reduce order between aten and inductor
@@ -8662,7 +8636,6 @@ def fn(a, b):
86628636
b = torch.rand(2, 2, 1, 4, 1).int()
86638637
self.common(fn, (a, b))
86648638

8665-
@skipIfRocmArch(NAVI_ARCH)
86668639
def test_argmax_argmin1(self):
86678640
def fn(x):
86688641
return (aten.argmax(x), aten.argmin(x))
@@ -8674,7 +8647,6 @@ def fn(x):
86748647
],
86758648
)
86768649

8677-
@skipIfRocmArch(NAVI_ARCH)
86788650
def test_argmax_argmin2(self):
86798651
def fn(x):
86808652
return (
@@ -8686,7 +8658,6 @@ def fn(x):
86868658

86878659
self.common(fn, (torch.randn([144, 144]),))
86888660

8689-
@skipIfRocmArch(NAVI_ARCH)
86908661
def test_argmax_argmin_with_duplicates(self):
86918662
def fn(x):
86928663
return (
@@ -8708,8 +8679,6 @@ def fn(x):
87088679
t1 = torch.randint(8, size=(1028, 1028))
87098680
self.common(fn, (t1,))
87108681

8711-
@skip_if_halide # nan behavior
8712-
@skipIfRocmArch(NAVI_ARCH)
87138682
def test_argmax_argmin_with_nan(self):
87148683
def fn(x):
87158684
return (
@@ -8833,7 +8802,6 @@ def fn(x):
88338802
],
88348803
)
88358804

8836-
@skipIfRocmArch(NAVI_ARCH)
88378805
def test_tmp_not_defined_issue1(self):
88388806
def forward(
88398807
primals_3,
@@ -9234,7 +9202,6 @@ def __torch_dispatch__(self, func, types, args=(), kwargs=None):
92349202
else:
92359203
self.assertEqual(len(inps), 0)
92369204

9237-
@skipIfRocmArch(NAVI_ARCH)
92389205
def test_dtype_mismatch_issue(self):
92399206
def fn(x):
92409207
attn = torch.nn.functional.pad(x, [0, 1])
@@ -12235,7 +12202,6 @@ def test_rnn_compile_safe(self):
1223512202

1223612203
class NanCheckerTest(TestCase):
1223712204
@config.patch("nan_asserts", True)
12238-
@skipIfRocmArch(NAVI_ARCH)
1223912205
def test_nan_checker_pass(self):
1224012206
def f(x):
1224112207
return torch.softmax(x, dim=-1)

0 commit comments

Comments
 (0)