@@ -1654,7 +1654,6 @@ def fn(x):
1654
1654
# make sure things also work if they aren't unrolled
1655
1655
self .common (fn , (torch .randn (8 , 3 ),))
1656
1656
1657
- @skipIfRocmArch (NAVI_ARCH )
1658
1657
def test_multilayer_sum_low_prec (self ):
1659
1658
# fp16 nyi for cpu
1660
1659
if self .device == "cpu" :
@@ -1665,7 +1664,6 @@ def fn(a):
1665
1664
1666
1665
self .common (fn , ((torch .rand ((10 , 3 , 352 , 352 ), dtype = torch .float16 ),)))
1667
1666
1668
- @skipIfRocmArch (NAVI_ARCH )
1669
1667
def test_multilayer_prime_size (self ):
1670
1668
def fn (a ):
1671
1669
return torch .max (a ), torch .sum (a )
@@ -1676,7 +1674,6 @@ def fn(a):
1676
1674
self .common (fn , (sample ,))
1677
1675
1678
1676
@skipCPUIf (IS_MACOS , "fails on macos" )
1679
- @skipIfRocmArch (NAVI_ARCH )
1680
1677
def test_multilayer_var (self ):
1681
1678
def fn (a ):
1682
1679
return torch .var (a )
@@ -2676,7 +2673,6 @@ def fn(a, b):
2676
2673
2677
2674
self .common (fn , (torch .randn (8 , 8 ), torch .randn (8 , 8 )))
2678
2675
2679
- @skipIfRocmArch (NAVI_ARCH )
2680
2676
def test_large_tensor_reduction (self ):
2681
2677
if not _has_sufficient_memory (self .device , 4.5 * 1024 ** 3 ): # 4.5 GiB
2682
2678
raise unittest .SkipTest ("insufficient memory" )
@@ -2697,7 +2693,6 @@ def fn(a):
2697
2693
expect = torch .tensor (2 , dtype = torch .int8 , device = self .device )
2698
2694
self .assertEqual (actual , expect )
2699
2695
2700
- @skipIfRocmArch (NAVI_ARCH )
2701
2696
def test_large_broadcast_reduction (self ):
2702
2697
if self .device == "cpu" :
2703
2698
raise unittest .SkipTest ("Fails on CPU" )
@@ -3772,7 +3767,6 @@ def test_conv2d_channels_last(self):
3772
3767
check_lowp = False ,
3773
3768
)
3774
3769
3775
- @skipIfRocmArch (NAVI_ARCH )
3776
3770
def test_conv2d_backward_channels_last (self ):
3777
3771
def fn (grad_output , inp , weight ):
3778
3772
convolution_backward_8 = torch .ops .aten .convolution_backward .default (
@@ -4532,7 +4526,6 @@ def fn(x, y):
4532
4526
self .assertEqual (a .stride (), c .stride ())
4533
4527
self .assertEqual (c .stride ()[2 ], 1 )
4534
4528
4535
- @skipIfRocmArch (NAVI_ARCH )
4536
4529
def test_std (self ):
4537
4530
def fn (x ):
4538
4531
return (
@@ -4575,7 +4568,6 @@ def test_batch_norm_2d(self):
4575
4568
4576
4569
# From yolov3
4577
4570
@with_tf32_off
4578
- @skipIfRocmArch (NAVI_ARCH )
4579
4571
def test_batch_norm_2d_2 (self ):
4580
4572
if self .device == "cpu" :
4581
4573
raise unittest .SkipTest (f"requires { GPU_TYPE } " )
@@ -4711,7 +4703,6 @@ def fn(x):
4711
4703
4712
4704
self .common (fn , (x ,))
4713
4705
4714
- @skipIfRocmArch (NAVI_ARCH )
4715
4706
def test_cauchy (self ):
4716
4707
def fn (x , y ):
4717
4708
return torch .sum (1 / (torch .unsqueeze (x , - 1 ) - y ))
@@ -6049,7 +6040,6 @@ def fn(a):
6049
6040
y = fn_compiled (x )
6050
6041
self .assertTrue (y is not x )
6051
6042
6052
- @skipIfRocmArch (NAVI_ARCH )
6053
6043
def test_l1_loss (self ):
6054
6044
def fn (a , b ):
6055
6045
return torch .nn .functional .l1_loss (a , b ), torch .nn .functional .mse_loss (a , b )
@@ -6447,7 +6437,6 @@ def fn(x):
6447
6437
fn , (torch .tensor ([1 , float ("inf" ), 2 , float ("-inf" ), float ("nan" )]),)
6448
6438
)
6449
6439
6450
- @skipIfRocmArch (NAVI_ARCH )
6451
6440
def test_any (self ):
6452
6441
def fn (x ):
6453
6442
return (
@@ -7196,7 +7185,6 @@ def fn(a, dim, index, b, reduce):
7196
7185
)
7197
7186
7198
7187
# issue #1150
7199
- @skipIfRocmArch (NAVI_ARCH )
7200
7188
def test_dense_mask_index (self ):
7201
7189
r"""
7202
7190
There will be a little difference for reduce order between aten and inductor
@@ -8172,7 +8160,6 @@ def fn(a, b):
8172
8160
b = torch .rand (2 , 2 , 1 , 4 , 1 ).int ()
8173
8161
self .common (fn , (a , b ))
8174
8162
8175
- @skipIfRocmArch (NAVI_ARCH )
8176
8163
def test_argmax_argmin1 (self ):
8177
8164
def fn (x ):
8178
8165
return (aten .argmax (x ), aten .argmin (x ))
@@ -8184,7 +8171,6 @@ def fn(x):
8184
8171
],
8185
8172
)
8186
8173
8187
- @skipIfRocmArch (NAVI_ARCH )
8188
8174
def test_argmax_argmin2 (self ):
8189
8175
def fn (x ):
8190
8176
return (
@@ -8196,7 +8182,6 @@ def fn(x):
8196
8182
8197
8183
self .common (fn , (torch .randn ([144 , 144 ]),))
8198
8184
8199
- @skipIfRocmArch (NAVI_ARCH )
8200
8185
def test_argmax_argmin_with_duplicates (self ):
8201
8186
def fn (x ):
8202
8187
return (
@@ -8218,7 +8203,6 @@ def fn(x):
8218
8203
t1 = torch .randint (8 , size = (1028 , 1028 ))
8219
8204
self .common (fn , (t1 ,))
8220
8205
8221
- @skipIfRocmArch (NAVI_ARCH )
8222
8206
def test_argmax_argmin_with_nan (self ):
8223
8207
def fn (x ):
8224
8208
return (
@@ -8351,7 +8335,6 @@ def fn(x):
8351
8335
],
8352
8336
)
8353
8337
8354
- @skipIfRocmArch (NAVI_ARCH )
8355
8338
def test_tmp_not_defined_issue1 (self ):
8356
8339
def forward (
8357
8340
primals_3 ,
@@ -8746,7 +8729,6 @@ def __torch_dispatch__(self, func, types, args=(), kwargs=None):
8746
8729
else :
8747
8730
self .assertEqual (len (inps ), 0 )
8748
8731
8749
- @skipIfRocmArch (NAVI_ARCH )
8750
8732
def test_dtype_mismatch_issue (self ):
8751
8733
def fn (x ):
8752
8734
attn = torch .nn .functional .pad (x , [0 , 1 ])
@@ -11414,7 +11396,6 @@ def test_rnn_compile_safe(self):
11414
11396
11415
11397
class NanCheckerTest (TestCase ):
11416
11398
@config .patch ("nan_asserts" , True )
11417
- @skipIfRocmArch (NAVI_ARCH )
11418
11399
def test_nan_checker_pass (self ):
11419
11400
def f (x ):
11420
11401
return torch .softmax (x , dim = - 1 )
0 commit comments