@@ -1406,17 +1406,17 @@ def f(x):
1406
1406
"""\
1407
1407
def forward(self, arg0_1: "Sym(s0)", arg1_1: "f32[s0][1]cpu"):
1408
1408
clone: "f32[s0][1]cpu" = torch.ops.aten.clone.default(arg1_1)
1409
- nonzero: "i64[u0, 1][1, 1 ]cpu" = torch.ops.aten.nonzero.default(clone); clone = None
1409
+ nonzero: "i64[u0, 1][1, u0 ]cpu" = torch.ops.aten.nonzero.default(clone); clone = None
1410
1410
sym_size_int_1: "Sym(u0)" = torch.ops.aten.sym_size.int(nonzero, 0)
1411
1411
ge_1: "Sym(u0 >= 0)" = sym_size_int_1 >= 0; sym_size_int_1 = None
1412
1412
_assert_scalar = torch.ops.aten._assert_scalar.default(ge_1, "Runtime assertion failed for expression u0 >= 0 on node 'ge'"); ge_1 = _assert_scalar = None
1413
- _to_copy: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten._to_copy.default(nonzero, dtype = torch.float32); nonzero = None
1413
+ _to_copy: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten._to_copy.default(nonzero, dtype = torch.float32); nonzero = None
1414
1414
auto_functionalized_v2 = torch.ops.higher_order.auto_functionalized_v2(torch.ops.mylib.foo.default, _x_base_index = 0, _x_alias = True, _y_base_index = 1, _y_alias = True, _all_bases = [arg1_1, _to_copy]); _to_copy = None
1415
1415
getitem_1: "f32[s0][1]cpu" = auto_functionalized_v2[1]
1416
- getitem_2: "f32[u0, 1][1, 1 ]cpu" = auto_functionalized_v2[2]; auto_functionalized_v2 = None
1416
+ getitem_2: "f32[u0, 1][1, u0 ]cpu" = auto_functionalized_v2[2]; auto_functionalized_v2 = None
1417
1417
copy_: "f32[s0][1]cpu" = torch.ops.aten.copy_.default(arg1_1, getitem_1); arg1_1 = copy_ = None
1418
1418
alias_1: "f32[s0][1]cpu" = torch.ops.aten.alias.default(getitem_1); getitem_1 = None
1419
- slice_2: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten.slice.Tensor(getitem_2); getitem_2 = None
1419
+ slice_2: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten.slice.Tensor(getitem_2); getitem_2 = None
1420
1420
return (alias_1, slice_2)""" , # noqa: B950
1421
1421
ignore_comments = True ,
1422
1422
ignore_empty_lines = True ,
@@ -1427,19 +1427,19 @@ def forward(self, arg0_1: "Sym(s0)", arg1_1: "f32[s0][1]cpu"):
1427
1427
"""\
1428
1428
def forward(self, arg0_1: "f32[2][1]cpu"):
1429
1429
clone: "f32[2][1]cpu" = torch.ops.aten.clone.default(arg0_1)
1430
- nonzero: "i64[u0, 1][1, 1 ]cpu" = torch.ops.aten.nonzero.default(clone); clone = None
1430
+ nonzero: "i64[u0, 1][1, u0 ]cpu" = torch.ops.aten.nonzero.default(clone); clone = None
1431
1431
sym_size_int: "Sym(u0)" = torch.ops.aten.sym_size.int(nonzero, 0)
1432
1432
ge_1: "Sym(u0 >= 0)" = sym_size_int >= 0
1433
1433
_assert_scalar = torch.ops.aten._assert_scalar.default(ge_1, "Runtime assertion failed for expression u0 >= 0 on node 'ge'"); ge_1 = _assert_scalar = None
1434
1434
le: "Sym(u0 <= 2)" = sym_size_int <= 2; sym_size_int = None
1435
1435
_assert_scalar_1 = torch.ops.aten._assert_scalar.default(le, "Runtime assertion failed for expression u0 <= 2 on node 'le'"); le = _assert_scalar_1 = None
1436
- _to_copy: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten._to_copy.default(nonzero, dtype = torch.float32); nonzero = None
1436
+ _to_copy: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten._to_copy.default(nonzero, dtype = torch.float32); nonzero = None
1437
1437
auto_functionalized_v2 = torch.ops.higher_order.auto_functionalized_v2(torch.ops.mylib.foo.default, _x_base_index = 0, _x_alias = True, _y_base_index = 1, _y_alias = True, _all_bases = [arg0_1, _to_copy]); _to_copy = None
1438
1438
getitem_1: "f32[2][1]cpu" = auto_functionalized_v2[1]
1439
- getitem_2: "f32[u0, 1][1, 1 ]cpu" = auto_functionalized_v2[2]; auto_functionalized_v2 = None
1439
+ getitem_2: "f32[u0, 1][1, u0 ]cpu" = auto_functionalized_v2[2]; auto_functionalized_v2 = None
1440
1440
copy_: "f32[2][1]cpu" = torch.ops.aten.copy_.default(arg0_1, getitem_1); arg0_1 = copy_ = None
1441
1441
alias_1: "f32[2][1]cpu" = torch.ops.aten.alias.default(getitem_1); getitem_1 = None
1442
- slice_2: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten.slice.Tensor(getitem_2); getitem_2 = None
1442
+ slice_2: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten.slice.Tensor(getitem_2); getitem_2 = None
1443
1443
return (alias_1, slice_2)""" , # noqa: B950
1444
1444
ignore_comments = True ,
1445
1445
ignore_empty_lines = True ,
@@ -1452,16 +1452,16 @@ def forward(self, arg0_1: "f32[2][1]cpu"):
1452
1452
graph_inductor ,
1453
1453
"""\
1454
1454
def forward(self, arg0_1: "Sym(s0)", arg1_1: "f32[s0][1]cpu"):
1455
- nonzero: "i64[u0, 1][1, 1 ]cpu" = torch.ops.aten.nonzero.default(arg1_1)
1455
+ nonzero: "i64[u0, 1][1, u0 ]cpu" = torch.ops.aten.nonzero.default(arg1_1)
1456
1456
sym_size_int_1: "Sym(u0)" = torch.ops.aten.sym_size.int(nonzero, 0)
1457
1457
ge_1: "Sym(u0 >= 0)" = sym_size_int_1 >= 0; sym_size_int_1 = None
1458
1458
_assert_scalar = torch.ops.aten._assert_scalar.default(ge_1, "Runtime assertion failed for expression u0 >= 0 on node 'ge'"); ge_1 = _assert_scalar = None
1459
- convert_element_type: "f32[u0, 1][1, 1 ]cpu" = torch.ops.prims.convert_element_type.default(nonzero, torch.float32); nonzero = None
1459
+ convert_element_type: "f32[u0, 1][1, u0 ]cpu" = torch.ops.prims.convert_element_type.default(nonzero, torch.float32); nonzero = None
1460
1460
alias_default: "f32[s0][1]cpu" = torch.ops.aten.alias.default(arg1_1)
1461
- alias_default_1: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten.alias.default(convert_element_type)
1461
+ alias_default_1: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten.alias.default(convert_element_type)
1462
1462
foo_default = torch.ops.mylib.foo.default(alias_default, alias_default_1); alias_default = alias_default_1 = foo_default = None
1463
1463
copy_: "f32[s0][1]cpu" = torch.ops.aten.copy_.default(arg1_1, arg1_1); copy_ = None
1464
- slice_2: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten.slice.Tensor(convert_element_type); convert_element_type = None
1464
+ slice_2: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten.slice.Tensor(convert_element_type); convert_element_type = None
1465
1465
return (arg1_1, slice_2)""" , # noqa: B950
1466
1466
ignore_comments = True ,
1467
1467
ignore_empty_lines = True ,
@@ -1471,18 +1471,18 @@ def forward(self, arg0_1: "Sym(s0)", arg1_1: "f32[s0][1]cpu"):
1471
1471
graph_inductor ,
1472
1472
"""\
1473
1473
def forward(self, arg0_1: "f32[2][1]cpu"):
1474
- nonzero: "i64[u0, 1][1, 1 ]cpu" = torch.ops.aten.nonzero.default(arg0_1)
1474
+ nonzero: "i64[u0, 1][1, u0 ]cpu" = torch.ops.aten.nonzero.default(arg0_1)
1475
1475
sym_size_int: "Sym(u0)" = torch.ops.aten.sym_size.int(nonzero, 0)
1476
1476
ge_1: "Sym(u0 >= 0)" = sym_size_int >= 0
1477
1477
_assert_scalar = torch.ops.aten._assert_scalar.default(ge_1, "Runtime assertion failed for expression u0 >= 0 on node 'ge'"); ge_1 = _assert_scalar = None
1478
1478
le: "Sym(u0 <= 2)" = sym_size_int <= 2; sym_size_int = None
1479
1479
_assert_scalar_1 = torch.ops.aten._assert_scalar.default(le, "Runtime assertion failed for expression u0 <= 2 on node 'le'"); le = _assert_scalar_1 = None
1480
- convert_element_type: "f32[u0, 1][1, 1 ]cpu" = torch.ops.prims.convert_element_type.default(nonzero, torch.float32); nonzero = None
1480
+ convert_element_type: "f32[u0, 1][1, u0 ]cpu" = torch.ops.prims.convert_element_type.default(nonzero, torch.float32); nonzero = None
1481
1481
alias_default: "f32[2][1]cpu" = torch.ops.aten.alias.default(arg0_1)
1482
- alias_default_1: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten.alias.default(convert_element_type)
1482
+ alias_default_1: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten.alias.default(convert_element_type)
1483
1483
foo_default = torch.ops.mylib.foo.default(alias_default, alias_default_1); alias_default = alias_default_1 = foo_default = None
1484
1484
copy_: "f32[2][1]cpu" = torch.ops.aten.copy_.default(arg0_1, arg0_1); copy_ = None
1485
- slice_2: "f32[u0, 1][1, 1 ]cpu" = torch.ops.aten.slice.Tensor(convert_element_type); convert_element_type = None
1485
+ slice_2: "f32[u0, 1][1, u0 ]cpu" = torch.ops.aten.slice.Tensor(convert_element_type); convert_element_type = None
1486
1486
return (arg0_1, slice_2)""" , # noqa: B950
1487
1487
ignore_comments = True ,
1488
1488
ignore_empty_lines = True ,
0 commit comments