@@ -408,14 +408,14 @@ define amdgpu_cs i32 @branch_divergent_simulated_negated_ballot_ne_zero_and(i32
408
408
; CHECK-NEXT: v_cmp_gt_u32_e32 vcc, 12, v0
409
409
; CHECK-NEXT: v_cmp_lt_u32_e64 s[0:1], 34, v1
410
410
; CHECK-NEXT: s_and_b64 vcc, vcc, s[0:1]
411
- ; CHECK-NEXT: s_cbranch_vccnz .LBB0_2
411
+ ; CHECK-NEXT: s_cbranch_vccnz .LBB20_2
412
412
; CHECK-NEXT: ; %bb.1: ; %true
413
413
; CHECK-NEXT: s_mov_b32 s0, 42
414
- ; CHECK-NEXT: s_branch .LBB0_3
415
- ; CHECK-NEXT: .LBB0_2 : ; %false
414
+ ; CHECK-NEXT: s_branch .LBB20_3
415
+ ; CHECK-NEXT: .LBB20_2 : ; %false
416
416
; CHECK-NEXT: s_mov_b32 s0, 33
417
- ; CHECK-NEXT: s_branch .LBB0_3
418
- ; CHECK-NEXT: .LBB0_3 :
417
+ ; CHECK-NEXT: s_branch .LBB20_3
418
+ ; CHECK-NEXT: .LBB20_3 :
419
419
%v1c = icmp ult i32 %v1 , 12
420
420
%v2c = icmp ugt i32 %v2 , 34
421
421
%c = and i1 %v1c , %v2c
@@ -437,14 +437,14 @@ define amdgpu_cs i32 @branch_uniform_simulated_negated_ballot_ne_zero_and(i32 in
437
437
; CHECK-NEXT: s_cselect_b64 s[0:1], -1, 0
438
438
; CHECK-NEXT: s_and_b64 s[0:1], s[2:3], s[0:1]
439
439
; CHECK-NEXT: s_and_b64 s[0:1], s[0:1], exec
440
- ; CHECK-NEXT: s_cbranch_scc1 .LBB1_2
440
+ ; CHECK-NEXT: s_cbranch_scc1 .LBB21_2
441
441
; CHECK-NEXT: ; %bb.1: ; %true
442
442
; CHECK-NEXT: s_mov_b32 s0, 42
443
- ; CHECK-NEXT: s_branch .LBB1_3
444
- ; CHECK-NEXT: .LBB1_2 : ; %false
443
+ ; CHECK-NEXT: s_branch .LBB21_3
444
+ ; CHECK-NEXT: .LBB21_2 : ; %false
445
445
; CHECK-NEXT: s_mov_b32 s0, 33
446
- ; CHECK-NEXT: s_branch .LBB1_3
447
- ; CHECK-NEXT: .LBB1_3 :
446
+ ; CHECK-NEXT: s_branch .LBB21_3
447
+ ; CHECK-NEXT: .LBB21_3 :
448
448
%v1c = icmp ult i32 %v1 , 12
449
449
%v2c = icmp ugt i32 %v2 , 34
450
450
%c = and i1 %v1c , %v2c
@@ -463,14 +463,14 @@ define amdgpu_cs i32 @branch_divergent_simulated_negated_ballot_eq_zero_and(i32
463
463
; CHECK-NEXT: v_cmp_gt_u32_e32 vcc, 12, v0
464
464
; CHECK-NEXT: v_cmp_lt_u32_e64 s[0:1], 34, v1
465
465
; CHECK-NEXT: s_and_b64 vcc, vcc, s[0:1]
466
- ; CHECK-NEXT: s_cbranch_vccnz .LBB2_2
466
+ ; CHECK-NEXT: s_cbranch_vccnz .LBB22_2
467
467
; CHECK-NEXT: ; %bb.1: ; %false
468
468
; CHECK-NEXT: s_mov_b32 s0, 33
469
- ; CHECK-NEXT: s_branch .LBB2_3
470
- ; CHECK-NEXT: .LBB2_2 : ; %true
469
+ ; CHECK-NEXT: s_branch .LBB22_3
470
+ ; CHECK-NEXT: .LBB22_2 : ; %true
471
471
; CHECK-NEXT: s_mov_b32 s0, 42
472
- ; CHECK-NEXT: s_branch .LBB2_3
473
- ; CHECK-NEXT: .LBB2_3 :
472
+ ; CHECK-NEXT: s_branch .LBB22_3
473
+ ; CHECK-NEXT: .LBB22_3 :
474
474
%v1c = icmp ult i32 %v1 , 12
475
475
%v2c = icmp ugt i32 %v2 , 34
476
476
%c = and i1 %v1c , %v2c
@@ -492,14 +492,14 @@ define amdgpu_cs i32 @branch_uniform_simulated_negated_ballot_eq_zero_and(i32 in
492
492
; CHECK-NEXT: s_cselect_b64 s[0:1], -1, 0
493
493
; CHECK-NEXT: s_and_b64 s[0:1], s[2:3], s[0:1]
494
494
; CHECK-NEXT: s_and_b64 s[0:1], s[0:1], exec
495
- ; CHECK-NEXT: s_cbranch_scc1 .LBB3_2
495
+ ; CHECK-NEXT: s_cbranch_scc1 .LBB23_2
496
496
; CHECK-NEXT: ; %bb.1: ; %false
497
497
; CHECK-NEXT: s_mov_b32 s0, 33
498
- ; CHECK-NEXT: s_branch .LBB3_3
499
- ; CHECK-NEXT: .LBB3_2 : ; %true
498
+ ; CHECK-NEXT: s_branch .LBB23_3
499
+ ; CHECK-NEXT: .LBB23_2 : ; %true
500
500
; CHECK-NEXT: s_mov_b32 s0, 42
501
- ; CHECK-NEXT: s_branch .LBB3_3
502
- ; CHECK-NEXT: .LBB3_3 :
501
+ ; CHECK-NEXT: s_branch .LBB23_3
502
+ ; CHECK-NEXT: .LBB23_3 :
503
503
%v1c = icmp ult i32 %v1 , 12
504
504
%v2c = icmp ugt i32 %v2 , 34
505
505
%c = and i1 %v1c , %v2c
0 commit comments