|
| 1 | +# NOTE: Assertions have been autogenerated by utils/update_mir_test_checks.py UTC_ARGS: --version 4 |
| 2 | +# RUN: llc -mtriple=amdgcn-amd-amdhsa -mcpu=gfx90a -run-pass=liveintervals -run-pass=twoaddressinstruction -verify-machineinstrs -o - %s | FileCheck --check-prefix=GFX90A %s |
| 3 | + |
| 4 | +--- |
| 5 | +name: aligned_partial_vgpr_64 |
| 6 | +tracksRegLiveness: true |
| 7 | +body: | |
| 8 | + bb.0: |
| 9 | + liveins: $sgpr0, $sgpr1, $sgpr2, $sgpr3, $sgpr4, $sgpr5, $sgpr6, $sgpr7, $vgpr0, $vgpr1 |
| 10 | +
|
| 11 | + ; GFX90A-LABEL: name: aligned_partial_vgpr_64 |
| 12 | + ; GFX90A: liveins: $sgpr0, $sgpr1, $sgpr2, $sgpr3, $sgpr4, $sgpr5, $sgpr6, $sgpr7, $vgpr0, $vgpr1 |
| 13 | + ; GFX90A-NEXT: {{ $}} |
| 14 | + ; GFX90A-NEXT: [[COPY:%[0-9]+]]:vgpr_32 = COPY $vgpr1 |
| 15 | + ; GFX90A-NEXT: [[COPY1:%[0-9]+]]:vgpr_32 = COPY $vgpr0 |
| 16 | + ; GFX90A-NEXT: [[COPY2:%[0-9]+]]:sgpr_32 = COPY $sgpr7 |
| 17 | + ; GFX90A-NEXT: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr6 |
| 18 | + ; GFX90A-NEXT: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr5 |
| 19 | + ; GFX90A-NEXT: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr4 |
| 20 | + ; GFX90A-NEXT: [[COPY6:%[0-9]+]]:sgpr_32 = COPY $sgpr3 |
| 21 | + ; GFX90A-NEXT: [[COPY7:%[0-9]+]]:sgpr_32 = COPY $sgpr2 |
| 22 | + ; GFX90A-NEXT: [[COPY8:%[0-9]+]]:sgpr_32 = COPY $sgpr1 |
| 23 | + ; GFX90A-NEXT: [[COPY9:%[0-9]+]]:sgpr_32 = COPY $sgpr0 |
| 24 | + ; GFX90A-NEXT: undef [[COPY10:%[0-9]+]].sub0:sgpr_256 = COPY [[COPY9]] |
| 25 | + ; GFX90A-NEXT: [[COPY10:%[0-9]+]].sub1:sgpr_256 = COPY [[COPY8]] |
| 26 | + ; GFX90A-NEXT: [[COPY10:%[0-9]+]].sub2:sgpr_256 = COPY [[COPY7]] |
| 27 | + ; GFX90A-NEXT: [[COPY10:%[0-9]+]].sub3:sgpr_256 = COPY [[COPY6]] |
| 28 | + ; GFX90A-NEXT: [[COPY10:%[0-9]+]].sub4:sgpr_256 = COPY [[COPY5]] |
| 29 | + ; GFX90A-NEXT: [[COPY10:%[0-9]+]].sub5:sgpr_256 = COPY [[COPY4]] |
| 30 | + ; GFX90A-NEXT: [[COPY10:%[0-9]+]].sub6:sgpr_256 = COPY [[COPY3]] |
| 31 | + ; GFX90A-NEXT: [[COPY10:%[0-9]+]].sub7:sgpr_256 = COPY [[COPY2]] |
| 32 | + ; GFX90A-NEXT: undef [[COPY11:%[0-9]+]].sub0:vreg_64_align2 = COPY [[COPY]] |
| 33 | + ; GFX90A-NEXT: [[COPY12:%[0-9]+]]:vgpr_32 = COPY [[COPY1]] |
| 34 | + ; GFX90A-NEXT: [[COPY12:%[0-9]+]]:vgpr_32 = IMAGE_ATOMIC_SWAP_V1_V1_gfx90a [[COPY12]], [[COPY11]].sub0, [[COPY10]], 1, -1, 1, 0, 0, 0, implicit $exec, implicit [[COPY11]] :: (volatile dereferenceable load store (s32), addrspace 8) |
| 35 | + ; GFX90A-NEXT: $vgpr0 = COPY [[COPY12]] |
| 36 | + ; GFX90A-NEXT: SI_RETURN_TO_EPILOG $vgpr0 |
| 37 | + %9:vgpr_32 = COPY $vgpr1 |
| 38 | + %8:vgpr_32 = COPY $vgpr0 |
| 39 | + %7:sgpr_32 = COPY $sgpr7 |
| 40 | + %6:sgpr_32 = COPY $sgpr6 |
| 41 | + %5:sgpr_32 = COPY $sgpr5 |
| 42 | + %4:sgpr_32 = COPY $sgpr4 |
| 43 | + %3:sgpr_32 = COPY $sgpr3 |
| 44 | + %2:sgpr_32 = COPY $sgpr2 |
| 45 | + %1:sgpr_32 = COPY $sgpr1 |
| 46 | + %0:sgpr_32 = COPY $sgpr0 |
| 47 | + %11:sgpr_256 = REG_SEQUENCE %0:sgpr_32, %subreg.sub0, %1:sgpr_32, %subreg.sub1, %2:sgpr_32, %subreg.sub2, %3:sgpr_32, %subreg.sub3, %4:sgpr_32, %subreg.sub4, %5:sgpr_32, %subreg.sub5, %6:sgpr_32, %subreg.sub6, %7:sgpr_32, %subreg.sub7 |
| 48 | + %14:vreg_64_align2 = REG_SEQUENCE %9:vgpr_32, %subreg.sub0, undef %13:vgpr_32, %subreg.sub1 |
| 49 | + %12:vgpr_32 = IMAGE_ATOMIC_SWAP_V1_V1_gfx90a %8:vgpr_32, %14.sub0:vreg_64_align2, %11:sgpr_256, 1, -1, 1, 0, 0, 0, implicit $exec, implicit %14:vreg_64_align2 :: (volatile dereferenceable load store (s32), addrspace 8) |
| 50 | + $vgpr0 = COPY %12:vgpr_32 |
| 51 | + SI_RETURN_TO_EPILOG $vgpr0 |
| 52 | +... |
0 commit comments