@@ -425,8 +425,7 @@ define amdgpu_kernel void @use_group_to_flat_addrspacecast(ptr addrspace(3) %ptr
425
425
;
426
426
; ATTRIBUTOR_HSA-LABEL: define {{[^@]+}}@use_group_to_flat_addrspacecast
427
427
; ATTRIBUTOR_HSA-SAME: (ptr addrspace(3) [[PTR:%.*]]) #[[ATTR12:[0-9]+]] {
428
- ; ATTRIBUTOR_HSA-NEXT: [[STOF:%.*]] = addrspacecast ptr addrspace(3) [[PTR]] to ptr
429
- ; ATTRIBUTOR_HSA-NEXT: store volatile i32 0, ptr [[STOF]], align 4
428
+ ; ATTRIBUTOR_HSA-NEXT: store volatile i32 0, ptr addrspace(3) [[PTR]], align 4
430
429
; ATTRIBUTOR_HSA-NEXT: ret void
431
430
;
432
431
%stof = addrspacecast ptr addrspace (3 ) %ptr to ptr
@@ -443,8 +442,7 @@ define amdgpu_kernel void @use_private_to_flat_addrspacecast(ptr addrspace(5) %p
443
442
;
444
443
; ATTRIBUTOR_HSA-LABEL: define {{[^@]+}}@use_private_to_flat_addrspacecast
445
444
; ATTRIBUTOR_HSA-SAME: (ptr addrspace(5) [[PTR:%.*]]) #[[ATTR12]] {
446
- ; ATTRIBUTOR_HSA-NEXT: [[STOF:%.*]] = addrspacecast ptr addrspace(5) [[PTR]] to ptr
447
- ; ATTRIBUTOR_HSA-NEXT: store volatile i32 0, ptr [[STOF]], align 4
445
+ ; ATTRIBUTOR_HSA-NEXT: store volatile i32 0, ptr addrspace(5) [[PTR]], align 4
448
446
; ATTRIBUTOR_HSA-NEXT: ret void
449
447
;
450
448
%stof = addrspacecast ptr addrspace (5 ) %ptr to ptr
@@ -478,23 +476,33 @@ define amdgpu_kernel void @use_flat_to_private_addrspacecast(ptr %ptr) #1 {
478
476
479
477
; No-op addrspacecast should not use queue ptr
480
478
define amdgpu_kernel void @use_global_to_flat_addrspacecast (ptr addrspace (1 ) %ptr ) #1 {
481
- ; HSA-LABEL: define {{[^@]+}}@use_global_to_flat_addrspacecast
482
- ; HSA-SAME: (ptr addrspace(1) [[PTR:%.*]]) #[[ATTR1]] {
483
- ; HSA-NEXT: [[STOF:%.*]] = addrspacecast ptr addrspace(1) [[PTR]] to ptr
484
- ; HSA-NEXT: store volatile i32 0, ptr [[STOF]], align 4
485
- ; HSA-NEXT: ret void
479
+ ; AKF_HSA-LABEL: define {{[^@]+}}@use_global_to_flat_addrspacecast
480
+ ; AKF_HSA-SAME: (ptr addrspace(1) [[PTR:%.*]]) #[[ATTR1]] {
481
+ ; AKF_HSA-NEXT: [[STOF:%.*]] = addrspacecast ptr addrspace(1) [[PTR]] to ptr
482
+ ; AKF_HSA-NEXT: store volatile i32 0, ptr [[STOF]], align 4
483
+ ; AKF_HSA-NEXT: ret void
484
+ ;
485
+ ; ATTRIBUTOR_HSA-LABEL: define {{[^@]+}}@use_global_to_flat_addrspacecast
486
+ ; ATTRIBUTOR_HSA-SAME: (ptr addrspace(1) [[PTR:%.*]]) #[[ATTR1]] {
487
+ ; ATTRIBUTOR_HSA-NEXT: store volatile i32 0, ptr addrspace(1) [[PTR]], align 4
488
+ ; ATTRIBUTOR_HSA-NEXT: ret void
486
489
;
487
490
%stof = addrspacecast ptr addrspace (1 ) %ptr to ptr
488
491
store volatile i32 0 , ptr %stof
489
492
ret void
490
493
}
491
494
492
495
define amdgpu_kernel void @use_constant_to_flat_addrspacecast (ptr addrspace (4 ) %ptr ) #1 {
493
- ; HSA-LABEL: define {{[^@]+}}@use_constant_to_flat_addrspacecast
494
- ; HSA-SAME: (ptr addrspace(4) [[PTR:%.*]]) #[[ATTR1]] {
495
- ; HSA-NEXT: [[STOF:%.*]] = addrspacecast ptr addrspace(4) [[PTR]] to ptr
496
- ; HSA-NEXT: [[LD:%.*]] = load volatile i32, ptr [[STOF]], align 4
497
- ; HSA-NEXT: ret void
496
+ ; AKF_HSA-LABEL: define {{[^@]+}}@use_constant_to_flat_addrspacecast
497
+ ; AKF_HSA-SAME: (ptr addrspace(4) [[PTR:%.*]]) #[[ATTR1]] {
498
+ ; AKF_HSA-NEXT: [[STOF:%.*]] = addrspacecast ptr addrspace(4) [[PTR]] to ptr
499
+ ; AKF_HSA-NEXT: [[LD:%.*]] = load volatile i32, ptr [[STOF]], align 4
500
+ ; AKF_HSA-NEXT: ret void
501
+ ;
502
+ ; ATTRIBUTOR_HSA-LABEL: define {{[^@]+}}@use_constant_to_flat_addrspacecast
503
+ ; ATTRIBUTOR_HSA-SAME: (ptr addrspace(4) [[PTR:%.*]]) #[[ATTR1]] {
504
+ ; ATTRIBUTOR_HSA-NEXT: [[LD:%.*]] = load volatile i32, ptr addrspace(4) [[PTR]], align 4
505
+ ; ATTRIBUTOR_HSA-NEXT: ret void
498
506
;
499
507
%stof = addrspacecast ptr addrspace (4 ) %ptr to ptr
500
508
%ld = load volatile i32 , ptr %stof
0 commit comments