@@ -870,31 +870,31 @@ define half @caller_half_return_stack2(half %x, half %y) nounwind {
870
870
; RV32I-NEXT: ADJCALLSTACKDOWN 4, 0, implicit-def $x2, implicit $x2
871
871
; RV32I-NEXT: [[ANYEXT:%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC]](s16)
872
872
; RV32I-NEXT: [[ANYEXT1:%[0-9]+]]:_(s32) = G_ANYEXT [[C]](s16)
873
- ; RV32I-NEXT: [[ANYEXT2 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC ]](s16 )
874
- ; RV32I-NEXT: [[ANYEXT3 :%[0-9]+]]:_(s32) = G_ANYEXT [[C1]](s16)
875
- ; RV32I-NEXT: [[ANYEXT4 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC ]](s16 )
876
- ; RV32I-NEXT: [[ANYEXT5 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC1]](s16)
877
- ; RV32I-NEXT: [[ANYEXT6 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC1 ]](s16 )
878
- ; RV32I-NEXT: [[ANYEXT7 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC1 ]](s16 )
879
- ; RV32I-NEXT: [[ANYEXT8 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC ]](s16 )
880
- ; RV32I-NEXT: [[COPY2 :%[0-9]+]]:_(p0) = COPY $x2
873
+ ; RV32I-NEXT: [[COPY2 :%[0-9]+]]:_(s32) = COPY [[ANYEXT ]](s32 )
874
+ ; RV32I-NEXT: [[ANYEXT2 :%[0-9]+]]:_(s32) = G_ANYEXT [[C1]](s16)
875
+ ; RV32I-NEXT: [[COPY3 :%[0-9]+]]:_(s32) = COPY [[ANYEXT ]](s32 )
876
+ ; RV32I-NEXT: [[ANYEXT3 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC1]](s16)
877
+ ; RV32I-NEXT: [[COPY4 :%[0-9]+]]:_(s32) = COPY [[ANYEXT3 ]](s32 )
878
+ ; RV32I-NEXT: [[COPY5 :%[0-9]+]]:_(s32) = COPY [[ANYEXT3 ]](s32 )
879
+ ; RV32I-NEXT: [[COPY6 :%[0-9]+]]:_(s32) = COPY [[ANYEXT ]](s32 )
880
+ ; RV32I-NEXT: [[COPY7 :%[0-9]+]]:_(p0) = COPY $x2
881
881
; RV32I-NEXT: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 0
882
- ; RV32I-NEXT: [[PTR_ADD:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY2 ]], [[C2]](s32)
883
- ; RV32I-NEXT: G_STORE [[ANYEXT8 ]](s32), [[PTR_ADD]](p0) :: (store (s32) into stack, align 16)
882
+ ; RV32I-NEXT: [[PTR_ADD:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY7 ]], [[C2]](s32)
883
+ ; RV32I-NEXT: G_STORE [[COPY6 ]](s32), [[PTR_ADD]](p0) :: (store (s32) into stack, align 16)
884
884
; RV32I-NEXT: $x10 = COPY [[ANYEXT]](s32)
885
885
; RV32I-NEXT: $x11 = COPY [[ANYEXT1]](s32)
886
- ; RV32I-NEXT: $x12 = COPY [[ANYEXT2 ]](s32)
887
- ; RV32I-NEXT: $x13 = COPY [[ANYEXT3 ]](s32)
888
- ; RV32I-NEXT: $x14 = COPY [[ANYEXT4 ]](s32)
889
- ; RV32I-NEXT: $x15 = COPY [[ANYEXT5 ]](s32)
890
- ; RV32I-NEXT: $x16 = COPY [[ANYEXT6 ]](s32)
891
- ; RV32I-NEXT: $x17 = COPY [[ANYEXT7 ]](s32)
886
+ ; RV32I-NEXT: $x12 = COPY [[COPY2 ]](s32)
887
+ ; RV32I-NEXT: $x13 = COPY [[ANYEXT2 ]](s32)
888
+ ; RV32I-NEXT: $x14 = COPY [[COPY3 ]](s32)
889
+ ; RV32I-NEXT: $x15 = COPY [[ANYEXT3 ]](s32)
890
+ ; RV32I-NEXT: $x16 = COPY [[COPY4 ]](s32)
891
+ ; RV32I-NEXT: $x17 = COPY [[COPY5 ]](s32)
892
892
; RV32I-NEXT: PseudoCALL target-flags(riscv-call) @callee_half_return_stack2, csr_ilp32_lp64, implicit-def $x1, implicit $x10, implicit $x11, implicit $x12, implicit $x13, implicit $x14, implicit $x15, implicit $x16, implicit $x17, implicit-def $x10
893
893
; RV32I-NEXT: ADJCALLSTACKUP 4, 0, implicit-def $x2, implicit $x2
894
- ; RV32I-NEXT: [[COPY3 :%[0-9]+]]:_(s32) = COPY $x10
895
- ; RV32I-NEXT: [[TRUNC2:%[0-9]+]]:_(s16) = G_TRUNC [[COPY3 ]](s32)
896
- ; RV32I-NEXT: [[ANYEXT9 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC2]](s16)
897
- ; RV32I-NEXT: $x10 = COPY [[ANYEXT9 ]](s32)
894
+ ; RV32I-NEXT: [[COPY8 :%[0-9]+]]:_(s32) = COPY $x10
895
+ ; RV32I-NEXT: [[TRUNC2:%[0-9]+]]:_(s16) = G_TRUNC [[COPY8 ]](s32)
896
+ ; RV32I-NEXT: [[ANYEXT4 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC2]](s16)
897
+ ; RV32I-NEXT: $x10 = COPY [[ANYEXT4 ]](s32)
898
898
; RV32I-NEXT: PseudoRET implicit $x10
899
899
;
900
900
; RV32IF-LABEL: name: caller_half_return_stack2
@@ -910,28 +910,28 @@ define half @caller_half_return_stack2(half %x, half %y) nounwind {
910
910
; RV32IF-NEXT: ADJCALLSTACKDOWN 0, 0, implicit-def $x2, implicit $x2
911
911
; RV32IF-NEXT: [[ANYEXT:%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC]](s16)
912
912
; RV32IF-NEXT: [[ANYEXT1:%[0-9]+]]:_(s32) = G_ANYEXT [[C]](s16)
913
- ; RV32IF-NEXT: [[ANYEXT2 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC ]](s16 )
914
- ; RV32IF-NEXT: [[ANYEXT3 :%[0-9]+]]:_(s32) = G_ANYEXT [[C1]](s16)
915
- ; RV32IF-NEXT: [[ANYEXT4 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC ]](s16 )
916
- ; RV32IF-NEXT: [[ANYEXT5 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC1]](s16)
917
- ; RV32IF-NEXT: [[ANYEXT6 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC1 ]](s16 )
918
- ; RV32IF-NEXT: [[ANYEXT7 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC1 ]](s16 )
919
- ; RV32IF-NEXT: [[ANYEXT8 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC ]](s16 )
913
+ ; RV32IF-NEXT: [[COPY2 :%[0-9]+]]:_(s32) = COPY [[ANYEXT ]](s32 )
914
+ ; RV32IF-NEXT: [[ANYEXT2 :%[0-9]+]]:_(s32) = G_ANYEXT [[C1]](s16)
915
+ ; RV32IF-NEXT: [[COPY3 :%[0-9]+]]:_(s32) = COPY [[ANYEXT ]](s32 )
916
+ ; RV32IF-NEXT: [[ANYEXT3 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC1]](s16)
917
+ ; RV32IF-NEXT: [[COPY4 :%[0-9]+]]:_(s32) = COPY [[ANYEXT3 ]](s32 )
918
+ ; RV32IF-NEXT: [[COPY5 :%[0-9]+]]:_(s32) = COPY [[ANYEXT3 ]](s32 )
919
+ ; RV32IF-NEXT: [[COPY6 :%[0-9]+]]:_(s32) = COPY [[ANYEXT ]](s32 )
920
920
; RV32IF-NEXT: $f10_f = COPY [[ANYEXT]](s32)
921
921
; RV32IF-NEXT: $f11_f = COPY [[ANYEXT1]](s32)
922
- ; RV32IF-NEXT: $f12_f = COPY [[ANYEXT2 ]](s32)
923
- ; RV32IF-NEXT: $f13_f = COPY [[ANYEXT3 ]](s32)
924
- ; RV32IF-NEXT: $f14_f = COPY [[ANYEXT4 ]](s32)
925
- ; RV32IF-NEXT: $f15_f = COPY [[ANYEXT5 ]](s32)
926
- ; RV32IF-NEXT: $f16_f = COPY [[ANYEXT6 ]](s32)
927
- ; RV32IF-NEXT: $f17_f = COPY [[ANYEXT7 ]](s32)
928
- ; RV32IF-NEXT: $x10 = COPY [[ANYEXT8 ]](s32)
922
+ ; RV32IF-NEXT: $f12_f = COPY [[COPY2 ]](s32)
923
+ ; RV32IF-NEXT: $f13_f = COPY [[ANYEXT2 ]](s32)
924
+ ; RV32IF-NEXT: $f14_f = COPY [[COPY3 ]](s32)
925
+ ; RV32IF-NEXT: $f15_f = COPY [[ANYEXT3 ]](s32)
926
+ ; RV32IF-NEXT: $f16_f = COPY [[COPY4 ]](s32)
927
+ ; RV32IF-NEXT: $f17_f = COPY [[COPY5 ]](s32)
928
+ ; RV32IF-NEXT: $x10 = COPY [[COPY6 ]](s32)
929
929
; RV32IF-NEXT: PseudoCALL target-flags(riscv-call) @callee_half_return_stack2, csr_ilp32f_lp64f, implicit-def $x1, implicit $f10_f, implicit $f11_f, implicit $f12_f, implicit $f13_f, implicit $f14_f, implicit $f15_f, implicit $f16_f, implicit $f17_f, implicit $x10, implicit-def $f10_f
930
930
; RV32IF-NEXT: ADJCALLSTACKUP 0, 0, implicit-def $x2, implicit $x2
931
- ; RV32IF-NEXT: [[COPY2 :%[0-9]+]]:_(s32) = COPY $f10_f
932
- ; RV32IF-NEXT: [[TRUNC2:%[0-9]+]]:_(s16) = G_TRUNC [[COPY2 ]](s32)
933
- ; RV32IF-NEXT: [[ANYEXT9 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC2]](s16)
934
- ; RV32IF-NEXT: $f10_f = COPY [[ANYEXT9 ]](s32)
931
+ ; RV32IF-NEXT: [[COPY7 :%[0-9]+]]:_(s32) = COPY $f10_f
932
+ ; RV32IF-NEXT: [[TRUNC2:%[0-9]+]]:_(s16) = G_TRUNC [[COPY7 ]](s32)
933
+ ; RV32IF-NEXT: [[ANYEXT4 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC2]](s16)
934
+ ; RV32IF-NEXT: $f10_f = COPY [[ANYEXT4 ]](s32)
935
935
; RV32IF-NEXT: PseudoRET implicit $f10_f
936
936
;
937
937
; RV32IZFH-LABEL: name: caller_half_return_stack2
@@ -972,31 +972,31 @@ define half @caller_half_return_stack2(half %x, half %y) nounwind {
972
972
; RV64I-NEXT: ADJCALLSTACKDOWN 8, 0, implicit-def $x2, implicit $x2
973
973
; RV64I-NEXT: [[ANYEXT:%[0-9]+]]:_(s64) = G_ANYEXT [[TRUNC]](s16)
974
974
; RV64I-NEXT: [[ANYEXT1:%[0-9]+]]:_(s64) = G_ANYEXT [[C]](s16)
975
- ; RV64I-NEXT: [[ANYEXT2 :%[0-9]+]]:_(s64) = G_ANYEXT [[TRUNC ]](s16 )
976
- ; RV64I-NEXT: [[ANYEXT3 :%[0-9]+]]:_(s64) = G_ANYEXT [[C1]](s16)
977
- ; RV64I-NEXT: [[ANYEXT4 :%[0-9]+]]:_(s64) = G_ANYEXT [[TRUNC ]](s16 )
978
- ; RV64I-NEXT: [[ANYEXT5 :%[0-9]+]]:_(s64) = G_ANYEXT [[TRUNC1]](s16)
979
- ; RV64I-NEXT: [[ANYEXT6 :%[0-9]+]]:_(s64) = G_ANYEXT [[TRUNC1 ]](s16 )
980
- ; RV64I-NEXT: [[ANYEXT7 :%[0-9]+]]:_(s64) = G_ANYEXT [[TRUNC1 ]](s16 )
981
- ; RV64I-NEXT: [[ANYEXT8 :%[0-9]+]]:_(s64) = G_ANYEXT [[TRUNC ]](s16 )
982
- ; RV64I-NEXT: [[COPY2 :%[0-9]+]]:_(p0) = COPY $x2
975
+ ; RV64I-NEXT: [[COPY2 :%[0-9]+]]:_(s64) = COPY [[ANYEXT ]](s64 )
976
+ ; RV64I-NEXT: [[ANYEXT2 :%[0-9]+]]:_(s64) = G_ANYEXT [[C1]](s16)
977
+ ; RV64I-NEXT: [[COPY3 :%[0-9]+]]:_(s64) = COPY [[ANYEXT ]](s64 )
978
+ ; RV64I-NEXT: [[ANYEXT3 :%[0-9]+]]:_(s64) = G_ANYEXT [[TRUNC1]](s16)
979
+ ; RV64I-NEXT: [[COPY4 :%[0-9]+]]:_(s64) = COPY [[ANYEXT3 ]](s64 )
980
+ ; RV64I-NEXT: [[COPY5 :%[0-9]+]]:_(s64) = COPY [[ANYEXT3 ]](s64 )
981
+ ; RV64I-NEXT: [[COPY6 :%[0-9]+]]:_(s64) = COPY [[ANYEXT ]](s64 )
982
+ ; RV64I-NEXT: [[COPY7 :%[0-9]+]]:_(p0) = COPY $x2
983
983
; RV64I-NEXT: [[C2:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
984
- ; RV64I-NEXT: [[PTR_ADD:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY2 ]], [[C2]](s64)
985
- ; RV64I-NEXT: G_STORE [[ANYEXT8 ]](s64), [[PTR_ADD]](p0) :: (store (s64) into stack, align 16)
984
+ ; RV64I-NEXT: [[PTR_ADD:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY7 ]], [[C2]](s64)
985
+ ; RV64I-NEXT: G_STORE [[COPY6 ]](s64), [[PTR_ADD]](p0) :: (store (s64) into stack, align 16)
986
986
; RV64I-NEXT: $x10 = COPY [[ANYEXT]](s64)
987
987
; RV64I-NEXT: $x11 = COPY [[ANYEXT1]](s64)
988
- ; RV64I-NEXT: $x12 = COPY [[ANYEXT2 ]](s64)
989
- ; RV64I-NEXT: $x13 = COPY [[ANYEXT3 ]](s64)
990
- ; RV64I-NEXT: $x14 = COPY [[ANYEXT4 ]](s64)
991
- ; RV64I-NEXT: $x15 = COPY [[ANYEXT5 ]](s64)
992
- ; RV64I-NEXT: $x16 = COPY [[ANYEXT6 ]](s64)
993
- ; RV64I-NEXT: $x17 = COPY [[ANYEXT7 ]](s64)
988
+ ; RV64I-NEXT: $x12 = COPY [[COPY2 ]](s64)
989
+ ; RV64I-NEXT: $x13 = COPY [[ANYEXT2 ]](s64)
990
+ ; RV64I-NEXT: $x14 = COPY [[COPY3 ]](s64)
991
+ ; RV64I-NEXT: $x15 = COPY [[ANYEXT3 ]](s64)
992
+ ; RV64I-NEXT: $x16 = COPY [[COPY4 ]](s64)
993
+ ; RV64I-NEXT: $x17 = COPY [[COPY5 ]](s64)
994
994
; RV64I-NEXT: PseudoCALL target-flags(riscv-call) @callee_half_return_stack2, csr_ilp32_lp64, implicit-def $x1, implicit $x10, implicit $x11, implicit $x12, implicit $x13, implicit $x14, implicit $x15, implicit $x16, implicit $x17, implicit-def $x10
995
995
; RV64I-NEXT: ADJCALLSTACKUP 8, 0, implicit-def $x2, implicit $x2
996
- ; RV64I-NEXT: [[COPY3 :%[0-9]+]]:_(s64) = COPY $x10
997
- ; RV64I-NEXT: [[TRUNC2:%[0-9]+]]:_(s16) = G_TRUNC [[COPY3 ]](s64)
998
- ; RV64I-NEXT: [[ANYEXT9 :%[0-9]+]]:_(s64) = G_ANYEXT [[TRUNC2]](s16)
999
- ; RV64I-NEXT: $x10 = COPY [[ANYEXT9 ]](s64)
996
+ ; RV64I-NEXT: [[COPY8 :%[0-9]+]]:_(s64) = COPY $x10
997
+ ; RV64I-NEXT: [[TRUNC2:%[0-9]+]]:_(s16) = G_TRUNC [[COPY8 ]](s64)
998
+ ; RV64I-NEXT: [[ANYEXT4 :%[0-9]+]]:_(s64) = G_ANYEXT [[TRUNC2]](s16)
999
+ ; RV64I-NEXT: $x10 = COPY [[ANYEXT4 ]](s64)
1000
1000
; RV64I-NEXT: PseudoRET implicit $x10
1001
1001
;
1002
1002
; RV64IF-LABEL: name: caller_half_return_stack2
@@ -1012,28 +1012,28 @@ define half @caller_half_return_stack2(half %x, half %y) nounwind {
1012
1012
; RV64IF-NEXT: ADJCALLSTACKDOWN 0, 0, implicit-def $x2, implicit $x2
1013
1013
; RV64IF-NEXT: [[ANYEXT:%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC]](s16)
1014
1014
; RV64IF-NEXT: [[ANYEXT1:%[0-9]+]]:_(s32) = G_ANYEXT [[C]](s16)
1015
- ; RV64IF-NEXT: [[ANYEXT2 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC ]](s16 )
1016
- ; RV64IF-NEXT: [[ANYEXT3 :%[0-9]+]]:_(s32) = G_ANYEXT [[C1]](s16)
1017
- ; RV64IF-NEXT: [[ANYEXT4 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC ]](s16 )
1018
- ; RV64IF-NEXT: [[ANYEXT5 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC1]](s16)
1019
- ; RV64IF-NEXT: [[ANYEXT6 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC1 ]](s16 )
1020
- ; RV64IF-NEXT: [[ANYEXT7 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC1 ]](s16 )
1015
+ ; RV64IF-NEXT: [[COPY2 :%[0-9]+]]:_(s32) = COPY [[ANYEXT ]](s32 )
1016
+ ; RV64IF-NEXT: [[ANYEXT2 :%[0-9]+]]:_(s32) = G_ANYEXT [[C1]](s16)
1017
+ ; RV64IF-NEXT: [[COPY3 :%[0-9]+]]:_(s32) = COPY [[ANYEXT ]](s32 )
1018
+ ; RV64IF-NEXT: [[ANYEXT3 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC1]](s16)
1019
+ ; RV64IF-NEXT: [[COPY4 :%[0-9]+]]:_(s32) = COPY [[ANYEXT3 ]](s32 )
1020
+ ; RV64IF-NEXT: [[COPY5 :%[0-9]+]]:_(s32) = COPY [[ANYEXT3 ]](s32 )
1021
1021
; RV64IF-NEXT: $f10_f = COPY [[ANYEXT]](s32)
1022
1022
; RV64IF-NEXT: $f11_f = COPY [[ANYEXT1]](s32)
1023
- ; RV64IF-NEXT: $f12_f = COPY [[ANYEXT2 ]](s32)
1024
- ; RV64IF-NEXT: $f13_f = COPY [[ANYEXT3 ]](s32)
1025
- ; RV64IF-NEXT: $f14_f = COPY [[ANYEXT4 ]](s32)
1026
- ; RV64IF-NEXT: $f15_f = COPY [[ANYEXT5 ]](s32)
1027
- ; RV64IF-NEXT: $f16_f = COPY [[ANYEXT6 ]](s32)
1028
- ; RV64IF-NEXT: $f17_f = COPY [[ANYEXT7 ]](s32)
1029
- ; RV64IF-NEXT: [[ANYEXT8 :%[0-9]+]]:_(s64) = G_ANYEXT [[TRUNC]](s16)
1030
- ; RV64IF-NEXT: $x10 = COPY [[ANYEXT8 ]](s64)
1023
+ ; RV64IF-NEXT: $f12_f = COPY [[COPY2 ]](s32)
1024
+ ; RV64IF-NEXT: $f13_f = COPY [[ANYEXT2 ]](s32)
1025
+ ; RV64IF-NEXT: $f14_f = COPY [[COPY3 ]](s32)
1026
+ ; RV64IF-NEXT: $f15_f = COPY [[ANYEXT3 ]](s32)
1027
+ ; RV64IF-NEXT: $f16_f = COPY [[COPY4 ]](s32)
1028
+ ; RV64IF-NEXT: $f17_f = COPY [[COPY5 ]](s32)
1029
+ ; RV64IF-NEXT: [[ANYEXT4 :%[0-9]+]]:_(s64) = G_ANYEXT [[TRUNC]](s16)
1030
+ ; RV64IF-NEXT: $x10 = COPY [[ANYEXT4 ]](s64)
1031
1031
; RV64IF-NEXT: PseudoCALL target-flags(riscv-call) @callee_half_return_stack2, csr_ilp32f_lp64f, implicit-def $x1, implicit $f10_f, implicit $f11_f, implicit $f12_f, implicit $f13_f, implicit $f14_f, implicit $f15_f, implicit $f16_f, implicit $f17_f, implicit $x10, implicit-def $f10_f
1032
1032
; RV64IF-NEXT: ADJCALLSTACKUP 0, 0, implicit-def $x2, implicit $x2
1033
- ; RV64IF-NEXT: [[COPY2 :%[0-9]+]]:_(s32) = COPY $f10_f
1034
- ; RV64IF-NEXT: [[TRUNC2:%[0-9]+]]:_(s16) = G_TRUNC [[COPY2 ]](s32)
1035
- ; RV64IF-NEXT: [[ANYEXT9 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC2]](s16)
1036
- ; RV64IF-NEXT: $f10_f = COPY [[ANYEXT9 ]](s32)
1033
+ ; RV64IF-NEXT: [[COPY6 :%[0-9]+]]:_(s32) = COPY $f10_f
1034
+ ; RV64IF-NEXT: [[TRUNC2:%[0-9]+]]:_(s16) = G_TRUNC [[COPY6 ]](s32)
1035
+ ; RV64IF-NEXT: [[ANYEXT5 :%[0-9]+]]:_(s32) = G_ANYEXT [[TRUNC2]](s16)
1036
+ ; RV64IF-NEXT: $f10_f = COPY [[ANYEXT5 ]](s32)
1037
1037
; RV64IF-NEXT: PseudoRET implicit $f10_f
1038
1038
;
1039
1039
; RV64IZFH-LABEL: name: caller_half_return_stack2
0 commit comments