@@ -1881,49 +1881,6 @@ entry:
1881
1881
ret <vscale x 1 x i64 > %a
1882
1882
}
1883
1883
1884
- define <vscale x 1 x i64 > @intrinsic_vadd_vx_sext_nxv1i64_nxv1i64_i64 (<vscale x 1 x i64 > %0 , i32 %1 , iXLen %2 ) nounwind {
1885
- ; RV32-LABEL: intrinsic_vadd_vx_sext_nxv1i64_nxv1i64_i64:
1886
- ; RV32: # %bb.0: # %entry
1887
- ; RV32-NEXT: vsetvli zero, a1, e64, m1, ta, mu
1888
- ; RV32-NEXT: vadd.vx v8, v8, a0
1889
- ; RV32-NEXT: ret
1890
- ;
1891
- ; RV64-LABEL: intrinsic_vadd_vx_sext_nxv1i64_nxv1i64_i64:
1892
- ; RV64: # %bb.0: # %entry
1893
- ; RV64-NEXT: sext.w a0, a0
1894
- ; RV64-NEXT: vsetvli zero, a1, e64, m1, ta, mu
1895
- ; RV64-NEXT: vadd.vx v8, v8, a0
1896
- ; RV64-NEXT: ret
1897
- entry:
1898
- %ext = sext i32 %1 to i64
1899
- %a = call <vscale x 1 x i64 > @llvm.riscv.vadd.nxv1i64.i64 (
1900
- <vscale x 1 x i64 > undef ,
1901
- <vscale x 1 x i64 > %0 ,
1902
- i64 %ext ,
1903
- iXLen %2 )
1904
-
1905
- ret <vscale x 1 x i64 > %a
1906
- }
1907
-
1908
- define <vscale x 1 x i64 > @intrinsic_vadd_vx_sextload_nxv1i64_nxv1i64_i64 (<vscale x 1 x i64 > %0 , i32* %1 , iXLen %2 ) nounwind {
1909
- ; CHECK-LABEL: intrinsic_vadd_vx_sextload_nxv1i64_nxv1i64_i64:
1910
- ; CHECK: # %bb.0: # %entry
1911
- ; CHECK-NEXT: lw a0, 0(a0)
1912
- ; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, mu
1913
- ; CHECK-NEXT: vadd.vx v8, v8, a0
1914
- ; CHECK-NEXT: ret
1915
- entry:
1916
- %load = load i32 , i32* %1
1917
- %ext = sext i32 %load to i64
1918
- %a = call <vscale x 1 x i64 > @llvm.riscv.vadd.nxv1i64.i64 (
1919
- <vscale x 1 x i64 > undef ,
1920
- <vscale x 1 x i64 > %0 ,
1921
- i64 %ext ,
1922
- iXLen %2 )
1923
-
1924
- ret <vscale x 1 x i64 > %a
1925
- }
1926
-
1927
1884
declare <vscale x 1 x i64 > @llvm.riscv.vadd.mask.nxv1i64.i64 (
1928
1885
<vscale x 1 x i64 >,
1929
1886
<vscale x 1 x i64 >,
0 commit comments