@@ -108,17 +108,17 @@ extern "C" {
108
108
#[link_name = "llvm.mips.addv.d"]
109
109
fn msa_addv_d(a: v2i64, b: v2i64) -> v2i64;
110
110
#[link_name = "llvm.mips.addvi.b"]
111
- fn msa_addvi_b(a: v16i8, b: u32 ) -> v16i8;
111
+ fn msa_addvi_b(a: v16i8, b: i32 ) -> v16i8;
112
112
#[link_name = "llvm.mips.addvi.h"]
113
- fn msa_addvi_h(a: v8i16, b: u32 ) -> v8i16;
113
+ fn msa_addvi_h(a: v8i16, b: i32 ) -> v8i16;
114
114
#[link_name = "llvm.mips.addvi.w"]
115
- fn msa_addvi_w(a: v4i32, b: u32 ) -> v4i32;
115
+ fn msa_addvi_w(a: v4i32, b: i32 ) -> v4i32;
116
116
#[link_name = "llvm.mips.addvi.d"]
117
- fn msa_addvi_d(a: v2i64, b: u32 ) -> v2i64;
117
+ fn msa_addvi_d(a: v2i64, b: i32 ) -> v2i64;
118
118
#[link_name = "llvm.mips.and.v"]
119
119
fn msa_and_v(a: v16u8, b: v16u8) -> v16u8;
120
120
#[link_name = "llvm.mips.andi.b"]
121
- fn msa_andi_b(a: v16u8, b: u32 ) -> v16u8;
121
+ fn msa_andi_b(a: v16u8, b: i32 ) -> v16u8;
122
122
#[link_name = "llvm.mips.asub.s.b"]
123
123
fn msa_asub_s_b(a: v16i8, b: v16i8) -> v16i8;
124
124
#[link_name = "llvm.mips.asub.s.h"]
@@ -719,13 +719,13 @@ extern "C" {
719
719
#[link_name = "llvm.mips.insve.d"]
720
720
fn msa_insve_d(a: v2i64, b: i32, c: v2i64) -> v2i64;
721
721
#[link_name = "llvm.mips.ld.b"]
722
- fn msa_ld_b(mem_addr: *mut i8 , b: i32) -> v16i8;
722
+ fn msa_ld_b(mem_addr: *mut u8 , b: i32) -> v16i8;
723
723
#[link_name = "llvm.mips.ld.h"]
724
- fn msa_ld_h(mem_addr: *mut i8 , b: i32) -> v8i16;
724
+ fn msa_ld_h(mem_addr: *mut u8 , b: i32) -> v8i16;
725
725
#[link_name = "llvm.mips.ld.w"]
726
- fn msa_ld_w(mem_addr: *mut i8 , b: i32) -> v4i32;
726
+ fn msa_ld_w(mem_addr: *mut u8 , b: i32) -> v4i32;
727
727
#[link_name = "llvm.mips.ld.d"]
728
- fn msa_ld_d(mem_addr: *mut i8 , b: i32) -> v2i64;
728
+ fn msa_ld_d(mem_addr: *mut u8 , b: i32) -> v2i64;
729
729
#[link_name = "llvm.mips.ldi.b"]
730
730
fn msa_ldi_b(a: i32) -> v16i8;
731
731
#[link_name = "llvm.mips.ldi.h"]
@@ -1063,13 +1063,13 @@ extern "C" {
1063
1063
#[link_name = "llvm.mips.srlri.d"]
1064
1064
fn msa_srlri_d(a: v2i64, b: i32) -> v2i64;
1065
1065
#[link_name = "llvm.mips.st.b"]
1066
- fn msa_st_b(a: v16i8, mem_addr: *mut i8 , imm_s10: i32) -> ();
1066
+ fn msa_st_b(a: v16i8, mem_addr: *mut u8 , imm_s10: i32) -> ();
1067
1067
#[link_name = "llvm.mips.st.h"]
1068
- fn msa_st_h(a: v8i16, mem_addr: *mut i8 , imm_s11: i32) -> ();
1068
+ fn msa_st_h(a: v8i16, mem_addr: *mut u8 , imm_s11: i32) -> ();
1069
1069
#[link_name = "llvm.mips.st.w"]
1070
- fn msa_st_w(a: v4i32, mem_addr: *mut i8 , imm_s12: i32) -> ();
1070
+ fn msa_st_w(a: v4i32, mem_addr: *mut u8 , imm_s12: i32) -> ();
1071
1071
#[link_name = "llvm.mips.st.d"]
1072
- fn msa_st_d(a: v2i64, mem_addr: *mut i8 , imm_s13: i32) -> ();
1072
+ fn msa_st_d(a: v2i64, mem_addr: *mut u8 , imm_s13: i32) -> ();
1073
1073
#[link_name = "llvm.mips.subs.s.b"]
1074
1074
fn msa_subs_s_b(a: v16i8, b: v16i8) -> v16i8;
1075
1075
#[link_name = "llvm.mips.subs.s.h"]
@@ -1410,7 +1410,7 @@ pub unsafe fn __msa_addv_d(a: v2i64, b: v2i64) -> v2i64 {
1410
1410
#[target_feature(enable = "msa")]
1411
1411
#[cfg_attr(test, assert_instr(addvi.b, imm5 = 0b10111))]
1412
1412
#[rustc_args_required_const(1)]
1413
- pub unsafe fn __msa_addvi_b(a: v16i8, imm5: u32 ) -> v16i8 {
1413
+ pub unsafe fn __msa_addvi_b(a: v16i8, imm5: i32 ) -> v16i8 {
1414
1414
macro_rules! call {
1415
1415
($imm5:expr) => {
1416
1416
msa_addvi_b(a, $imm5)
@@ -1429,7 +1429,7 @@ pub unsafe fn __msa_addvi_b(a: v16i8, imm5: u32) -> v16i8 {
1429
1429
#[target_feature(enable = "msa")]
1430
1430
#[cfg_attr(test, assert_instr(addvi.h, imm5 = 0b10111))]
1431
1431
#[rustc_args_required_const(1)]
1432
- pub unsafe fn __msa_addvi_h(a: v8i16, imm5: u32 ) -> v8i16 {
1432
+ pub unsafe fn __msa_addvi_h(a: v8i16, imm5: i32 ) -> v8i16 {
1433
1433
macro_rules! call {
1434
1434
($imm5:expr) => {
1435
1435
msa_addvi_h(a, $imm5)
@@ -1448,7 +1448,7 @@ pub unsafe fn __msa_addvi_h(a: v8i16, imm5: u32) -> v8i16 {
1448
1448
#[target_feature(enable = "msa")]
1449
1449
#[cfg_attr(test, assert_instr(addvi.w, imm5 = 0b10111))]
1450
1450
#[rustc_args_required_const(1)]
1451
- pub unsafe fn __msa_addvi_w(a: v4i32, imm5: u32 ) -> v4i32 {
1451
+ pub unsafe fn __msa_addvi_w(a: v4i32, imm5: i32 ) -> v4i32 {
1452
1452
macro_rules! call {
1453
1453
($imm5:expr) => {
1454
1454
msa_addvi_w(a, $imm5)
@@ -1467,7 +1467,7 @@ pub unsafe fn __msa_addvi_w(a: v4i32, imm5: u32) -> v4i32 {
1467
1467
#[target_feature(enable = "msa")]
1468
1468
#[cfg_attr(test, assert_instr(addvi.d, imm5 = 0b10111))]
1469
1469
#[rustc_args_required_const(1)]
1470
- pub unsafe fn __msa_addvi_d(a: v2i64, imm5: u32 ) -> v2i64 {
1470
+ pub unsafe fn __msa_addvi_d(a: v2i64, imm5: i32 ) -> v2i64 {
1471
1471
macro_rules! call {
1472
1472
($imm5:expr) => {
1473
1473
msa_addvi_d(a, $imm5)
@@ -1500,7 +1500,7 @@ pub unsafe fn __msa_and_v(a: v16u8, b: v16u8) -> v16u8 {
1500
1500
#[target_feature(enable = "msa")]
1501
1501
#[cfg_attr(test, assert_instr(andi.b, imm8 = 0b10010111))]
1502
1502
#[rustc_args_required_const(1)]
1503
- pub unsafe fn __msa_andi_b(a: v16u8, imm8: u32 ) -> v16u8 {
1503
+ pub unsafe fn __msa_andi_b(a: v16u8, imm8: i32 ) -> v16u8 {
1504
1504
macro_rules! call {
1505
1505
($imm8:expr) => {
1506
1506
msa_andi_b(a, $imm8)
@@ -5946,7 +5946,7 @@ pub unsafe fn __msa_insve_d(a: v2i64, imm1: i32, c: v2i64) -> v2i64 {
5946
5946
#[target_feature(enable = "msa")]
5947
5947
#[cfg_attr(test, assert_instr(ld.b, imm_s10 = 0b1111111111))]
5948
5948
#[rustc_args_required_const(1)]
5949
- pub unsafe fn __msa_ld_b(mem_addr: *mut i8 , imm_s10: i32) -> v16i8 {
5949
+ pub unsafe fn __msa_ld_b(mem_addr: *mut u8 , imm_s10: i32) -> v16i8 {
5950
5950
macro_rules! call {
5951
5951
($imm_s10:expr) => {
5952
5952
msa_ld_b(mem_addr, $imm_s10)
@@ -5965,7 +5965,7 @@ pub unsafe fn __msa_ld_b(mem_addr: *mut i8, imm_s10: i32) -> v16i8 {
5965
5965
#[target_feature(enable = "msa")]
5966
5966
#[cfg_attr(test, assert_instr(ld.h, imm_s11 = 0b11111111111))]
5967
5967
#[rustc_args_required_const(1)]
5968
- pub unsafe fn __msa_ld_h(mem_addr: *mut i8 , imm_s11: i32) -> v8i16 {
5968
+ pub unsafe fn __msa_ld_h(mem_addr: *mut u8 , imm_s11: i32) -> v8i16 {
5969
5969
macro_rules! call {
5970
5970
($imm_s11:expr) => {
5971
5971
msa_ld_h(mem_addr, $imm_s11)
@@ -5984,7 +5984,7 @@ pub unsafe fn __msa_ld_h(mem_addr: *mut i8, imm_s11: i32) -> v8i16 {
5984
5984
#[target_feature(enable = "msa")]
5985
5985
#[cfg_attr(test, assert_instr(ld.w, imm_s12 = 0b111111111111))]
5986
5986
#[rustc_args_required_const(1)]
5987
- pub unsafe fn __msa_ld_w(mem_addr: *mut i8 , imm_s12: i32) -> v4i32 {
5987
+ pub unsafe fn __msa_ld_w(mem_addr: *mut u8 , imm_s12: i32) -> v4i32 {
5988
5988
macro_rules! call {
5989
5989
($imm_s12:expr) => {
5990
5990
msa_ld_w(mem_addr, $imm_s12)
@@ -6003,7 +6003,7 @@ pub unsafe fn __msa_ld_w(mem_addr: *mut i8, imm_s12: i32) -> v4i32 {
6003
6003
#[target_feature(enable = "msa")]
6004
6004
#[cfg_attr(test, assert_instr(ld.d, imm_s13 = 0b1111111111111))]
6005
6005
#[rustc_args_required_const(1)]
6006
- pub unsafe fn __msa_ld_d(mem_addr: *mut i8 , imm_s13: i32) -> v2i64 {
6006
+ pub unsafe fn __msa_ld_d(mem_addr: *mut u8 , imm_s13: i32) -> v2i64 {
6007
6007
macro_rules! call {
6008
6008
($imm_s13:expr) => {
6009
6009
msa_ld_d(mem_addr, $imm_s13)
@@ -8706,7 +8706,7 @@ pub unsafe fn __msa_srlri_d(a: v2i64, imm6: i32) -> v2i64 {
8706
8706
#[target_feature(enable = "msa")]
8707
8707
#[cfg_attr(test, assert_instr(st.b, imm_s10 = 0b1111111111))]
8708
8708
#[rustc_args_required_const(2)]
8709
- pub unsafe fn __msa_st_b(a: v16i8, mem_addr: *mut i8 , imm_s10: i32) -> () {
8709
+ pub unsafe fn __msa_st_b(a: v16i8, mem_addr: *mut u8 , imm_s10: i32) -> () {
8710
8710
macro_rules! call {
8711
8711
($imm_s10:expr) => {
8712
8712
msa_st_b(a, mem_addr, $imm_s10)
@@ -8725,7 +8725,7 @@ pub unsafe fn __msa_st_b(a: v16i8, mem_addr: *mut i8, imm_s10: i32) -> () {
8725
8725
#[target_feature(enable = "msa")]
8726
8726
#[cfg_attr(test, assert_instr(st.h, imm_s11 = 0b11111111111))]
8727
8727
#[rustc_args_required_const(2)]
8728
- pub unsafe fn __msa_st_h(a: v8i16, mem_addr: *mut i8 , imm_s11: i32) -> () {
8728
+ pub unsafe fn __msa_st_h(a: v8i16, mem_addr: *mut u8 , imm_s11: i32) -> () {
8729
8729
macro_rules! call {
8730
8730
($imm_s11:expr) => {
8731
8731
msa_st_h(a, mem_addr, $imm_s11)
@@ -8744,7 +8744,7 @@ pub unsafe fn __msa_st_h(a: v8i16, mem_addr: *mut i8, imm_s11: i32) -> () {
8744
8744
#[target_feature(enable = "msa")]
8745
8745
#[cfg_attr(test, assert_instr(st.w, imm_s12 = 0b111111111111))]
8746
8746
#[rustc_args_required_const(2)]
8747
- pub unsafe fn __msa_st_w(a: v4i32, mem_addr: *mut i8 , imm_s12: i32) -> () {
8747
+ pub unsafe fn __msa_st_w(a: v4i32, mem_addr: *mut u8 , imm_s12: i32) -> () {
8748
8748
macro_rules! call {
8749
8749
($imm_s12:expr) => {
8750
8750
msa_st_w(a, mem_addr, $imm_s12)
@@ -8763,7 +8763,7 @@ pub unsafe fn __msa_st_w(a: v4i32, mem_addr: *mut i8, imm_s12: i32) -> () {
8763
8763
#[target_feature(enable = "msa")]
8764
8764
#[cfg_attr(test, assert_instr(st.d, imm_s13 = 0b1111111111111))]
8765
8765
#[rustc_args_required_const(2)]
8766
- pub unsafe fn __msa_st_d(a: v2i64, mem_addr: *mut i8 , imm_s13: i32) -> () {
8766
+ pub unsafe fn __msa_st_d(a: v2i64, mem_addr: *mut u8 , imm_s13: i32) -> () {
8767
8767
macro_rules! call {
8768
8768
($imm_s13:expr) => {
8769
8769
msa_st_d(a, mem_addr, $imm_s13)
@@ -14811,7 +14811,7 @@ mod tests {
14811
14811
16, 17, 18, 19, 20, 21, 22, 23,
14812
14812
24, 25, 26, 27, 28, 29, 30, 31
14813
14813
];
14814
- let p = &mut a[4] as *mut _ as *mut i8 ;
14814
+ let p = &mut a[4] as *mut _ as *mut u8 ;
14815
14815
#[rustfmt::skip]
14816
14816
let r = i8x16::new(
14817
14817
13, 14, 15, 16,
@@ -14830,7 +14830,7 @@ mod tests {
14830
14830
0, 1, 2, 3, 4, 5, 6, 7,
14831
14831
8, 9, 10, 11, 12, 13, 14, 15
14832
14832
];
14833
- let p = &mut a[4] as *mut _ as *mut i8 ;
14833
+ let p = &mut a[4] as *mut _ as *mut u8 ;
14834
14834
#[rustfmt::skip]
14835
14835
let r = i16x8::new(3, 4, 5, 6, 7, 8, 9, 10);
14836
14836
@@ -14841,7 +14841,7 @@ mod tests {
14841
14841
unsafe fn test_msa_ld_w() {
14842
14842
#[rustfmt::skip]
14843
14843
let mut a : [i32; 8] = [0, 1, 2, 3, 4, 5, 6, 7];
14844
- let p = &mut a[3] as *mut _ as *mut i8 ;
14844
+ let p = &mut a[3] as *mut _ as *mut u8 ;
14845
14845
#[rustfmt::skip]
14846
14846
let r = i32x4::new(2, 3, 4, 5);
14847
14847
@@ -14852,7 +14852,7 @@ mod tests {
14852
14852
unsafe fn test_msa_ld_d() {
14853
14853
#[rustfmt::skip]
14854
14854
let mut a : [i64; 8] = [0, 1, 2, 3, 4, 5, 6, 7];
14855
- let p = &mut a[4] as *mut _ as *mut i8 ;
14855
+ let p = &mut a[4] as *mut _ as *mut u8 ;
14856
14856
#[rustfmt::skip]
14857
14857
let r = i64x2::new(0, 1);
14858
14858
@@ -17825,7 +17825,7 @@ mod tests {
17825
17825
let mut arr: [i16; 8] = [0, 0, 0, 0, 0, 0, 0, 0];
17826
17826
#[rustfmt::skip]
17827
17827
let r : [i16; 8] = [13, 14, 15, 16, 17, 18, 19, 20];
17828
- __msa_st_h(::mem::transmute(a), arr.as_mut_ptr() as *mut i8 , 0);
17828
+ __msa_st_h(::mem::transmute(a), arr.as_mut_ptr() as *mut u8 , 0);
17829
17829
assert_eq!(arr, r);
17830
17830
}
17831
17831
@@ -17836,7 +17836,7 @@ mod tests {
17836
17836
let mut arr: [i32; 4] = [0, 0, 0, 0];
17837
17837
#[rustfmt::skip]
17838
17838
let r : [i32; 4] = [13, 14, 15, 16];
17839
- __msa_st_w(::mem::transmute(a), arr.as_mut_ptr() as *mut i8 , 0);
17839
+ __msa_st_w(::mem::transmute(a), arr.as_mut_ptr() as *mut u8 , 0);
17840
17840
assert_eq!(arr, r);
17841
17841
}
17842
17842
@@ -17847,7 +17847,7 @@ mod tests {
17847
17847
let mut arr: [i64; 2] = [0, 0];
17848
17848
#[rustfmt::skip]
17849
17849
let r : [i64; 2] = [13, 14];
17850
- __msa_st_d(::mem::transmute(a), arr.as_mut_ptr() as *mut i8 , 0);
17850
+ __msa_st_d(::mem::transmute(a), arr.as_mut_ptr() as *mut u8 , 0);
17851
17851
assert_eq!(arr, r);
17852
17852
}
17853
17853
0 commit comments