|
1 | 1 | ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
2 | 2 | ; RUN: llc -mtriple=riscv32 -mattr=+v -target-abi=ilp32 \
|
3 |
| -; RUN: -verify-machineinstrs < %s | FileCheck %s |
| 3 | +; RUN: -verify-machineinstrs < %s | FileCheck -check-prefixes=CHECK,RV32 %s |
4 | 4 | ; RUN: llc -mtriple=riscv64 -mattr=+v -target-abi=lp64 \
|
5 |
| -; RUN: -verify-machineinstrs < %s | FileCheck %s |
| 5 | +; RUN: -verify-machineinstrs < %s | FileCheck -check-prefixes=CHECK,RV64 %s |
6 | 6 |
|
7 | 7 | define <vscale x 1 x i32> @vnsrl_wx_i64_nxv1i32(<vscale x 1 x i64> %va, i64 %b) {
|
8 | 8 | ; CHECK-LABEL: vnsrl_wx_i64_nxv1i32:
|
@@ -632,3 +632,77 @@ define <vscale x 8 x i32> @vnsrl_wi_i32_nxv8i32_zext(<vscale x 8 x i64> %va) {
|
632 | 632 | %y = trunc <vscale x 8 x i64> %x to <vscale x 8 x i32>
|
633 | 633 | ret <vscale x 8 x i32> %y
|
634 | 634 | }
|
| 635 | + |
| 636 | +define <vscale x 1 x i16> @vnsrl_wx_i64_nxv1i16(<vscale x 1 x i32> %va, i64 %b) { |
| 637 | +; RV32-LABEL: vnsrl_wx_i64_nxv1i16: |
| 638 | +; RV32: # %bb.0: |
| 639 | +; RV32-NEXT: addi sp, sp, -16 |
| 640 | +; RV32-NEXT: .cfi_def_cfa_offset 16 |
| 641 | +; RV32-NEXT: sw a1, 12(sp) |
| 642 | +; RV32-NEXT: sw a0, 8(sp) |
| 643 | +; RV32-NEXT: addi a0, sp, 8 |
| 644 | +; RV32-NEXT: vsetvli a1, zero, e32, mf2, ta, ma |
| 645 | +; RV32-NEXT: vlse64.v v9, (a0), zero |
| 646 | +; RV32-NEXT: vnsrl.wi v9, v9, 0 |
| 647 | +; RV32-NEXT: vsrl.vv v8, v8, v9 |
| 648 | +; RV32-NEXT: vsetvli zero, zero, e16, mf4, ta, ma |
| 649 | +; RV32-NEXT: vnsrl.wi v8, v8, 0 |
| 650 | +; RV32-NEXT: addi sp, sp, 16 |
| 651 | +; RV32-NEXT: ret |
| 652 | +; |
| 653 | +; RV64-LABEL: vnsrl_wx_i64_nxv1i16: |
| 654 | +; RV64: # %bb.0: |
| 655 | +; RV64-NEXT: vsetvli a1, zero, e64, m1, ta, ma |
| 656 | +; RV64-NEXT: vmv.v.x v9, a0 |
| 657 | +; RV64-NEXT: vsetvli zero, zero, e32, mf2, ta, ma |
| 658 | +; RV64-NEXT: vnsrl.wi v9, v9, 0 |
| 659 | +; RV64-NEXT: vsrl.vv v8, v8, v9 |
| 660 | +; RV64-NEXT: vsetvli zero, zero, e16, mf4, ta, ma |
| 661 | +; RV64-NEXT: vnsrl.wi v8, v8, 0 |
| 662 | +; RV64-NEXT: ret |
| 663 | + %head = insertelement <vscale x 1 x i64> poison, i64 %b, i32 0 |
| 664 | + %splat = shufflevector <vscale x 1 x i64> %head, <vscale x 1 x i64> poison, <vscale x 1 x i32> zeroinitializer |
| 665 | + %vb = trunc <vscale x 1 x i64> %splat to <vscale x 1 x i32> |
| 666 | + %x = lshr <vscale x 1 x i32> %va, %vb |
| 667 | + %y = trunc <vscale x 1 x i32> %x to <vscale x 1 x i16> |
| 668 | + ret <vscale x 1 x i16> %y |
| 669 | +} |
| 670 | + |
| 671 | +define <vscale x 1 x i8> @vnsrl_wx_i64_nxv1i8(<vscale x 1 x i16> %va, i64 %b) { |
| 672 | +; RV32-LABEL: vnsrl_wx_i64_nxv1i8: |
| 673 | +; RV32: # %bb.0: |
| 674 | +; RV32-NEXT: addi sp, sp, -16 |
| 675 | +; RV32-NEXT: .cfi_def_cfa_offset 16 |
| 676 | +; RV32-NEXT: sw a1, 12(sp) |
| 677 | +; RV32-NEXT: sw a0, 8(sp) |
| 678 | +; RV32-NEXT: addi a0, sp, 8 |
| 679 | +; RV32-NEXT: vsetvli a1, zero, e32, mf2, ta, ma |
| 680 | +; RV32-NEXT: vlse64.v v9, (a0), zero |
| 681 | +; RV32-NEXT: vnsrl.wi v9, v9, 0 |
| 682 | +; RV32-NEXT: vsetvli zero, zero, e16, mf4, ta, ma |
| 683 | +; RV32-NEXT: vnsrl.wi v9, v9, 0 |
| 684 | +; RV32-NEXT: vsrl.vv v8, v8, v9 |
| 685 | +; RV32-NEXT: vsetvli zero, zero, e8, mf8, ta, ma |
| 686 | +; RV32-NEXT: vnsrl.wi v8, v8, 0 |
| 687 | +; RV32-NEXT: addi sp, sp, 16 |
| 688 | +; RV32-NEXT: ret |
| 689 | +; |
| 690 | +; RV64-LABEL: vnsrl_wx_i64_nxv1i8: |
| 691 | +; RV64: # %bb.0: |
| 692 | +; RV64-NEXT: vsetvli a1, zero, e64, m1, ta, ma |
| 693 | +; RV64-NEXT: vmv.v.x v9, a0 |
| 694 | +; RV64-NEXT: vsetvli zero, zero, e32, mf2, ta, ma |
| 695 | +; RV64-NEXT: vnsrl.wi v9, v9, 0 |
| 696 | +; RV64-NEXT: vsetvli zero, zero, e16, mf4, ta, ma |
| 697 | +; RV64-NEXT: vnsrl.wi v9, v9, 0 |
| 698 | +; RV64-NEXT: vsrl.vv v8, v8, v9 |
| 699 | +; RV64-NEXT: vsetvli zero, zero, e8, mf8, ta, ma |
| 700 | +; RV64-NEXT: vnsrl.wi v8, v8, 0 |
| 701 | +; RV64-NEXT: ret |
| 702 | + %head = insertelement <vscale x 1 x i64> poison, i64 %b, i32 0 |
| 703 | + %splat = shufflevector <vscale x 1 x i64> %head, <vscale x 1 x i64> poison, <vscale x 1 x i32> zeroinitializer |
| 704 | + %vb = trunc <vscale x 1 x i64> %splat to <vscale x 1 x i16> |
| 705 | + %x = lshr <vscale x 1 x i16> %va, %vb |
| 706 | + %y = trunc <vscale x 1 x i16> %x to <vscale x 1 x i8> |
| 707 | + ret <vscale x 1 x i8> %y |
| 708 | +} |
0 commit comments