|
| 1 | +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py |
| 2 | +; RUN: llc -mtriple=aarch64-linux-gnu -mattr=+sve2p2 < %s | FileCheck %s --check-prefixes=CHECK |
| 3 | +; RUN: llc -mtriple=aarch64-linux-gnu -mattr=+sme2p2 -force-streaming < %s | FileCheck %s --check-prefixes=CHECK |
| 4 | + |
| 5 | +; |
| 6 | +; COMPACT |
| 7 | +; |
| 8 | + |
| 9 | +define <vscale x 16 x i8> @compact_i8(<vscale x 16 x i1> %pg, <vscale x 16 x i8> %a) { |
| 10 | +; CHECK-LABEL: compact_i8: |
| 11 | +; CHECK: // %bb.0: |
| 12 | +; CHECK-NEXT: compact z0.b, p0, z0.b |
| 13 | +; CHECK-NEXT: ret |
| 14 | + %out = call <vscale x 16 x i8> @llvm.aarch64.sve.compact.nxv16i8(<vscale x 16 x i1> %pg, |
| 15 | + <vscale x 16 x i8> %a) |
| 16 | + ret <vscale x 16 x i8> %out |
| 17 | +} |
| 18 | + |
| 19 | +define <vscale x 8 x i16> @compact_i16(<vscale x 8 x i1> %pg, <vscale x 8 x i16> %a) { |
| 20 | +; CHECK-LABEL: compact_i16: |
| 21 | +; CHECK: // %bb.0: |
| 22 | +; CHECK-NEXT: compact z0.h, p0, z0.h |
| 23 | +; CHECK-NEXT: ret |
| 24 | + %out = call <vscale x 8 x i16> @llvm.aarch64.sve.compact.nxv8i16(<vscale x 8 x i1> %pg, |
| 25 | + <vscale x 8 x i16> %a) |
| 26 | + ret <vscale x 8 x i16> %out |
| 27 | +} |
| 28 | + |
| 29 | +define <vscale x 8 x half> @compact_f16(<vscale x 8 x i1> %pg, <vscale x 8 x half> %a) { |
| 30 | +; CHECK-LABEL: compact_f16: |
| 31 | +; CHECK: // %bb.0: |
| 32 | +; CHECK-NEXT: compact z0.h, p0, z0.h |
| 33 | +; CHECK-NEXT: ret |
| 34 | + %out = call <vscale x 8 x half> @llvm.aarch64.sve.compact.nxv8f16(<vscale x 8 x i1> %pg, |
| 35 | + <vscale x 8 x half> %a) |
| 36 | + ret <vscale x 8 x half> %out |
| 37 | +} |
| 38 | + |
| 39 | +define <vscale x 8 x bfloat> @compact_bf16(<vscale x 8 x i1> %pg, <vscale x 8 x bfloat> %a) { |
| 40 | +; CHECK-LABEL: compact_bf16: |
| 41 | +; CHECK: // %bb.0: |
| 42 | +; CHECK-NEXT: compact z0.h, p0, z0.h |
| 43 | +; CHECK-NEXT: ret |
| 44 | + %out = call <vscale x 8 x bfloat> @llvm.aarch64.sve.compact.nxv8bf16(<vscale x 8 x i1> %pg, |
| 45 | + <vscale x 8 x bfloat> %a) |
| 46 | + ret <vscale x 8 x bfloat> %out |
| 47 | +} |
| 48 | + |
| 49 | +declare <vscale x 16 x i8> @llvm.aarch64.sve.compact.nxv16i8(<vscale x 16 x i1>, <vscale x 16 x i8>) |
| 50 | +declare <vscale x 8 x i16> @llvm.aarch64.sve.compact.nxv8i16(<vscale x 8 x i1>, <vscale x 8 x i16>) |
| 51 | +declare <vscale x 8 x half> @llvm.aarch64.sve.compact.nxv8f16(<vscale x 8 x i1>, <vscale x 8 x half>) |
| 52 | +declare <vscale x 8 x bfloat> @llvm.aarch64.sve.compact.nxv8bf16(<vscale x 8 x i1>, <vscale x 8 x bfloat>) |
0 commit comments