|
| 1 | +; NOTE: Assertions have been autogenerated by utils/update_test_checks.py |
| 2 | +; RUN: opt < %s -passes=slp-vectorizer -S -mtriple=riscv64-unknown-linux -mattr=+v -slp-threshold=-100 | FileCheck %s |
| 3 | + |
| 4 | +define i64 @test(ptr %p) { |
| 5 | +; CHECK-LABEL: @test( |
| 6 | +; CHECK-NEXT: entry: |
| 7 | +; CHECK-NEXT: [[ARRAYIDX_4:%.*]] = getelementptr inbounds i64, ptr [[P:%.*]], i64 4 |
| 8 | +; CHECK-NEXT: [[TMP0:%.*]] = load <4 x i64>, ptr [[P]], align 4 |
| 9 | +; CHECK-NEXT: [[TMP1:%.*]] = load <2 x i64>, ptr [[ARRAYIDX_4]], align 4 |
| 10 | +; CHECK-NEXT: [[TMP2:%.*]] = shufflevector <4 x i64> [[TMP0]], <4 x i64> poison, <8 x i32> <i32 poison, i32 poison, i32 poison, i32 poison, i32 poison, i32 poison, i32 0, i32 0> |
| 11 | +; CHECK-NEXT: [[TMP3:%.*]] = call <8 x i64> @llvm.vector.insert.v8i64.v4i64(<8 x i64> [[TMP2]], <4 x i64> [[TMP0]], i64 0) |
| 12 | +; CHECK-NEXT: [[TMP4:%.*]] = call <8 x i64> @llvm.vector.insert.v8i64.v2i64(<8 x i64> [[TMP3]], <2 x i64> [[TMP1]], i64 4) |
| 13 | +; CHECK-NEXT: [[TMP5:%.*]] = mul <8 x i64> [[TMP4]], <i64 42, i64 42, i64 42, i64 42, i64 42, i64 42, i64 42, i64 42> |
| 14 | +; CHECK-NEXT: [[TMP6:%.*]] = call i64 @llvm.vector.reduce.add.v8i64(<8 x i64> [[TMP5]]) |
| 15 | +; CHECK-NEXT: ret i64 [[TMP6]] |
| 16 | +; |
| 17 | +entry: |
| 18 | + %arrayidx.1 = getelementptr inbounds i64, ptr %p, i64 1 |
| 19 | + %arrayidx.2 = getelementptr inbounds i64, ptr %p, i64 2 |
| 20 | + %arrayidx.3 = getelementptr inbounds i64, ptr %p, i64 3 |
| 21 | + %arrayidx.4 = getelementptr inbounds i64, ptr %p, i64 4 |
| 22 | + %arrayidx.5 = getelementptr inbounds i64, ptr %p, i64 5 |
| 23 | + %tmp = load i64, ptr %p, align 4 |
| 24 | + %mul = mul i64 %tmp, 42 |
| 25 | + %tmp1 = load i64, ptr %arrayidx.1, align 4 |
| 26 | + %mul1 = mul i64 %tmp1, 42 |
| 27 | + %add = add i64 %mul, %mul1 |
| 28 | + %tmp2 = load i64, ptr %arrayidx.2, align 4 |
| 29 | + %mul2 = mul i64 %tmp2, 42 |
| 30 | + %add1 = add i64 %add, %mul2 |
| 31 | + %tmp3 = load i64, ptr %arrayidx.3, align 4 |
| 32 | + %mul3 = mul i64 %tmp3, 42 |
| 33 | + %add2 = add i64 %add1, %mul3 |
| 34 | + %tmp4 = load i64, ptr %arrayidx.4, align 4 |
| 35 | + %mul4 = mul i64 %tmp4, 42 |
| 36 | + %add3 = add i64 %add2, %mul4 |
| 37 | + %tmp5 = load i64, ptr %arrayidx.5, align 4 |
| 38 | + %mul5 = mul i64 %tmp5, 42 |
| 39 | + %add4 = add i64 %add3, %mul5 |
| 40 | + %mul6 = mul i64 %tmp, 42 |
| 41 | + %add5 = add i64 %add4, %mul6 |
| 42 | + %mul7 = mul i64 %tmp, 42 |
| 43 | + %add6 = add i64 %add5, %mul7 |
| 44 | + ret i64 %add6 |
| 45 | +} |
0 commit comments