xref: /llvm-project/llvm/test/Transforms/InstCombine/AArch64/tbl1.ll (revision ebbfdca586d5543c13617b15d0cdf5b7fdc4fd4a)
1; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
2; RUN: opt < %s -passes=instcombine -S | FileCheck %s
3
4target datalayout = "e-m:e-i8:8:32-i16:16:32-i64:64-i128:128-n32:64-S128"
5target triple = "aarch64"
6
7; Turning a table lookup intrinsic into a shuffle vector instruction
8; can be beneficial. If the mask used for the lookup is the constant
9; vector {7,6,5,4,3,2,1,0}, then the back-end generates rev64
10; instructions instead.
11
12define <8 x i8> @tbl1_8x8(<16 x i8> %vec) {
13; CHECK-LABEL: @tbl1_8x8(
14; CHECK-NEXT:  entry:
15; CHECK-NEXT:    [[TMP0:%.*]] = shufflevector <16 x i8> [[VEC:%.*]], <16 x i8> poison, <8 x i32> <i32 7, i32 6, i32 5, i32 4, i32 3, i32 2, i32 1, i32 0>
16; CHECK-NEXT:    ret <8 x i8> [[TMP0]]
17;
18entry:
19  %tbl1 = call <8 x i8> @llvm.aarch64.neon.tbl1.v8i8(<16 x i8> %vec, <8 x i8> <i8 7, i8 6, i8 5, i8 4, i8 3, i8 2, i8 1, i8 0>)
20  ret <8 x i8> %tbl1
21}
22
23; Bail the optimization if a mask index is out of range.
24define <8 x i8> @tbl1_8x8_out_of_range(<16 x i8> %vec) {
25; CHECK-LABEL: @tbl1_8x8_out_of_range(
26; CHECK-NEXT:  entry:
27; CHECK-NEXT:    [[TBL1:%.*]] = call <8 x i8> @llvm.aarch64.neon.tbl1.v8i8(<16 x i8> [[VEC:%.*]], <8 x i8> <i8 8, i8 6, i8 5, i8 4, i8 3, i8 2, i8 1, i8 0>)
28; CHECK-NEXT:    ret <8 x i8> [[TBL1]]
29;
30entry:
31  %tbl1 = call <8 x i8> @llvm.aarch64.neon.tbl1.v8i8(<16 x i8> %vec, <8 x i8> <i8 8, i8 6, i8 5, i8 4, i8 3, i8 2, i8 1, i8 0>)
32  ret <8 x i8> %tbl1
33}
34
35; Bail the optimization if the size of the return vector is not 8 elements.
36define <16 x i8> @tbl1_16x8(<16 x i8> %vec) {
37; CHECK-LABEL: @tbl1_16x8(
38; CHECK-NEXT:  entry:
39; CHECK-NEXT:    [[TBL1:%.*]] = call <16 x i8> @llvm.aarch64.neon.tbl1.v16i8(<16 x i8> [[VEC:%.*]], <16 x i8> <i8 15, i8 14, i8 13, i8 12, i8 11, i8 10, i8 9, i8 8, i8 7, i8 6, i8 5, i8 4, i8 3, i8 2, i8 1, i8 0>)
40; CHECK-NEXT:    ret <16 x i8> [[TBL1]]
41;
42entry:
43  %tbl1 = call <16 x i8> @llvm.aarch64.neon.tbl1.v16i8(<16 x i8> %vec, <16 x i8> <i8 15, i8 14, i8 13, i8 12, i8 11, i8 10, i8 9, i8 8, i8 7, i8 6, i8 5, i8 4, i8 3, i8 2, i8 1, i8 0>)
44  ret <16 x i8> %tbl1
45}
46
47; Bail the optimization if the elements of the return vector are not of type i8.
48define <8 x i16> @tbl1_8x16(<16 x i8> %vec) {
49; CHECK-LABEL: @tbl1_8x16(
50; CHECK-NEXT:  entry:
51; CHECK-NEXT:    [[TBL1:%.*]] = call <8 x i16> @llvm.aarch64.neon.tbl1.v8i16(<16 x i8> [[VEC:%.*]], <8 x i16> <i16 0, i16 1, i16 2, i16 3, i16 4, i16 5, i16 6, i16 7>)
52; CHECK-NEXT:    ret <8 x i16> [[TBL1]]
53;
54entry:
55  %tbl1 = call <8 x i16> @llvm.aarch64.neon.tbl1.v8i16(<16 x i8> %vec, <8 x i16> <i16 0, i16 1, i16 2, i16 3, i16 4, i16 5, i16 6, i16 7>)
56  ret <8 x i16> %tbl1
57}
58
59; The type <8 x i16> is not a valid return type for this intrinsic,
60; but we want to test that the optimization won't trigger for vector
61; elements of type different than i8.
62declare <8 x i16> @llvm.aarch64.neon.tbl1.v8i16(<16 x i8>, <8 x i16>)
63
64declare <8 x i8> @llvm.aarch64.neon.tbl1.v8i8(<16 x i8>, <8 x i8>)
65declare <16 x i8> @llvm.aarch64.neon.tbl1.v16i8(<16 x i8>, <16 x i8>)
66