xref: /llvm-project/llvm/test/Transforms/AggressiveInstCombine/trunc_lshr.ll (revision 7c802f985f2c28985c233d05b559a4e8c92110ae)
1; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
2; RUN: opt < %s -passes=aggressive-instcombine -S | FileCheck %s
3
4define i16 @lshr_15(i16 %x) {
5; CHECK-LABEL: @lshr_15(
6; CHECK-NEXT:    [[LSHR:%.*]] = lshr i16 [[X:%.*]], 15
7; CHECK-NEXT:    ret i16 [[LSHR]]
8;
9  %zext = zext i16 %x to i32
10  %lshr = lshr i32 %zext, 15
11  %trunc = trunc i32 %lshr to i16
12  ret i16 %trunc
13}
14
15; Negative test
16
17define i16 @lshr_16(i16 %x) {
18; CHECK-LABEL: @lshr_16(
19; CHECK-NEXT:    [[ZEXT:%.*]] = zext i16 [[X:%.*]] to i32
20; CHECK-NEXT:    [[LSHR:%.*]] = lshr i32 [[ZEXT]], 16
21; CHECK-NEXT:    [[TRUNC:%.*]] = trunc i32 [[LSHR]] to i16
22; CHECK-NEXT:    ret i16 [[TRUNC]]
23;
24  %zext = zext i16 %x to i32
25  %lshr = lshr i32 %zext, 16
26  %trunc = trunc i32 %lshr to i16
27  ret i16 %trunc
28}
29
30; Negative test
31
32define i16 @lshr_var_shift_amount(i8 %x, i8 %amt) {
33; CHECK-LABEL: @lshr_var_shift_amount(
34; CHECK-NEXT:    [[Z:%.*]] = zext i8 [[X:%.*]] to i32
35; CHECK-NEXT:    [[ZA:%.*]] = zext i8 [[AMT:%.*]] to i32
36; CHECK-NEXT:    [[S:%.*]] = lshr i32 [[Z]], [[ZA]]
37; CHECK-NEXT:    [[A:%.*]] = add i32 [[S]], [[Z]]
38; CHECK-NEXT:    [[S2:%.*]] = lshr i32 [[A]], 2
39; CHECK-NEXT:    [[T:%.*]] = trunc i32 [[S2]] to i16
40; CHECK-NEXT:    ret i16 [[T]]
41;
42  %z = zext i8 %x to i32
43  %za = zext i8 %amt to i32
44  %s = lshr i32 %z, %za
45  %a = add i32 %s, %z
46  %s2 = lshr i32 %a, 2
47  %t = trunc i32 %s2 to i16
48  ret i16 %t
49}
50
51define i16 @lshr_var_bounded_shift_amount(i8 %x, i8 %amt) {
52; CHECK-LABEL: @lshr_var_bounded_shift_amount(
53; CHECK-NEXT:    [[Z:%.*]] = zext i8 [[X:%.*]] to i16
54; CHECK-NEXT:    [[ZA:%.*]] = zext i8 [[AMT:%.*]] to i16
55; CHECK-NEXT:    [[ZA2:%.*]] = and i16 [[ZA]], 15
56; CHECK-NEXT:    [[S:%.*]] = lshr i16 [[Z]], [[ZA2]]
57; CHECK-NEXT:    [[A:%.*]] = add i16 [[S]], [[Z]]
58; CHECK-NEXT:    [[S2:%.*]] = lshr i16 [[A]], 2
59; CHECK-NEXT:    ret i16 [[S2]]
60;
61  %z = zext i8 %x to i32
62  %za = zext i8 %amt to i32
63  %za2 = and i32 %za, 15
64  %s = lshr i32 %z, %za2
65  %a = add i32 %s, %z
66  %s2 = lshr i32 %a, 2
67  %t = trunc i32 %s2 to i16
68  ret i16 %t
69}
70
71; Negative test
72
73define i32 @lshr_check_no_overflow(i32 %x, i16 %amt) {
74; CHECK-LABEL: @lshr_check_no_overflow(
75; CHECK-NEXT:    [[ZEXT:%.*]] = zext i32 [[X:%.*]] to i64
76; CHECK-NEXT:    [[SEXT:%.*]] = sext i16 [[AMT:%.*]] to i64
77; CHECK-NEXT:    [[AND:%.*]] = and i64 [[SEXT]], 4294967295
78; CHECK-NEXT:    [[SHL:%.*]] = lshr i64 [[ZEXT]], [[AND]]
79; CHECK-NEXT:    [[TRUNC:%.*]] = trunc i64 [[SHL]] to i32
80; CHECK-NEXT:    ret i32 [[TRUNC]]
81;
82  %zext = zext i32 %x to i64
83  %sext = sext i16 %amt to i64
84  %and = and i64 %sext, 4294967295
85  %shl = lshr i64 %zext, %and
86  %trunc = trunc i64 %shl to i32
87  ret i32 %trunc
88}
89
90define void @lshr_big_dag(ptr %a, i8 %b, i8 %c) {
91; CHECK-LABEL: @lshr_big_dag(
92; CHECK-NEXT:    [[ZEXT1:%.*]] = zext i8 [[B:%.*]] to i16
93; CHECK-NEXT:    [[ZEXT2:%.*]] = zext i8 [[C:%.*]] to i16
94; CHECK-NEXT:    [[ADD1:%.*]] = add i16 [[ZEXT1]], [[ZEXT2]]
95; CHECK-NEXT:    [[SFT1:%.*]] = and i16 [[ADD1]], 15
96; CHECK-NEXT:    [[SHR1:%.*]] = lshr i16 [[ADD1]], [[SFT1]]
97; CHECK-NEXT:    [[ADD2:%.*]] = add i16 [[ADD1]], [[SHR1]]
98; CHECK-NEXT:    [[SFT2:%.*]] = and i16 [[ADD2]], 7
99; CHECK-NEXT:    [[SHR2:%.*]] = lshr i16 [[ADD2]], [[SFT2]]
100; CHECK-NEXT:    store i16 [[SHR2]], ptr [[A:%.*]], align 2
101; CHECK-NEXT:    ret void
102;
103  %zext1 = zext i8 %b to i32
104  %zext2 = zext i8 %c to i32
105  %add1 = add i32 %zext1, %zext2
106  %sft1 = and i32 %add1, 15
107  %shr1 = lshr i32 %add1, %sft1
108  %add2 = add i32 %add1, %shr1
109  %sft2 = and i32 %add2, 7
110  %shr2 = lshr i32 %add2, %sft2
111  %trunc = trunc i32 %shr2 to i16
112  store i16 %trunc, ptr %a, align 2
113  ret void
114}
115
116; Negative test
117
118define i8 @lshr_check_not_i8_trunc(i16 %x) {
119; CHECK-LABEL: @lshr_check_not_i8_trunc(
120; CHECK-NEXT:    [[LSHR:%.*]] = lshr i16 [[X:%.*]], 1
121; CHECK-NEXT:    [[ZEXT2:%.*]] = zext i16 [[LSHR]] to i32
122; CHECK-NEXT:    [[LSHR2:%.*]] = lshr i32 [[ZEXT2]], 2
123; CHECK-NEXT:    [[TRUNC:%.*]] = trunc i32 [[LSHR2]] to i8
124; CHECK-NEXT:    ret i8 [[TRUNC]]
125;
126  %lshr = lshr i16 %x, 1
127  %zext2 = zext i16 %lshr to i32
128  %lshr2 = lshr i32 %zext2, 2
129  %trunc = trunc i32 %lshr2 to i8
130  ret i8 %trunc
131}
132
133define <2 x i16> @lshr_vector(<2 x i8> %x) {
134; CHECK-LABEL: @lshr_vector(
135; CHECK-NEXT:    [[Z:%.*]] = zext <2 x i8> [[X:%.*]] to <2 x i16>
136; CHECK-NEXT:    [[ZA:%.*]] = and <2 x i16> [[Z]], <i16 7, i16 8>
137; CHECK-NEXT:    [[S:%.*]] = lshr <2 x i16> [[Z]], [[ZA]]
138; CHECK-NEXT:    [[A:%.*]] = add <2 x i16> [[S]], [[Z]]
139; CHECK-NEXT:    [[S2:%.*]] = lshr <2 x i16> [[A]], <i16 4, i16 5>
140; CHECK-NEXT:    ret <2 x i16> [[S2]]
141;
142  %z = zext <2 x i8> %x to <2 x i32>
143  %za = and <2 x i32> %z, <i32 7, i32 8>
144  %s = lshr <2 x i32> %z, %za
145  %a = add <2 x i32> %s, %z
146  %s2 = lshr <2 x i32> %a, <i32 4, i32 5>
147  %t = trunc <2 x i32> %s2 to <2 x i16>
148  ret <2 x i16> %t
149}
150
151; Negative test - can only fold to <2 x i16>, requiring new vector type
152
153define <2 x i8> @lshr_vector_no_new_vector_type(<2 x i8> %x) {
154; CHECK-LABEL: @lshr_vector_no_new_vector_type(
155; CHECK-NEXT:    [[Z:%.*]] = zext <2 x i8> [[X:%.*]] to <2 x i32>
156; CHECK-NEXT:    [[ZA:%.*]] = and <2 x i32> [[Z]], <i32 7, i32 8>
157; CHECK-NEXT:    [[S:%.*]] = lshr <2 x i32> [[Z]], [[ZA]]
158; CHECK-NEXT:    [[A:%.*]] = add <2 x i32> [[S]], [[Z]]
159; CHECK-NEXT:    [[S2:%.*]] = lshr <2 x i32> [[A]], <i32 4, i32 5>
160; CHECK-NEXT:    [[T:%.*]] = trunc <2 x i32> [[S2]] to <2 x i8>
161; CHECK-NEXT:    ret <2 x i8> [[T]]
162;
163  %z = zext <2 x i8> %x to <2 x i32>
164  %za = and <2 x i32> %z, <i32 7, i32 8>
165  %s = lshr <2 x i32> %z, %za
166  %a = add <2 x i32> %s, %z
167  %s2 = lshr <2 x i32> %a, <i32 4, i32 5>
168  %t = trunc <2 x i32> %s2 to <2 x i8>
169  ret <2 x i8> %t
170}
171
172; Negative test
173
174define <2 x i16> @lshr_vector_large_shift_amount(<2 x i8> %x) {
175; CHECK-LABEL: @lshr_vector_large_shift_amount(
176; CHECK-NEXT:    [[Z:%.*]] = zext <2 x i8> [[X:%.*]] to <2 x i32>
177; CHECK-NEXT:    [[ZA:%.*]] = and <2 x i32> [[Z]], <i32 7, i32 8>
178; CHECK-NEXT:    [[S:%.*]] = lshr <2 x i32> [[Z]], [[ZA]]
179; CHECK-NEXT:    [[A:%.*]] = add <2 x i32> [[S]], [[Z]]
180; CHECK-NEXT:    [[S2:%.*]] = lshr <2 x i32> [[A]], <i32 16, i32 5>
181; CHECK-NEXT:    [[T:%.*]] = trunc <2 x i32> [[S2]] to <2 x i16>
182; CHECK-NEXT:    ret <2 x i16> [[T]]
183;
184  %z = zext <2 x i8> %x to <2 x i32>
185  %za = and <2 x i32> %z, <i32 7, i32 8>
186  %s = lshr <2 x i32> %z, %za
187  %a = add <2 x i32> %s, %z
188  %s2 = lshr <2 x i32> %a, <i32 16, i32 5>
189  %t = trunc <2 x i32> %s2 to <2 x i16>
190  ret <2 x i16> %t
191}
192
193define i16 @lshr_exact(i16 %x) {
194; CHECK-LABEL: @lshr_exact(
195; CHECK-NEXT:    [[LSHR:%.*]] = lshr exact i16 [[X:%.*]], 15
196; CHECK-NEXT:    ret i16 [[LSHR]]
197;
198  %zext = zext i16 %x to i32
199  %lshr = lshr exact i32 %zext, 15
200  %trunc = trunc i32 %lshr to i16
201  ret i16 %trunc
202}
203
204; Negative test
205
206define i16 @lshr_negative_operand(i16 %x) {
207; CHECK-LABEL: @lshr_negative_operand(
208; CHECK-NEXT:    [[ZEXT:%.*]] = zext i16 [[X:%.*]] to i32
209; CHECK-NEXT:    [[XOR:%.*]] = xor i32 -1, [[ZEXT]]
210; CHECK-NEXT:    [[LSHR2:%.*]] = lshr i32 [[XOR]], 2
211; CHECK-NEXT:    [[TRUNC:%.*]] = trunc i32 [[LSHR2]] to i16
212; CHECK-NEXT:    ret i16 [[TRUNC]]
213;
214  %zext = zext i16 %x to i32
215  %xor = xor i32 -1, %zext
216  %lshr2 = lshr i32 %xor, 2
217  %trunc = trunc i32 %lshr2 to i16
218  ret i16 %trunc
219}
220
221; We may encounter unoptimized IR as below,
222; so don't crash by assuming that we can
223; apply instruction flags (exact) if there
224; is no instruction.
225
226define i8 @non_canonical_crash() {
227; CHECK-LABEL: @non_canonical_crash(
228; CHECK-NEXT:    ret i8 8
229;
230  %sh = lshr i32 33, 2
231  %tr = trunc i32 %sh to i8
232  ret i8 %tr
233}
234