Lines Matching +full:0 +full:xc3

7 ; CHECK:       # %bb.0: # %entry
8 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
9 ; CHECK-NEXT: cmpoxadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xe0,0x07]
10 ; CHECK-NEXT: retq # encoding: [0xc3]
13 ; EGPR: # %bb.0: # %entry
14 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
15 ; EGPR-NEXT: cmpoxadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xe0,0x07]
16 ; EGPR-NEXT: retq # encoding: [0xc3]
18 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 0)
19 ret i32 %0
26 ; CHECK: # %bb.0: # %entry
27 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
28 ; CHECK-NEXT: cmpoxadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xe0,0x07]
29 ; CHECK-NEXT: retq # encoding: [0xc3]
32 ; EGPR: # %bb.0: # %entry
33 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
34 ; EGPR-NEXT: cmpoxadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xe0,0x07]
35 ; EGPR-NEXT: retq # encoding: [0xc3]
37 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 0)
38 ret i64 %0
45 ; CHECK: # %bb.0: # %entry
46 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
47 ; CHECK-NEXT: cmpnoxadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xe1,0x07]
48 ; CHECK-NEXT: retq # encoding: [0xc3]
51 ; EGPR: # %bb.0: # %entry
52 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
53 ; EGPR-NEXT: cmpnoxadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xe1,0x07]
54 ; EGPR-NEXT: retq # encoding: [0xc3]
56 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 1)
57 ret i32 %0
62 ; CHECK: # %bb.0: # %entry
63 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
64 ; CHECK-NEXT: cmpnoxadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xe1,0x07]
65 ; CHECK-NEXT: retq # encoding: [0xc3]
68 ; EGPR: # %bb.0: # %entry
69 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
70 ; EGPR-NEXT: cmpnoxadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xe1,0x07]
71 ; EGPR-NEXT: retq # encoding: [0xc3]
73 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 1)
74 ret i64 %0
79 ; CHECK: # %bb.0: # %entry
80 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
81 ; CHECK-NEXT: cmpbxadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xe2,0x07]
82 ; CHECK-NEXT: retq # encoding: [0xc3]
85 ; EGPR: # %bb.0: # %entry
86 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
87 ; EGPR-NEXT: cmpbxadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xe2,0x07]
88 ; EGPR-NEXT: retq # encoding: [0xc3]
90 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 2)
91 ret i32 %0
96 ; CHECK: # %bb.0: # %entry
97 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
98 ; CHECK-NEXT: cmpbxadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xe2,0x07]
99 ; CHECK-NEXT: retq # encoding: [0xc3]
102 ; EGPR: # %bb.0: # %entry
103 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
104 ; EGPR-NEXT: cmpbxadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xe2,0x07]
105 ; EGPR-NEXT: retq # encoding: [0xc3]
107 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 2)
108 ret i64 %0
113 ; CHECK: # %bb.0: # %entry
114 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
115 ; CHECK-NEXT: cmpaexadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xe3,0x07]
116 ; CHECK-NEXT: retq # encoding: [0xc3]
119 ; EGPR: # %bb.0: # %entry
120 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
121 ; EGPR-NEXT: cmpaexadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xe3,0x07]
122 ; EGPR-NEXT: retq # encoding: [0xc3]
124 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 3)
125 ret i32 %0
130 ; CHECK: # %bb.0: # %entry
131 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
132 ; CHECK-NEXT: cmpaexadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xe3,0x07]
133 ; CHECK-NEXT: retq # encoding: [0xc3]
136 ; EGPR: # %bb.0: # %entry
137 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
138 ; EGPR-NEXT: cmpaexadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xe3,0x07]
139 ; EGPR-NEXT: retq # encoding: [0xc3]
141 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 3)
142 ret i64 %0
147 ; CHECK: # %bb.0: # %entry
148 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
149 ; CHECK-NEXT: cmpexadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xe4,0x07]
150 ; CHECK-NEXT: retq # encoding: [0xc3]
153 ; EGPR: # %bb.0: # %entry
154 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
155 ; EGPR-NEXT: cmpexadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xe4,0x07]
156 ; EGPR-NEXT: retq # encoding: [0xc3]
158 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 4)
159 ret i32 %0
164 ; CHECK: # %bb.0: # %entry
165 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
166 ; CHECK-NEXT: cmpexadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xe4,0x07]
167 ; CHECK-NEXT: retq # encoding: [0xc3]
170 ; EGPR: # %bb.0: # %entry
171 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
172 ; EGPR-NEXT: cmpexadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xe4,0x07]
173 ; EGPR-NEXT: retq # encoding: [0xc3]
175 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 4)
176 ret i64 %0
181 ; CHECK: # %bb.0: # %entry
182 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
183 ; CHECK-NEXT: cmpnexadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xe5,0x07]
184 ; CHECK-NEXT: retq # encoding: [0xc3]
187 ; EGPR: # %bb.0: # %entry
188 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
189 ; EGPR-NEXT: cmpnexadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xe5,0x07]
190 ; EGPR-NEXT: retq # encoding: [0xc3]
192 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 5)
193 ret i32 %0
198 ; CHECK: # %bb.0: # %entry
199 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
200 ; CHECK-NEXT: cmpnexadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xe5,0x07]
201 ; CHECK-NEXT: retq # encoding: [0xc3]
204 ; EGPR: # %bb.0: # %entry
205 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
206 ; EGPR-NEXT: cmpnexadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xe5,0x07]
207 ; EGPR-NEXT: retq # encoding: [0xc3]
209 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 5)
210 ret i64 %0
215 ; CHECK: # %bb.0: # %entry
216 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
217 ; CHECK-NEXT: cmpbexadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xe6,0x07]
218 ; CHECK-NEXT: retq # encoding: [0xc3]
221 ; EGPR: # %bb.0: # %entry
222 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
223 ; EGPR-NEXT: cmpbexadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xe6,0x07]
224 ; EGPR-NEXT: retq # encoding: [0xc3]
226 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 6)
227 ret i32 %0
232 ; CHECK: # %bb.0: # %entry
233 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
234 ; CHECK-NEXT: cmpbexadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xe6,0x07]
235 ; CHECK-NEXT: retq # encoding: [0xc3]
238 ; EGPR: # %bb.0: # %entry
239 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
240 ; EGPR-NEXT: cmpbexadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xe6,0x07]
241 ; EGPR-NEXT: retq # encoding: [0xc3]
243 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 6)
244 ret i64 %0
249 ; CHECK: # %bb.0: # %entry
250 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
251 ; CHECK-NEXT: cmpaxadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xe7,0x07]
252 ; CHECK-NEXT: retq # encoding: [0xc3]
255 ; EGPR: # %bb.0: # %entry
256 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
257 ; EGPR-NEXT: cmpaxadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xe7,0x07]
258 ; EGPR-NEXT: retq # encoding: [0xc3]
260 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 7)
261 ret i32 %0
266 ; CHECK: # %bb.0: # %entry
267 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
268 ; CHECK-NEXT: cmpaxadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xe7,0x07]
269 ; CHECK-NEXT: retq # encoding: [0xc3]
272 ; EGPR: # %bb.0: # %entry
273 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
274 ; EGPR-NEXT: cmpaxadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xe7,0x07]
275 ; EGPR-NEXT: retq # encoding: [0xc3]
277 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 7)
278 ret i64 %0
283 ; CHECK: # %bb.0: # %entry
284 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
285 ; CHECK-NEXT: cmpsxadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xe8,0x07]
286 ; CHECK-NEXT: retq # encoding: [0xc3]
289 ; EGPR: # %bb.0: # %entry
290 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
291 ; EGPR-NEXT: cmpsxadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xe8,0x07]
292 ; EGPR-NEXT: retq # encoding: [0xc3]
294 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 8)
295 ret i32 %0
300 ; CHECK: # %bb.0: # %entry
301 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
302 ; CHECK-NEXT: cmpsxadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xe8,0x07]
303 ; CHECK-NEXT: retq # encoding: [0xc3]
306 ; EGPR: # %bb.0: # %entry
307 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
308 ; EGPR-NEXT: cmpsxadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xe8,0x07]
309 ; EGPR-NEXT: retq # encoding: [0xc3]
311 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 8)
312 ret i64 %0
317 ; CHECK: # %bb.0: # %entry
318 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
319 ; CHECK-NEXT: cmpnsxadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xe9,0x07]
320 ; CHECK-NEXT: retq # encoding: [0xc3]
323 ; EGPR: # %bb.0: # %entry
324 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
325 ; EGPR-NEXT: cmpnsxadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xe9,0x07]
326 ; EGPR-NEXT: retq # encoding: [0xc3]
328 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 9)
329 ret i32 %0
334 ; CHECK: # %bb.0: # %entry
335 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
336 ; CHECK-NEXT: cmpnsxadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xe9,0x07]
337 ; CHECK-NEXT: retq # encoding: [0xc3]
340 ; EGPR: # %bb.0: # %entry
341 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
342 ; EGPR-NEXT: cmpnsxadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xe9,0x07]
343 ; EGPR-NEXT: retq # encoding: [0xc3]
345 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 9)
346 ret i64 %0
351 ; CHECK: # %bb.0: # %entry
352 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
353 ; CHECK-NEXT: cmppxadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xea,0x07]
354 ; CHECK-NEXT: retq # encoding: [0xc3]
357 ; EGPR: # %bb.0: # %entry
358 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
359 ; EGPR-NEXT: cmppxadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xea,0x07]
360 ; EGPR-NEXT: retq # encoding: [0xc3]
362 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 10)
363 ret i32 %0
368 ; CHECK: # %bb.0: # %entry
369 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
370 ; CHECK-NEXT: cmppxadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xea,0x07]
371 ; CHECK-NEXT: retq # encoding: [0xc3]
374 ; EGPR: # %bb.0: # %entry
375 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
376 ; EGPR-NEXT: cmppxadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xea,0x07]
377 ; EGPR-NEXT: retq # encoding: [0xc3]
379 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 10)
380 ret i64 %0
385 ; CHECK: # %bb.0: # %entry
386 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
387 ; CHECK-NEXT: cmpnpxadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xeb,0x07]
388 ; CHECK-NEXT: retq # encoding: [0xc3]
391 ; EGPR: # %bb.0: # %entry
392 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
393 ; EGPR-NEXT: cmpnpxadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xeb,0x07]
394 ; EGPR-NEXT: retq # encoding: [0xc3]
396 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 11)
397 ret i32 %0
402 ; CHECK: # %bb.0: # %entry
403 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
404 ; CHECK-NEXT: cmpnpxadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xeb,0x07]
405 ; CHECK-NEXT: retq # encoding: [0xc3]
408 ; EGPR: # %bb.0: # %entry
409 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
410 ; EGPR-NEXT: cmpnpxadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xeb,0x07]
411 ; EGPR-NEXT: retq # encoding: [0xc3]
413 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 11)
414 ret i64 %0
419 ; CHECK: # %bb.0: # %entry
420 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
421 ; CHECK-NEXT: cmplxadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xec,0x07]
422 ; CHECK-NEXT: retq # encoding: [0xc3]
425 ; EGPR: # %bb.0: # %entry
426 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
427 ; EGPR-NEXT: cmplxadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xec,0x07]
428 ; EGPR-NEXT: retq # encoding: [0xc3]
430 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 12)
431 ret i32 %0
436 ; CHECK: # %bb.0: # %entry
437 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
438 ; CHECK-NEXT: cmplxadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xec,0x07]
439 ; CHECK-NEXT: retq # encoding: [0xc3]
442 ; EGPR: # %bb.0: # %entry
443 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
444 ; EGPR-NEXT: cmplxadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xec,0x07]
445 ; EGPR-NEXT: retq # encoding: [0xc3]
447 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 12)
448 ret i64 %0
453 ; CHECK: # %bb.0: # %entry
454 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
455 ; CHECK-NEXT: cmpgexadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xed,0x07]
456 ; CHECK-NEXT: retq # encoding: [0xc3]
459 ; EGPR: # %bb.0: # %entry
460 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
461 ; EGPR-NEXT: cmpgexadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xed,0x07]
462 ; EGPR-NEXT: retq # encoding: [0xc3]
464 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 13)
465 ret i32 %0
470 ; CHECK: # %bb.0: # %entry
471 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
472 ; CHECK-NEXT: cmpgexadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xed,0x07]
473 ; CHECK-NEXT: retq # encoding: [0xc3]
476 ; EGPR: # %bb.0: # %entry
477 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
478 ; EGPR-NEXT: cmpgexadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xed,0x07]
479 ; EGPR-NEXT: retq # encoding: [0xc3]
481 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 13)
482 ret i64 %0
487 ; CHECK: # %bb.0: # %entry
488 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
489 ; CHECK-NEXT: cmplexadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xee,0x07]
490 ; CHECK-NEXT: retq # encoding: [0xc3]
493 ; EGPR: # %bb.0: # %entry
494 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
495 ; EGPR-NEXT: cmplexadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xee,0x07]
496 ; EGPR-NEXT: retq # encoding: [0xc3]
498 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 14)
499 ret i32 %0
504 ; CHECK: # %bb.0: # %entry
505 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
506 ; CHECK-NEXT: cmplexadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xee,0x07]
507 ; CHECK-NEXT: retq # encoding: [0xc3]
510 ; EGPR: # %bb.0: # %entry
511 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
512 ; EGPR-NEXT: cmplexadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xee,0x07]
513 ; EGPR-NEXT: retq # encoding: [0xc3]
515 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 14)
516 ret i64 %0
521 ; CHECK: # %bb.0: # %entry
522 ; CHECK-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
523 ; CHECK-NEXT: cmpgxadd %edx, %eax, (%rdi) # encoding: [0xc4,0xe2,0x69,0xef,0x07]
524 ; CHECK-NEXT: retq # encoding: [0xc3]
527 ; EGPR: # %bb.0: # %entry
528 ; EGPR-NEXT: movl %esi, %eax # encoding: [0x89,0xf0]
529 ; EGPR-NEXT: cmpgxadd %edx, %eax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0x69,0xef,0x07]
530 ; EGPR-NEXT: retq # encoding: [0xc3]
532 %0 = tail call i32 @llvm.x86.cmpccxadd32(ptr %__A, i32 %__B, i32 %__C, i32 15)
533 ret i32 %0
538 ; CHECK: # %bb.0: # %entry
539 ; CHECK-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
540 ; CHECK-NEXT: cmpgxadd %rdx, %rax, (%rdi) # encoding: [0xc4,0xe2,0xe9,0xef,0x07]
541 ; CHECK-NEXT: retq # encoding: [0xc3]
544 ; EGPR: # %bb.0: # %entry
545 ; EGPR-NEXT: movq %rsi, %rax # encoding: [0x48,0x89,0xf0]
546 ; EGPR-NEXT: cmpgxadd %rdx, %rax, (%rdi) # EVEX TO VEX Compression encoding: [0xc4,0xe2,0xe9,0xef,0x07]
547 ; EGPR-NEXT: retq # encoding: [0xc3]
549 %0 = tail call i64 @llvm.x86.cmpccxadd64(ptr %__A, i64 %__B, i64 %__C, i32 15)
550 ret i64 %0