Lines Matching defs:b

17 // CHECK-A32: store [2 x i64] %b.coerce, ptr %coerce.dive, align 8
32 void test_vst1_f16_x2(float16_t *a, float16x4x2_t b) {
33 vst1_f16_x2(a, b);
41 // CHECK-A32: store [3 x i64] %b.coerce, ptr %coerce.dive, align 8
61 void test_vst1_f16_x3(float16_t *a, float16x4x3_t b) {
62 vst1_f16_x3(a, b);
70 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
95 void test_vst1_f16_x4(float16_t *a, float16x4x4_t b) {
96 vst1_f16_x4(a, b);
104 // CHECK-A32: store [2 x i64] %b.coerce, ptr %coerce.dive, align 8
119 void test_vst1_f32_x2(float32_t *a, float32x2x2_t b) {
120 vst1_f32_x2(a, b);
128 // CHECK-A32: store [3 x i64] %b.coerce, ptr %coerce.dive, align 8
148 void test_vst1_f32_x3(float32_t *a, float32x2x3_t b) {
149 vst1_f32_x3(a, b);
157 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
182 void test_vst1_f32_x4(float32_t *a, float32x2x4_t b) {
183 vst1_f32_x4(a, b);
191 // CHECK-A32: store [2 x i64] %b.coerce, ptr %coerce.dive, align 8
206 void test_vst1_p16_x2(poly16_t *a, poly16x4x2_t b) {
207 vst1_p16_x2(a, b);
215 // CHECK-A32: store [3 x i64] %b.coerce, ptr %coerce.dive, align 8
235 void test_vst1_p16_x3(poly16_t *a, poly16x4x3_t b) {
236 vst1_p16_x3(a, b);
244 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
269 void test_vst1_p16_x4(poly16_t *a, poly16x4x4_t b) {
270 vst1_p16_x4(a, b);
278 // CHECK-A32: store [2 x i64] %b.coerce, ptr %coerce.dive, align 8
289 void test_vst1_p8_x2(poly8_t *a, poly8x8x2_t b) {
290 vst1_p8_x2(a, b);
298 // CHECK-A32: store [3 x i64] %b.coerce, ptr %coerce.dive, align 8
312 void test_vst1_p8_x3(poly8_t *a, poly8x8x3_t b) {
313 vst1_p8_x3(a, b);
321 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
338 void test_vst1_p8_x4(poly8_t *a, poly8x8x4_t b) {
339 vst1_p8_x4(a, b);
347 // CHECK-A32: store [2 x i64] %b.coerce, ptr %coerce.dive, align 8
362 void test_vst1_s16_x2(int16_t *a, int16x4x2_t b) {
363 vst1_s16_x2(a, b);
371 // CHECK-A32: store [3 x i64] %b.coerce, ptr %coerce.dive, align 8
391 void test_vst1_s16_x3(int16_t *a, int16x4x3_t b) {
392 vst1_s16_x3(a, b);
400 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
425 void test_vst1_s16_x4(int16_t *a, int16x4x4_t b) {
426 vst1_s16_x4(a, b);
434 // CHECK-A32: store [2 x i64] %b.coerce, ptr %coerce.dive, align 8
449 void test_vst1_s32_x2(int32_t *a, int32x2x2_t b) {
450 vst1_s32_x2(a, b);
458 // CHECK-A32: store [3 x i64] %b.coerce, ptr %coerce.dive, align 8
478 void test_vst1_s32_x3(int32_t *a, int32x2x3_t b) {
479 vst1_s32_x3(a, b);
487 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
512 void test_vst1_s32_x4(int32_t *a, int32x2x4_t b) {
513 vst1_s32_x4(a, b);
521 // CHECK-A32: store [2 x i64] %b.coerce, ptr %coerce.dive, align 8
536 void test_vst1_s64_x2(int64_t *a, int64x1x2_t b) {
537 vst1_s64_x2(a, b);
545 // CHECK-A32: store [3 x i64] %b.coerce, ptr %coerce.dive, align 8
565 void test_vst1_s64_x3(int64_t *a, int64x1x3_t b) {
566 vst1_s64_x3(a, b);
574 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
599 void test_vst1_s64_x4(int64_t *a, int64x1x4_t b) {
600 vst1_s64_x4(a, b);
608 // CHECK-A32: store [2 x i64] %b.coerce, ptr %coerce.dive, align 8
619 void test_vst1_s8_x2(int8_t *a, int8x8x2_t b) {
620 vst1_s8_x2(a, b);
628 // CHECK-A32: store [3 x i64] %b.coerce, ptr %coerce.dive, align 8
642 void test_vst1_s8_x3(int8_t *a, int8x8x3_t b) {
643 vst1_s8_x3(a, b);
651 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
668 void test_vst1_s8_x4(int8_t *a, int8x8x4_t b) {
669 vst1_s8_x4(a, b);
677 // CHECK-A32: store [2 x i64] %b.coerce, ptr %coerce.dive, align 8
692 void test_vst1_u16_x2(uint16_t *a, uint16x4x2_t b) {
693 vst1_u16_x2(a, b);
701 // CHECK-A32: store [3 x i64] %b.coerce, ptr %coerce.dive, align 8
721 void test_vst1_u16_x3(uint16_t *a, uint16x4x3_t b) {
722 vst1_u16_x3(a, b);
730 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
755 void test_vst1_u16_x4(uint16_t *a, uint16x4x4_t b) {
756 vst1_u16_x4(a, b);
764 // CHECK-A32: store [2 x i64] %b.coerce, ptr %coerce.dive, align 8
779 void test_vst1_u32_x2(uint32_t *a, uint32x2x2_t b) {
780 vst1_u32_x2(a, b);
788 // CHECK-A32: store [3 x i64] %b.coerce, ptr %coerce.dive, align 8
808 void test_vst1_u32_x3(uint32_t *a, uint32x2x3_t b) {
809 vst1_u32_x3(a, b);
817 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
842 void test_vst1_u32_x4(uint32_t *a, uint32x2x4_t b) {
843 vst1_u32_x4(a, b);
851 // CHECK-A32: store [2 x i64] %b.coerce, ptr %coerce.dive, align 8
866 void test_vst1_u64_x2(uint64_t *a, uint64x1x2_t b) {
867 vst1_u64_x2(a, b);
875 // CHECK-A32: store [3 x i64] %b.coerce, ptr %coerce.dive, align 8
895 void test_vst1_u64_x3(uint64_t *a, uint64x1x3_t b) {
896 vst1_u64_x3(a, b);
904 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
929 void test_vst1_u64_x4(uint64_t *a, uint64x1x4_t b) {
930 vst1_u64_x4(a, b);
938 // CHECK-A32: store [2 x i64] %b.coerce, ptr %coerce.dive, align 8
949 void test_vst1_u8_x2(uint8_t *a, uint8x8x2_t b) {
950 vst1_u8_x2(a, b);
958 // CHECK-A32: store [3 x i64] %b.coerce, ptr %coerce.dive, align 8
972 void test_vst1_u8_x3(uint8_t *a, uint8x8x3_t b) {
973 vst1_u8_x3(a, b);
981 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
998 void test_vst1_u8_x4(uint8_t *a, uint8x8x4_t b) {
999 vst1_u8_x4(a, b);
1007 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
1022 void test_vst1q_f16_x2(float16_t *a, float16x8x2_t b) {
1023 vst1q_f16_x2(a, b);
1031 // CHECK-A32: store [6 x i64] %b.coerce, ptr %coerce.dive, align 8
1051 void test_vst1q_f16_x3(float16_t *a, float16x8x3_t b) {
1052 vst1q_f16_x3(a, b);
1060 // CHECK-A32: store [8 x i64] %b.coerce, ptr %coerce.dive, align 8
1085 void test_vst1q_f16_x4(float16_t *a, float16x8x4_t b) {
1086 vst1q_f16_x4(a, b);
1094 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
1109 void test_vst1q_f32_x2(float32_t *a, float32x4x2_t b) {
1110 vst1q_f32_x2(a, b);
1118 // CHECK-A32: store [6 x i64] %b.coerce, ptr %coerce.dive, align 8
1138 void test_vst1q_f32_x3(float32_t *a, float32x4x3_t b) {
1139 vst1q_f32_x3(a, b);
1147 // CHECK-A32: store [8 x i64] %b.coerce, ptr %coerce.dive, align 8
1172 void test_vst1q_f32_x4(float32_t *a, float32x4x4_t b) {
1173 vst1q_f32_x4(a, b);
1181 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
1196 void test_vst1q_p16_x2(poly16_t *a, poly16x8x2_t b) {
1197 vst1q_p16_x2(a, b);
1205 // CHECK-A32: store [6 x i64] %b.coerce, ptr %coerce.dive, align 8
1225 void test_vst1q_p16_x3(poly16_t *a, poly16x8x3_t b) {
1226 vst1q_p16_x3(a, b);
1234 // CHECK-A32: store [8 x i64] %b.coerce, ptr %coerce.dive, align 8
1259 void test_vst1q_p16_x4(poly16_t *a, poly16x8x4_t b) {
1260 vst1q_p16_x4(a, b);
1268 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
1279 void test_vst1q_p8_x2(poly8_t *a, poly8x16x2_t b) {
1280 vst1q_p8_x2(a, b);
1288 // CHECK-A32: store [6 x i64] %b.coerce, ptr %coerce.dive, align 8
1302 void test_vst1q_p8_x3(poly8_t *a, poly8x16x3_t b) {
1303 vst1q_p8_x3(a, b);
1311 // CHECK-A32: store [8 x i64] %b.coerce, ptr %coerce.dive, align 8
1328 void test_vst1q_p8_x4(poly8_t *a, poly8x16x4_t b) {
1329 vst1q_p8_x4(a, b);
1337 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
1352 void test_vst1q_s16_x2(int16_t *a, int16x8x2_t b) {
1353 vst1q_s16_x2(a, b);
1361 // CHECK-A32: store [6 x i64] %b.coerce, ptr %coerce.dive, align 8
1381 void test_vst1q_s16_x3(int16_t *a, int16x8x3_t b) {
1382 vst1q_s16_x3(a, b);
1390 // CHECK-A32: store [8 x i64] %b.coerce, ptr %coerce.dive, align 8
1415 void test_vst1q_s16_x4(int16_t *a, int16x8x4_t b) {
1416 vst1q_s16_x4(a, b);
1424 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
1439 void test_vst1q_s32_x2(int32_t *a, int32x4x2_t b) {
1440 vst1q_s32_x2(a, b);
1448 // CHECK-A32: store [6 x i64] %b.coerce, ptr %coerce.dive, align 8
1468 void test_vst1q_s32_x3(int32_t *a, int32x4x3_t b) {
1469 vst1q_s32_x3(a, b);
1477 // CHECK-A32: store [8 x i64] %b.coerce, ptr %coerce.dive, align 8
1502 void test_vst1q_s32_x4(int32_t *a, int32x4x4_t b) {
1503 vst1q_s32_x4(a, b);
1511 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
1526 void test_vst1q_s64_x2(int64_t *a, int64x2x2_t b) {
1527 vst1q_s64_x2(a, b);
1535 // CHECK-A32: store [6 x i64] %b.coerce, ptr %coerce.dive, align 8
1555 void test_vst1q_s64_x3(int64_t *a, int64x2x3_t b) {
1556 vst1q_s64_x3(a, b);
1564 // CHECK-A32: store [8 x i64] %b.coerce, ptr %coerce.dive, align 8
1589 void test_vst1q_s64_x4(int64_t *a, int64x2x4_t b) {
1590 vst1q_s64_x4(a, b);
1598 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
1609 void test_vst1q_s8_x2(int8_t *a, int8x16x2_t b) {
1610 vst1q_s8_x2(a, b);
1618 // CHECK-A32: store [6 x i64] %b.coerce, ptr %coerce.dive, align 8
1632 void test_vst1q_s8_x3(int8_t *a, int8x16x3_t b) {
1633 vst1q_s8_x3(a, b);
1641 // CHECK-A32: store [8 x i64] %b.coerce, ptr %coerce.dive, align 8
1658 void test_vst1q_s8_x4(int8_t *a, int8x16x4_t b) {
1659 vst1q_s8_x4(a, b);
1667 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
1682 void test_vst1q_u16_x2(uint16_t *a, uint16x8x2_t b) {
1683 vst1q_u16_x2(a, b);
1691 // CHECK-A32: store [6 x i64] %b.coerce, ptr %coerce.dive, align 8
1711 void test_vst1q_u16_x3(uint16_t *a, uint16x8x3_t b) {
1712 vst1q_u16_x3(a, b);
1720 // CHECK-A32: store [8 x i64] %b.coerce, ptr %coerce.dive, align 8
1745 void test_vst1q_u16_x4(uint16_t *a, uint16x8x4_t b) {
1746 vst1q_u16_x4(a, b);
1754 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
1769 void test_vst1q_u32_x2(uint32_t *a, uint32x4x2_t b) {
1770 vst1q_u32_x2(a, b);
1778 // CHECK-A32: store [6 x i64] %b.coerce, ptr %coerce.dive, align 8
1798 void test_vst1q_u32_x3(uint32_t *a, uint32x4x3_t b) {
1799 vst1q_u32_x3(a, b);
1807 // CHECK-A32: store [8 x i64] %b.coerce, ptr %coerce.dive, align 8
1832 void test_vst1q_u32_x4(uint32_t *a, uint32x4x4_t b) {
1833 vst1q_u32_x4(a, b);
1841 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
1856 void test_vst1q_u64_x2(uint64_t *a, uint64x2x2_t b) {
1857 vst1q_u64_x2(a, b);
1865 // CHECK-A32: store [6 x i64] %b.coerce, ptr %coerce.dive, align 8
1885 void test_vst1q_u64_x3(uint64_t *a, uint64x2x3_t b) {
1886 vst1q_u64_x3(a, b);
1894 // CHECK-A32: store [8 x i64] %b.coerce, ptr %coerce.dive, align 8
1919 void test_vst1q_u64_x4(uint64_t *a, uint64x2x4_t b) {
1920 vst1q_u64_x4(a, b);
1928 // CHECK-A32: store [4 x i64] %b.coerce, ptr %coerce.dive, align 8
1939 void test_vst1q_u8_x2(uint8_t *a, uint8x16x2_t b) {
1940 vst1q_u8_x2(a, b);
1948 // CHECK-A32: store [6 x i64] %b.coerce, ptr %coerce.dive, align 8
1962 void test_vst1q_u8_x3(uint8_t *a, uint8x16x3_t b) {
1963 vst1q_u8_x3(a, b);
1971 // CHECK-A32: store [8 x i64] %b.coerce, ptr %coerce.dive, align 8
1988 void test_vst1q_u8_x4(uint8_t *a, uint8x16x4_t b) {
1989 vst1q_u8_x4(a, b);