Lines Matching full:base
11 // CHECK-NEXT: [[TMP1:%.*]] = load <8 x half>, ptr [[BASE:%.*]], align 2
14 float16x8_t test_vld1q_f16(const float16_t *base) in test_vld1q_f16() argument
17 return vld1q(base); in test_vld1q_f16()
19 return vld1q_f16(base); in test_vld1q_f16()
25 // CHECK-NEXT: [[TMP1:%.*]] = load <4 x float>, ptr [[BASE:%.*]], align 4
28 float32x4_t test_vld1q_f32(const float32_t *base) in test_vld1q_f32() argument
31 return vld1q(base); in test_vld1q_f32()
33 return vld1q_f32(base); in test_vld1q_f32()
39 // CHECK-NEXT: [[TMP1:%.*]] = load <16 x i8>, ptr [[BASE:%.*]], align 1
42 int8x16_t test_vld1q_s8(const int8_t *base) in test_vld1q_s8() argument
45 return vld1q(base); in test_vld1q_s8()
47 return vld1q_s8(base); in test_vld1q_s8()
53 // CHECK-NEXT: [[TMP1:%.*]] = load <8 x i16>, ptr [[BASE:%.*]], align 2
56 int16x8_t test_vld1q_s16(const int16_t *base) in test_vld1q_s16() argument
59 return vld1q(base); in test_vld1q_s16()
61 return vld1q_s16(base); in test_vld1q_s16()
67 // CHECK-NEXT: [[TMP1:%.*]] = load <4 x i32>, ptr [[BASE:%.*]], align 4
70 int32x4_t test_vld1q_s32(const int32_t *base) in test_vld1q_s32() argument
73 return vld1q(base); in test_vld1q_s32()
75 return vld1q_s32(base); in test_vld1q_s32()
81 // CHECK-NEXT: [[TMP1:%.*]] = load <16 x i8>, ptr [[BASE:%.*]], align 1
84 uint8x16_t test_vld1q_u8(const uint8_t *base) in test_vld1q_u8() argument
87 return vld1q(base); in test_vld1q_u8()
89 return vld1q_u8(base); in test_vld1q_u8()
95 // CHECK-NEXT: [[TMP1:%.*]] = load <8 x i16>, ptr [[BASE:%.*]], align 2
98 uint16x8_t test_vld1q_u16(const uint16_t *base) in test_vld1q_u16() argument
101 return vld1q(base); in test_vld1q_u16()
103 return vld1q_u16(base); in test_vld1q_u16()
109 // CHECK-NEXT: [[TMP1:%.*]] = load <4 x i32>, ptr [[BASE:%.*]], align 4
112 uint32x4_t test_vld1q_u32(const uint32_t *base) in test_vld1q_u32() argument
115 return vld1q(base); in test_vld1q_u32()
117 return vld1q_u32(base); in test_vld1q_u32()
125 // CHECK-NEXT: [[TMP3:%.*]] = call <8 x half> @llvm.masked.load.v8f16.p0(ptr [[BASE:%.*]], i32 2…
128 float16x8_t test_vld1q_z_f16(const float16_t *base, mve_pred16_t p) in test_vld1q_z_f16() argument
131 return vld1q_z(base, p); in test_vld1q_z_f16()
133 return vld1q_z_f16(base, p); in test_vld1q_z_f16()
141 // CHECK-NEXT: [[TMP3:%.*]] = call <4 x float> @llvm.masked.load.v4f32.p0(ptr [[BASE:%.*]], i32 …
144 float32x4_t test_vld1q_z_f32(const float32_t *base, mve_pred16_t p) in test_vld1q_z_f32() argument
147 return vld1q_z(base, p); in test_vld1q_z_f32()
149 return vld1q_z_f32(base, p); in test_vld1q_z_f32()
157 // CHECK-NEXT: [[TMP3:%.*]] = call <16 x i8> @llvm.masked.load.v16i8.p0(ptr [[BASE:%.*]], i32 1,…
160 int8x16_t test_vld1q_z_s8(const int8_t *base, mve_pred16_t p) in test_vld1q_z_s8() argument
163 return vld1q_z(base, p); in test_vld1q_z_s8()
165 return vld1q_z_s8(base, p); in test_vld1q_z_s8()
173 // CHECK-NEXT: [[TMP3:%.*]] = call <8 x i16> @llvm.masked.load.v8i16.p0(ptr [[BASE:%.*]], i32 2,…
176 int16x8_t test_vld1q_z_s16(const int16_t *base, mve_pred16_t p) in test_vld1q_z_s16() argument
179 return vld1q_z(base, p); in test_vld1q_z_s16()
181 return vld1q_z_s16(base, p); in test_vld1q_z_s16()
189 // CHECK-NEXT: [[TMP3:%.*]] = call <4 x i32> @llvm.masked.load.v4i32.p0(ptr [[BASE:%.*]], i32 4,…
192 int32x4_t test_vld1q_z_s32(const int32_t *base, mve_pred16_t p) in test_vld1q_z_s32() argument
195 return vld1q_z(base, p); in test_vld1q_z_s32()
197 return vld1q_z_s32(base, p); in test_vld1q_z_s32()
205 // CHECK-NEXT: [[TMP3:%.*]] = call <16 x i8> @llvm.masked.load.v16i8.p0(ptr [[BASE:%.*]], i32 1,…
208 uint8x16_t test_vld1q_z_u8(const uint8_t *base, mve_pred16_t p) in test_vld1q_z_u8() argument
211 return vld1q_z(base, p); in test_vld1q_z_u8()
213 return vld1q_z_u8(base, p); in test_vld1q_z_u8()
221 // CHECK-NEXT: [[TMP3:%.*]] = call <8 x i16> @llvm.masked.load.v8i16.p0(ptr [[BASE:%.*]], i32 2,…
224 uint16x8_t test_vld1q_z_u16(const uint16_t *base, mve_pred16_t p) in test_vld1q_z_u16() argument
227 return vld1q_z(base, p); in test_vld1q_z_u16()
229 return vld1q_z_u16(base, p); in test_vld1q_z_u16()
237 // CHECK-NEXT: [[TMP3:%.*]] = call <4 x i32> @llvm.masked.load.v4i32.p0(ptr [[BASE:%.*]], i32 4,…
240 uint32x4_t test_vld1q_z_u32(const uint32_t *base, mve_pred16_t p) in test_vld1q_z_u32() argument
243 return vld1q_z(base, p); in test_vld1q_z_u32()
245 return vld1q_z_u32(base, p); in test_vld1q_z_u32()
251 // CHECK-NEXT: [[TMP1:%.*]] = load <16 x i8>, ptr [[BASE:%.*]], align 1
254 int8x16_t test_vldrbq_s8(const int8_t *base) in test_vldrbq_s8() argument
256 return vldrbq_s8(base); in test_vldrbq_s8()
261 // CHECK-NEXT: [[TMP1:%.*]] = load <8 x i8>, ptr [[BASE:%.*]], align 1
265 int16x8_t test_vldrbq_s16(const int8_t *base) in test_vldrbq_s16() argument
267 return vldrbq_s16(base); in test_vldrbq_s16()
272 // CHECK-NEXT: [[TMP1:%.*]] = load <4 x i8>, ptr [[BASE:%.*]], align 1
276 int32x4_t test_vldrbq_s32(const int8_t *base) in test_vldrbq_s32() argument
278 return vldrbq_s32(base); in test_vldrbq_s32()
283 // CHECK-NEXT: [[TMP1:%.*]] = load <16 x i8>, ptr [[BASE:%.*]], align 1
286 uint8x16_t test_vldrbq_u8(const uint8_t *base) in test_vldrbq_u8() argument
288 return vldrbq_u8(base); in test_vldrbq_u8()
293 // CHECK-NEXT: [[TMP1:%.*]] = load <8 x i8>, ptr [[BASE:%.*]], align 1
297 uint16x8_t test_vldrbq_u16(const uint8_t *base) in test_vldrbq_u16() argument
299 return vldrbq_u16(base); in test_vldrbq_u16()
304 // CHECK-NEXT: [[TMP1:%.*]] = load <4 x i8>, ptr [[BASE:%.*]], align 1
308 uint32x4_t test_vldrbq_u32(const uint8_t *base) in test_vldrbq_u32() argument
310 return vldrbq_u32(base); in test_vldrbq_u32()
317 // CHECK-NEXT: [[TMP3:%.*]] = call <16 x i8> @llvm.masked.load.v16i8.p0(ptr [[BASE:%.*]], i32 1,…
320 int8x16_t test_vldrbq_z_s8(const int8_t *base, mve_pred16_t p) in test_vldrbq_z_s8() argument
322 return vldrbq_z_s8(base, p); in test_vldrbq_z_s8()
329 // CHECK-NEXT: [[TMP3:%.*]] = call <8 x i8> @llvm.masked.load.v8i8.p0(ptr [[BASE:%.*]], i32 1, <…
333 int16x8_t test_vldrbq_z_s16(const int8_t *base, mve_pred16_t p) in test_vldrbq_z_s16() argument
335 return vldrbq_z_s16(base, p); in test_vldrbq_z_s16()
342 // CHECK-NEXT: [[TMP3:%.*]] = call <4 x i8> @llvm.masked.load.v4i8.p0(ptr [[BASE:%.*]], i32 1, <…
346 int32x4_t test_vldrbq_z_s32(const int8_t *base, mve_pred16_t p) in test_vldrbq_z_s32() argument
348 return vldrbq_z_s32(base, p); in test_vldrbq_z_s32()
355 // CHECK-NEXT: [[TMP3:%.*]] = call <16 x i8> @llvm.masked.load.v16i8.p0(ptr [[BASE:%.*]], i32 1,…
358 uint8x16_t test_vldrbq_z_u8(const uint8_t *base, mve_pred16_t p) in test_vldrbq_z_u8() argument
360 return vldrbq_z_u8(base, p); in test_vldrbq_z_u8()
367 // CHECK-NEXT: [[TMP3:%.*]] = call <8 x i8> @llvm.masked.load.v8i8.p0(ptr [[BASE:%.*]], i32 1, <…
371 uint16x8_t test_vldrbq_z_u16(const uint8_t *base, mve_pred16_t p) in test_vldrbq_z_u16() argument
373 return vldrbq_z_u16(base, p); in test_vldrbq_z_u16()
380 // CHECK-NEXT: [[TMP3:%.*]] = call <4 x i8> @llvm.masked.load.v4i8.p0(ptr [[BASE:%.*]], i32 1, <…
384 uint32x4_t test_vldrbq_z_u32(const uint8_t *base, mve_pred16_t p) in test_vldrbq_z_u32() argument
386 return vldrbq_z_u32(base, p); in test_vldrbq_z_u32()
391 // CHECK-NEXT: [[TMP1:%.*]] = load <8 x half>, ptr [[BASE:%.*]], align 2
394 float16x8_t test_vldrhq_f16(const float16_t *base) in test_vldrhq_f16() argument
396 return vldrhq_f16(base); in test_vldrhq_f16()
401 // CHECK-NEXT: [[TMP1:%.*]] = load <8 x i16>, ptr [[BASE:%.*]], align 2
404 int16x8_t test_vldrhq_s16(const int16_t *base) in test_vldrhq_s16() argument
406 return vldrhq_s16(base); in test_vldrhq_s16()
411 // CHECK-NEXT: [[TMP1:%.*]] = load <4 x i16>, ptr [[BASE:%.*]], align 2
415 int32x4_t test_vldrhq_s32(const int16_t *base) in test_vldrhq_s32() argument
417 return vldrhq_s32(base); in test_vldrhq_s32()
422 // CHECK-NEXT: [[TMP1:%.*]] = load <8 x i16>, ptr [[BASE:%.*]], align 2
425 uint16x8_t test_vldrhq_u16(const uint16_t *base) in test_vldrhq_u16() argument
427 return vldrhq_u16(base); in test_vldrhq_u16()
432 // CHECK-NEXT: [[TMP1:%.*]] = load <4 x i16>, ptr [[BASE:%.*]], align 2
436 uint32x4_t test_vldrhq_u32(const uint16_t *base) in test_vldrhq_u32() argument
438 return vldrhq_u32(base); in test_vldrhq_u32()
445 // CHECK-NEXT: [[TMP3:%.*]] = call <8 x half> @llvm.masked.load.v8f16.p0(ptr [[BASE:%.*]], i32 2…
448 float16x8_t test_vldrhq_z_f16(const float16_t *base, mve_pred16_t p) in test_vldrhq_z_f16() argument
450 return vldrhq_z_f16(base, p); in test_vldrhq_z_f16()
457 // CHECK-NEXT: [[TMP3:%.*]] = call <8 x i16> @llvm.masked.load.v8i16.p0(ptr [[BASE:%.*]], i32 2,…
460 int16x8_t test_vldrhq_z_s16(const int16_t *base, mve_pred16_t p) in test_vldrhq_z_s16() argument
462 return vldrhq_z_s16(base, p); in test_vldrhq_z_s16()
469 // CHECK-NEXT: [[TMP3:%.*]] = call <4 x i16> @llvm.masked.load.v4i16.p0(ptr [[BASE:%.*]], i32 2,…
473 int32x4_t test_vldrhq_z_s32(const int16_t *base, mve_pred16_t p) in test_vldrhq_z_s32() argument
475 return vldrhq_z_s32(base, p); in test_vldrhq_z_s32()
482 // CHECK-NEXT: [[TMP3:%.*]] = call <8 x i16> @llvm.masked.load.v8i16.p0(ptr [[BASE:%.*]], i32 2,…
485 uint16x8_t test_vldrhq_z_u16(const uint16_t *base, mve_pred16_t p) in test_vldrhq_z_u16() argument
487 return vldrhq_z_u16(base, p); in test_vldrhq_z_u16()
494 // CHECK-NEXT: [[TMP3:%.*]] = call <4 x i16> @llvm.masked.load.v4i16.p0(ptr [[BASE:%.*]], i32 2,…
498 uint32x4_t test_vldrhq_z_u32(const uint16_t *base, mve_pred16_t p) in test_vldrhq_z_u32() argument
500 return vldrhq_z_u32(base, p); in test_vldrhq_z_u32()
505 // CHECK-NEXT: [[TMP1:%.*]] = load <4 x float>, ptr [[BASE:%.*]], align 4
508 float32x4_t test_vldrwq_f32(const float32_t *base) in test_vldrwq_f32() argument
510 return vldrwq_f32(base); in test_vldrwq_f32()
515 // CHECK-NEXT: [[TMP1:%.*]] = load <4 x i32>, ptr [[BASE:%.*]], align 4
518 int32x4_t test_vldrwq_s32(const int32_t *base) in test_vldrwq_s32() argument
520 return vldrwq_s32(base); in test_vldrwq_s32()
525 // CHECK-NEXT: [[TMP1:%.*]] = load <4 x i32>, ptr [[BASE:%.*]], align 4
528 uint32x4_t test_vldrwq_u32(const uint32_t *base) in test_vldrwq_u32() argument
530 return vldrwq_u32(base); in test_vldrwq_u32()
537 // CHECK-NEXT: [[TMP3:%.*]] = call <4 x float> @llvm.masked.load.v4f32.p0(ptr [[BASE:%.*]], i32 …
540 float32x4_t test_vldrwq_z_f32(const float32_t *base, mve_pred16_t p) in test_vldrwq_z_f32() argument
542 return vldrwq_z_f32(base, p); in test_vldrwq_z_f32()
549 // CHECK-NEXT: [[TMP3:%.*]] = call <4 x i32> @llvm.masked.load.v4i32.p0(ptr [[BASE:%.*]], i32 4,…
552 int32x4_t test_vldrwq_z_s32(const int32_t *base, mve_pred16_t p) in test_vldrwq_z_s32() argument
554 return vldrwq_z_s32(base, p); in test_vldrwq_z_s32()
561 // CHECK-NEXT: [[TMP3:%.*]] = call <4 x i32> @llvm.masked.load.v4i32.p0(ptr [[BASE:%.*]], i32 4,…
564 uint32x4_t test_vldrwq_z_u32(const uint32_t *base, mve_pred16_t p) in test_vldrwq_z_u32() argument
566 return vldrwq_z_u32(base, p); in test_vldrwq_z_u32()
571 // CHECK-NEXT: store <8 x half> [[VALUE:%.*]], ptr [[BASE:%.*]], align 2
574 void test_vst1q_f16(float16_t *base, float16x8_t value) in test_vst1q_f16() argument
577 vst1q(base, value); in test_vst1q_f16()
579 vst1q_f16(base, value); in test_vst1q_f16()
585 // CHECK-NEXT: store <4 x float> [[VALUE:%.*]], ptr [[BASE:%.*]], align 4
588 void test_vst1q_f32(float32_t *base, float32x4_t value) in test_vst1q_f32() argument
591 vst1q(base, value); in test_vst1q_f32()
593 vst1q_f32(base, value); in test_vst1q_f32()
599 // CHECK-NEXT: store <16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], align 1
602 void test_vst1q_s8(int8_t *base, int8x16_t value) in test_vst1q_s8() argument
605 vst1q(base, value); in test_vst1q_s8()
607 vst1q_s8(base, value); in test_vst1q_s8()
613 // CHECK-NEXT: store <8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], align 2
616 void test_vst1q_s16(int16_t *base, int16x8_t value) in test_vst1q_s16() argument
619 vst1q(base, value); in test_vst1q_s16()
621 vst1q_s16(base, value); in test_vst1q_s16()
627 // CHECK-NEXT: store <4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], align 4
630 void test_vst1q_s32(int32_t *base, int32x4_t value) in test_vst1q_s32() argument
633 vst1q(base, value); in test_vst1q_s32()
635 vst1q_s32(base, value); in test_vst1q_s32()
641 // CHECK-NEXT: store <16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], align 1
644 void test_vst1q_u8(uint8_t *base, uint8x16_t value) in test_vst1q_u8() argument
647 vst1q(base, value); in test_vst1q_u8()
649 vst1q_u8(base, value); in test_vst1q_u8()
655 // CHECK-NEXT: store <8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], align 2
658 void test_vst1q_u16(uint16_t *base, uint16x8_t value) in test_vst1q_u16() argument
661 vst1q(base, value); in test_vst1q_u16()
663 vst1q_u16(base, value); in test_vst1q_u16()
669 // CHECK-NEXT: store <4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], align 4
672 void test_vst1q_u32(uint32_t *base, uint32x4_t value) in test_vst1q_u32() argument
675 vst1q(base, value); in test_vst1q_u32()
677 vst1q_u32(base, value); in test_vst1q_u32()
685 // CHECK-NEXT: call void @llvm.masked.store.v8f16.p0(<8 x half> [[VALUE:%.*]], ptr [[BASE:%.*]],…
688 void test_vst1q_p_f16(float16_t *base, float16x8_t value, mve_pred16_t p) in test_vst1q_p_f16() argument
691 vst1q_p(base, value, p); in test_vst1q_p_f16()
693 vst1q_p_f16(base, value, p); in test_vst1q_p_f16()
701 // CHECK-NEXT: call void @llvm.masked.store.v4f32.p0(<4 x float> [[VALUE:%.*]], ptr [[BASE:%.*]]…
704 void test_vst1q_p_f32(float32_t *base, float32x4_t value, mve_pred16_t p) in test_vst1q_p_f32() argument
707 vst1q_p(base, value, p); in test_vst1q_p_f32()
709 vst1q_p_f32(base, value, p); in test_vst1q_p_f32()
717 // CHECK-NEXT: call void @llvm.masked.store.v16i8.p0(<16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], …
720 void test_vst1q_p_s8(int8_t *base, int8x16_t value, mve_pred16_t p) in test_vst1q_p_s8() argument
723 vst1q_p(base, value, p); in test_vst1q_p_s8()
725 vst1q_p_s8(base, value, p); in test_vst1q_p_s8()
733 // CHECK-NEXT: call void @llvm.masked.store.v8i16.p0(<8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], …
736 void test_vst1q_p_s16(int16_t *base, int16x8_t value, mve_pred16_t p) in test_vst1q_p_s16() argument
739 vst1q_p(base, value, p); in test_vst1q_p_s16()
741 vst1q_p_s16(base, value, p); in test_vst1q_p_s16()
749 // CHECK-NEXT: call void @llvm.masked.store.v4i32.p0(<4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], …
752 void test_vst1q_p_s32(int32_t *base, int32x4_t value, mve_pred16_t p) in test_vst1q_p_s32() argument
755 vst1q_p(base, value, p); in test_vst1q_p_s32()
757 vst1q_p_s32(base, value, p); in test_vst1q_p_s32()
765 // CHECK-NEXT: call void @llvm.masked.store.v16i8.p0(<16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], …
768 void test_vst1q_p_u8(uint8_t *base, uint8x16_t value, mve_pred16_t p) in test_vst1q_p_u8() argument
771 vst1q_p(base, value, p); in test_vst1q_p_u8()
773 vst1q_p_u8(base, value, p); in test_vst1q_p_u8()
781 // CHECK-NEXT: call void @llvm.masked.store.v8i16.p0(<8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], …
784 void test_vst1q_p_u16(uint16_t *base, uint16x8_t value, mve_pred16_t p) in test_vst1q_p_u16() argument
787 vst1q_p(base, value, p); in test_vst1q_p_u16()
789 vst1q_p_u16(base, value, p); in test_vst1q_p_u16()
797 // CHECK-NEXT: call void @llvm.masked.store.v4i32.p0(<4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], …
800 void test_vst1q_p_u32(uint32_t *base, uint32x4_t value, mve_pred16_t p) in test_vst1q_p_u32() argument
803 vst1q_p(base, value, p); in test_vst1q_p_u32()
805 vst1q_p_u32(base, value, p); in test_vst1q_p_u32()
811 // CHECK-NEXT: store <16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], align 1
814 void test_vstrbq_s8(int8_t *base, int8x16_t value) in test_vstrbq_s8() argument
817 vstrbq(base, value); in test_vstrbq_s8()
819 vstrbq_s8(base, value); in test_vstrbq_s8()
826 // CHECK-NEXT: store <8 x i8> [[TMP0]], ptr [[BASE:%.*]], align 1
829 void test_vstrbq_s16(int8_t *base, int16x8_t value) in test_vstrbq_s16() argument
832 vstrbq(base, value); in test_vstrbq_s16()
834 vstrbq_s16(base, value); in test_vstrbq_s16()
841 // CHECK-NEXT: store <4 x i8> [[TMP0]], ptr [[BASE:%.*]], align 1
844 void test_vstrbq_s32(int8_t *base, int32x4_t value) in test_vstrbq_s32() argument
847 vstrbq(base, value); in test_vstrbq_s32()
849 vstrbq_s32(base, value); in test_vstrbq_s32()
855 // CHECK-NEXT: store <16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], align 1
858 void test_vstrbq_u8(uint8_t *base, uint8x16_t value) in test_vstrbq_u8() argument
861 vstrbq(base, value); in test_vstrbq_u8()
863 vstrbq_u8(base, value); in test_vstrbq_u8()
870 // CHECK-NEXT: store <8 x i8> [[TMP0]], ptr [[BASE:%.*]], align 1
873 void test_vstrbq_u16(uint8_t *base, uint16x8_t value) in test_vstrbq_u16() argument
876 vstrbq(base, value); in test_vstrbq_u16()
878 vstrbq_u16(base, value); in test_vstrbq_u16()
885 // CHECK-NEXT: store <4 x i8> [[TMP0]], ptr [[BASE:%.*]], align 1
888 void test_vstrbq_u32(uint8_t *base, uint32x4_t value) in test_vstrbq_u32() argument
891 vstrbq(base, value); in test_vstrbq_u32()
893 vstrbq_u32(base, value); in test_vstrbq_u32()
901 // CHECK-NEXT: call void @llvm.masked.store.v16i8.p0(<16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], …
904 void test_vstrbq_p_s8(int8_t *base, int8x16_t value, mve_pred16_t p) in test_vstrbq_p_s8() argument
907 vstrbq_p(base, value, p); in test_vstrbq_p_s8()
909 vstrbq_p_s8(base, value, p); in test_vstrbq_p_s8()
918 // CHECK-NEXT: call void @llvm.masked.store.v8i8.p0(<8 x i8> [[TMP0]], ptr [[BASE:%.*]], i32 1, …
921 void test_vstrbq_p_s16(int8_t *base, int16x8_t value, mve_pred16_t p) in test_vstrbq_p_s16() argument
924 vstrbq_p(base, value, p); in test_vstrbq_p_s16()
926 vstrbq_p_s16(base, value, p); in test_vstrbq_p_s16()
935 // CHECK-NEXT: call void @llvm.masked.store.v4i8.p0(<4 x i8> [[TMP0]], ptr [[BASE:%.*]], i32 1, …
938 void test_vstrbq_p_s32(int8_t *base, int32x4_t value, mve_pred16_t p) in test_vstrbq_p_s32() argument
941 vstrbq_p(base, value, p); in test_vstrbq_p_s32()
943 vstrbq_p_s32(base, value, p); in test_vstrbq_p_s32()
951 // CHECK-NEXT: call void @llvm.masked.store.v16i8.p0(<16 x i8> [[VALUE:%.*]], ptr [[BASE:%.*]], …
954 void test_vstrbq_p_u8(uint8_t *base, uint8x16_t value, mve_pred16_t p) in test_vstrbq_p_u8() argument
957 vstrbq_p(base, value, p); in test_vstrbq_p_u8()
959 vstrbq_p_u8(base, value, p); in test_vstrbq_p_u8()
968 // CHECK-NEXT: call void @llvm.masked.store.v8i8.p0(<8 x i8> [[TMP0]], ptr [[BASE:%.*]], i32 1, …
971 void test_vstrbq_p_u16(uint8_t *base, uint16x8_t value, mve_pred16_t p) in test_vstrbq_p_u16() argument
974 vstrbq_p(base, value, p); in test_vstrbq_p_u16()
976 vstrbq_p_u16(base, value, p); in test_vstrbq_p_u16()
985 // CHECK-NEXT: call void @llvm.masked.store.v4i8.p0(<4 x i8> [[TMP0]], ptr [[BASE:%.*]], i32 1, …
988 void test_vstrbq_p_u32(uint8_t *base, uint32x4_t value, mve_pred16_t p) in test_vstrbq_p_u32() argument
991 vstrbq_p(base, value, p); in test_vstrbq_p_u32()
993 vstrbq_p_u32(base, value, p); in test_vstrbq_p_u32()
999 // CHECK-NEXT: store <8 x half> [[VALUE:%.*]], ptr [[BASE:%.*]], align 2
1002 void test_vstrhq_f16(float16_t *base, float16x8_t value) in test_vstrhq_f16() argument
1005 vstrhq(base, value); in test_vstrhq_f16()
1007 vstrhq_f16(base, value); in test_vstrhq_f16()
1013 // CHECK-NEXT: store <8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], align 2
1016 void test_vstrhq_s16(int16_t *base, int16x8_t value) in test_vstrhq_s16() argument
1019 vstrhq(base, value); in test_vstrhq_s16()
1021 vstrhq_s16(base, value); in test_vstrhq_s16()
1028 // CHECK-NEXT: store <4 x i16> [[TMP0]], ptr [[BASE:%.*]], align 2
1031 void test_vstrhq_s32(int16_t *base, int32x4_t value) in test_vstrhq_s32() argument
1034 vstrhq(base, value); in test_vstrhq_s32()
1036 vstrhq_s32(base, value); in test_vstrhq_s32()
1042 // CHECK-NEXT: store <8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], align 2
1045 void test_vstrhq_u16(uint16_t *base, uint16x8_t value) in test_vstrhq_u16() argument
1048 vstrhq(base, value); in test_vstrhq_u16()
1050 vstrhq_u16(base, value); in test_vstrhq_u16()
1057 // CHECK-NEXT: store <4 x i16> [[TMP0]], ptr [[BASE:%.*]], align 2
1060 void test_vstrhq_u32(uint16_t *base, uint32x4_t value) in test_vstrhq_u32() argument
1063 vstrhq(base, value); in test_vstrhq_u32()
1065 vstrhq_u32(base, value); in test_vstrhq_u32()
1073 // CHECK-NEXT: call void @llvm.masked.store.v8f16.p0(<8 x half> [[VALUE:%.*]], ptr [[BASE:%.*]],…
1076 void test_vstrhq_p_f16(float16_t *base, float16x8_t value, mve_pred16_t p) in test_vstrhq_p_f16() argument
1079 vstrhq_p(base, value, p); in test_vstrhq_p_f16()
1081 vstrhq_p_f16(base, value, p); in test_vstrhq_p_f16()
1089 // CHECK-NEXT: call void @llvm.masked.store.v8i16.p0(<8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], …
1092 void test_vstrhq_p_s16(int16_t *base, int16x8_t value, mve_pred16_t p) in test_vstrhq_p_s16() argument
1095 vstrhq_p(base, value, p); in test_vstrhq_p_s16()
1097 vstrhq_p_s16(base, value, p); in test_vstrhq_p_s16()
1106 // CHECK-NEXT: call void @llvm.masked.store.v4i16.p0(<4 x i16> [[TMP0]], ptr [[BASE:%.*]], i32 2…
1109 void test_vstrhq_p_s32(int16_t *base, int32x4_t value, mve_pred16_t p) in test_vstrhq_p_s32() argument
1112 vstrhq_p(base, value, p); in test_vstrhq_p_s32()
1114 vstrhq_p_s32(base, value, p); in test_vstrhq_p_s32()
1122 // CHECK-NEXT: call void @llvm.masked.store.v8i16.p0(<8 x i16> [[VALUE:%.*]], ptr [[BASE:%.*]], …
1125 void test_vstrhq_p_u16(uint16_t *base, uint16x8_t value, mve_pred16_t p) in test_vstrhq_p_u16() argument
1128 vstrhq_p(base, value, p); in test_vstrhq_p_u16()
1130 vstrhq_p_u16(base, value, p); in test_vstrhq_p_u16()
1139 // CHECK-NEXT: call void @llvm.masked.store.v4i16.p0(<4 x i16> [[TMP0]], ptr [[BASE:%.*]], i32 2…
1142 void test_vstrhq_p_u32(uint16_t *base, uint32x4_t value, mve_pred16_t p) in test_vstrhq_p_u32() argument
1145 vstrhq_p(base, value, p); in test_vstrhq_p_u32()
1147 vstrhq_p_u32(base, value, p); in test_vstrhq_p_u32()
1153 // CHECK-NEXT: store <4 x float> [[VALUE:%.*]], ptr [[BASE:%.*]], align 4
1156 void test_vstrwq_f32(float32_t *base, float32x4_t value) in test_vstrwq_f32() argument
1159 vstrwq(base, value); in test_vstrwq_f32()
1161 vstrwq_f32(base, value); in test_vstrwq_f32()
1167 // CHECK-NEXT: store <4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], align 4
1170 void test_vstrwq_s32(int32_t *base, int32x4_t value) in test_vstrwq_s32() argument
1173 vstrwq(base, value); in test_vstrwq_s32()
1175 vstrwq_s32(base, value); in test_vstrwq_s32()
1181 // CHECK-NEXT: store <4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], align 4
1184 void test_vstrwq_u32(uint32_t *base, uint32x4_t value) in test_vstrwq_u32() argument
1187 vstrwq(base, value); in test_vstrwq_u32()
1189 vstrwq_u32(base, value); in test_vstrwq_u32()
1197 // CHECK-NEXT: call void @llvm.masked.store.v4f32.p0(<4 x float> [[VALUE:%.*]], ptr [[BASE:%.*]]…
1200 void test_vstrwq_p_f32(float32_t *base, float32x4_t value, mve_pred16_t p) in test_vstrwq_p_f32() argument
1203 vstrwq_p(base, value, p); in test_vstrwq_p_f32()
1205 vstrwq_p_f32(base, value, p); in test_vstrwq_p_f32()
1213 // CHECK-NEXT: call void @llvm.masked.store.v4i32.p0(<4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], …
1216 void test_vstrwq_p_s32(int32_t *base, int32x4_t value, mve_pred16_t p) in test_vstrwq_p_s32() argument
1219 vstrwq_p(base, value, p); in test_vstrwq_p_s32()
1221 vstrwq_p_s32(base, value, p); in test_vstrwq_p_s32()
1229 // CHECK-NEXT: call void @llvm.masked.store.v4i32.p0(<4 x i32> [[VALUE:%.*]], ptr [[BASE:%.*]], …
1232 void test_vstrwq_p_u32(uint32_t *base, uint32x4_t value, mve_pred16_t p) in test_vstrwq_p_u32() argument
1235 vstrwq_p(base, value, p); in test_vstrwq_p_u32()
1237 vstrwq_p_u32(base, value, p); in test_vstrwq_p_u32()