Lines Matching refs:slice
6 define void @multi_vector_add_single_vg1x2_s(i32 %slice, <vscale x 4 x float> %zn0, <vscale x 4 x f…
15 call void @llvm.aarch64.sme.fmla.single.vg1x2.nxv4f32(i32 %slice,
18 %slice.7 = add i32 %slice, 7
19 call void @llvm.aarch64.sme.fmla.single.vg1x2.nxv4f32(i32 %slice.7,
25 define void @multi_vector_add_single_vg1x2_d(i32 %slice, <vscale x 2 x double> %zn0, <vscale x 2 x …
34 call void @llvm.aarch64.sme.fmla.single.vg1x2.nxv2f64(i32 %slice,
37 %slice.7 = add i32 %slice, 7
38 call void @llvm.aarch64.sme.fmla.single.vg1x2.nxv2f64(i32 %slice.7,
44 define void @multi_vector_add_single_vg1x4_s(i32 %slice, <vscale x 4 x float> %zn0, <vscale x 4 x f…
56 call void @llvm.aarch64.sme.fmla.single.vg1x4.nxv4f32(i32 %slice,
60 %slice.7 = add i32 %slice, 7
61 call void @llvm.aarch64.sme.fmla.single.vg1x4.nxv4f32(i32 %slice.7,
68 define void @multi_vector_add_single_vg1x4_d(i32 %slice, <vscale x 2 x double> %zn0, <vscale x 2 x …
80 call void @llvm.aarch64.sme.fmla.single.vg1x4.nxv2f64(i32 %slice,
84 %slice.7 = add i32 %slice, 7
85 call void @llvm.aarch64.sme.fmla.single.vg1x4.nxv2f64(i32 %slice.7,
94 define void @multi_vector_sub_single_vg1x2_s(i32 %slice, <vscale x 4 x float> %zn0, <vscale x 4 x f…
103 call void @llvm.aarch64.sme.fmls.single.vg1x2.nxv4f32(i32 %slice,
106 %slice.7 = add i32 %slice, 7
107 call void @llvm.aarch64.sme.fmls.single.vg1x2.nxv4f32(i32 %slice.7,
113 define void @multi_vector_sub_single_vg1x2_d(i32 %slice, <vscale x 2 x double> %zn0, <vscale x 2 x …
122 call void @llvm.aarch64.sme.fmls.single.vg1x2.nxv2f64(i32 %slice,
125 %slice.7 = add i32 %slice, 7
126 call void @llvm.aarch64.sme.fmls.single.vg1x2.nxv2f64(i32 %slice.7,
132 define void @multi_vector_sub_single_vg1x4_s(i32 %slice, <vscale x 4 x float> %zn0, <vscale x 4 x f…
144 call void @llvm.aarch64.sme.fmls.single.vg1x4.nxv4f32(i32 %slice,
148 %slice.7 = add i32 %slice, 7
149 call void @llvm.aarch64.sme.fmls.single.vg1x4.nxv4f32(i32 %slice.7,
156 define void @multi_vector_sub_single_vg1x4_d(i32 %slice, <vscale x 2 x double> %zn0, <vscale x 2 x …
168 call void @llvm.aarch64.sme.fmls.single.vg1x4.nxv2f64(i32 %slice,
172 %slice.7 = add i32 %slice, 7
173 call void @llvm.aarch64.sme.fmls.single.vg1x4.nxv2f64(i32 %slice.7,
182 define void @multi_vector_add_vg1x2_s(i32 %slice, <vscale x 4 x float> %zn0, <vscale x 4 x float> %…
194 call void @llvm.aarch64.sme.fmla.vg1x2.nxv4f32(i32 %slice,
197 %slice.7 = add i32 %slice, 7
198 call void @llvm.aarch64.sme.fmla.vg1x2.nxv4f32(i32 %slice.7,
204 define void @multi_vector_add_vg1x2_d(i32 %slice, <vscale x 2 x double> %zn0, <vscale x 2 x double>…
216 call void @llvm.aarch64.sme.fmla.vg1x2.nxv2f64(i32 %slice,
219 %slice.7 = add i32 %slice, 7
220 call void @llvm.aarch64.sme.fmla.vg1x2.nxv2f64(i32 %slice.7,
227 define void @multi_vector_add_vg1x2_s_regclass(i32 %slice, <vscale x 4 x float> %zn0, <vscale x 4 x…
238 call void @llvm.aarch64.sme.fmla.vg1x2.nxv4f32(i32 %slice,
244 define void @multi_vector_add_vg1x4_s(i32 %slice, <vscale x 4 x float> %zn0, <vscale x 4 x float> %…
260 call void @llvm.aarch64.sme.fmla.vg1x4.nxv4f32(i32 %slice,
263 %slice.7 = add i32 %slice, 7
264 call void @llvm.aarch64.sme.fmla.vg1x4.nxv4f32(i32 %slice.7,
270 define void @multi_vector_add_vg1x4_d(i32 %slice, <vscale x 2 x double> %zn0, <vscale x 2 x double>…
286 call void @llvm.aarch64.sme.fmla.vg1x4.nxv2f64(i32 %slice,
289 %slice.7 = add i32 %slice, 7
290 call void @llvm.aarch64.sme.fmla.vg1x4.nxv2f64(i32 %slice.7,
297 define void @multi_vector_add_vg1x4_s_regclass(i32 %slice, <vscale x 4 x float> %zn0, <vscale x 4 x…
312 call void @llvm.aarch64.sme.fmla.vg1x4.nxv4f32(i32 %slice,
320 define void @multi_vector_sub_vg1x2_s(i32 %slice, <vscale x 4 x float> %zn0, <vscale x 4 x float> %…
332 call void @llvm.aarch64.sme.fmls.vg1x2.nxv4f32(i32 %slice,
335 %slice.7 = add i32 %slice, 7
336 call void @llvm.aarch64.sme.fmls.vg1x2.nxv4f32(i32 %slice.7,
342 define void @multi_vector_sub_vg1x2_d(i32 %slice, <vscale x 2 x double> %zn0, <vscale x 2 x double>…
354 call void @llvm.aarch64.sme.fmls.vg1x2.nxv2f64(i32 %slice,
357 %slice.7 = add i32 %slice, 7
358 call void @llvm.aarch64.sme.fmls.vg1x2.nxv2f64(i32 %slice.7,
364 define void @multi_vector_sub_vg1x4_s(i32 %slice, <vscale x 4 x float> %zn0, <vscale x 4 x float> %…
380 call void @llvm.aarch64.sme.fmls.vg1x4.nxv4f32(i32 %slice,
383 %slice.7 = add i32 %slice, 7
384 call void @llvm.aarch64.sme.fmls.vg1x4.nxv4f32(i32 %slice.7,
390 define void @multi_vector_sub_vg1x4_d(i32 %slice, <vscale x 2 x double> %zn0, <vscale x 2 x double>…
406 call void @llvm.aarch64.sme.fmls.vg1x4.nxv2f64(i32 %slice,
409 %slice.7 = add i32 %slice, 7
410 call void @llvm.aarch64.sme.fmls.vg1x4.nxv2f64(i32 %slice.7,
418 define void @multi_vector_add_lane_vg1x2_s(i32 %slice, <vscale x 4 x float> %zn0, <vscale x 4 x flo…
427 call void @llvm.aarch64.sme.fmla.lane.vg1x2.nxv4f32(i32 %slice,
430 %slice.7 = add i32 %slice, 7
431 call void @llvm.aarch64.sme.fmla.lane.vg1x2.nxv4f32(i32 %slice.7,
437 define void @multi_vector_add_lane_vg1x2_d(i32 %slice, <vscale x 2 x double> %zn0, <vscale x 2 x do…
446 call void @llvm.aarch64.sme.fmla.lane.vg1x2.nxv2f64(i32 %slice,
449 %slice.7 = add i32 %slice, 7
450 call void @llvm.aarch64.sme.fmla.lane.vg1x2.nxv2f64(i32 %slice.7,
457 define void @multi_vector_add_lane_vg1x2_s_regclass(i32 %slice, <vscale x 4 x float> %zn0, <vscale …
465 call void @llvm.aarch64.sme.fmla.lane.vg1x2.nxv4f32(i32 %slice,
471 define void @multi_vector_add_lane_vg1x4_s(i32 %slice, <vscale x 4 x float> %zn0, <vscale x 4 x flo…
483 call void @llvm.aarch64.sme.fmla.lane.vg1x4.nxv4f32(i32 %slice,
487 %slice.7 = add i32 %slice, 7
488 call void @llvm.aarch64.sme.fmla.lane.vg1x4.nxv4f32(i32 %slice.7,
495 define void @multi_vector_add_lane_vg1x4_d(i32 %slice, <vscale x 2 x double> %zn0, <vscale x 2 x do…
507 call void @llvm.aarch64.sme.fmla.lane.vg1x4.nxv2f64(i32 %slice,
511 %slice.7 = add i32 %slice, 7
512 call void @llvm.aarch64.sme.fmla.lane.vg1x4.nxv2f64(i32 %slice.7,
520 define void @multi_vector_add_lane_vg1x4_s_regclass(i32 %slice, <vscale x 4 x float> %zn0, <vscale …
531 call void @llvm.aarch64.sme.fmla.lane.vg1x4.nxv4f32(i32 %slice,
540 define void @multi_vector_sub_lane_vg1x2_s(i32 %slice, <vscale x 4 x float> %zn0, <vscale x 4 x flo…
549 call void @llvm.aarch64.sme.fmls.lane.vg1x2.nxv4f32(i32 %slice,
552 %slice.7 = add i32 %slice, 7
553 call void @llvm.aarch64.sme.fmls.lane.vg1x2.nxv4f32(i32 %slice.7,
559 define void @multi_vector_sub_lane_vg1x2_d(i32 %slice, <vscale x 2 x double> %zn0, <vscale x 2 x do…
568 call void @llvm.aarch64.sme.fmls.lane.vg1x2.nxv2f64(i32 %slice,
571 %slice.7 = add i32 %slice, 7
572 call void @llvm.aarch64.sme.fmls.lane.vg1x2.nxv2f64(i32 %slice.7,
578 define void @multi_vector_sub_lane_vg1x4_s(i32 %slice, <vscale x 4 x float> %zn0, <vscale x 4 x flo…
590 call void @llvm.aarch64.sme.fmls.lane.vg1x4.nxv4f32(i32 %slice,
594 %slice.7 = add i32 %slice, 7
595 call void @llvm.aarch64.sme.fmls.lane.vg1x4.nxv4f32(i32 %slice.7,
602 define void @multi_vector_sub_lane_vg1x4_d(i32 %slice, <vscale x 2 x double> %zn0, <vscale x 2 x do…
614 call void @llvm.aarch64.sme.fmls.lane.vg1x4.nxv2f64(i32 %slice,
618 %slice.7 = add i32 %slice, 7
619 call void @llvm.aarch64.sme.fmls.lane.vg1x4.nxv2f64(i32 %slice.7,