Lines Matching defs:_1

41     __lsx_vsll_b(__m128i _1, __m128i _2) {  in __lsx_vsll_b()
47 __lsx_vsll_h(__m128i _1, __m128i _2) { in __lsx_vsll_h()
53 __lsx_vsll_w(__m128i _1, __m128i _2) { in __lsx_vsll_w()
59 __lsx_vsll_d(__m128i _1, __m128i _2) { in __lsx_vsll_d()
63 #define __lsx_vslli_b(/*__m128i*/ _1, /*ui3*/ _2) \ argument
66 #define __lsx_vslli_h(/*__m128i*/ _1, /*ui4*/ _2) \ argument
69 #define __lsx_vslli_w(/*__m128i*/ _1, /*ui5*/ _2) \ argument
72 #define __lsx_vslli_d(/*__m128i*/ _1, /*ui6*/ _2) \ argument
77 __lsx_vsra_b(__m128i _1, __m128i _2) { in __lsx_vsra_b()
83 __lsx_vsra_h(__m128i _1, __m128i _2) { in __lsx_vsra_h()
89 __lsx_vsra_w(__m128i _1, __m128i _2) { in __lsx_vsra_w()
95 __lsx_vsra_d(__m128i _1, __m128i _2) { in __lsx_vsra_d()
99 #define __lsx_vsrai_b(/*__m128i*/ _1, /*ui3*/ _2) \ argument
102 #define __lsx_vsrai_h(/*__m128i*/ _1, /*ui4*/ _2) \ argument
105 #define __lsx_vsrai_w(/*__m128i*/ _1, /*ui5*/ _2) \ argument
108 #define __lsx_vsrai_d(/*__m128i*/ _1, /*ui6*/ _2) \ argument
113 __lsx_vsrar_b(__m128i _1, __m128i _2) { in __lsx_vsrar_b()
119 __lsx_vsrar_h(__m128i _1, __m128i _2) { in __lsx_vsrar_h()
125 __lsx_vsrar_w(__m128i _1, __m128i _2) { in __lsx_vsrar_w()
131 __lsx_vsrar_d(__m128i _1, __m128i _2) { in __lsx_vsrar_d()
135 #define __lsx_vsrari_b(/*__m128i*/ _1, /*ui3*/ _2) \ argument
138 #define __lsx_vsrari_h(/*__m128i*/ _1, /*ui4*/ _2) \ argument
141 #define __lsx_vsrari_w(/*__m128i*/ _1, /*ui5*/ _2) \ argument
144 #define __lsx_vsrari_d(/*__m128i*/ _1, /*ui6*/ _2) \ argument
149 __lsx_vsrl_b(__m128i _1, __m128i _2) { in __lsx_vsrl_b()
155 __lsx_vsrl_h(__m128i _1, __m128i _2) { in __lsx_vsrl_h()
161 __lsx_vsrl_w(__m128i _1, __m128i _2) { in __lsx_vsrl_w()
167 __lsx_vsrl_d(__m128i _1, __m128i _2) { in __lsx_vsrl_d()
171 #define __lsx_vsrli_b(/*__m128i*/ _1, /*ui3*/ _2) \ argument
174 #define __lsx_vsrli_h(/*__m128i*/ _1, /*ui4*/ _2) \ argument
177 #define __lsx_vsrli_w(/*__m128i*/ _1, /*ui5*/ _2) \ argument
180 #define __lsx_vsrli_d(/*__m128i*/ _1, /*ui6*/ _2) \ argument
185 __lsx_vsrlr_b(__m128i _1, __m128i _2) { in __lsx_vsrlr_b()
191 __lsx_vsrlr_h(__m128i _1, __m128i _2) { in __lsx_vsrlr_h()
197 __lsx_vsrlr_w(__m128i _1, __m128i _2) { in __lsx_vsrlr_w()
203 __lsx_vsrlr_d(__m128i _1, __m128i _2) { in __lsx_vsrlr_d()
207 #define __lsx_vsrlri_b(/*__m128i*/ _1, /*ui3*/ _2) \ argument
210 #define __lsx_vsrlri_h(/*__m128i*/ _1, /*ui4*/ _2) \ argument
213 #define __lsx_vsrlri_w(/*__m128i*/ _1, /*ui5*/ _2) \ argument
216 #define __lsx_vsrlri_d(/*__m128i*/ _1, /*ui6*/ _2) \ argument
221 __lsx_vbitclr_b(__m128i _1, __m128i _2) { in __lsx_vbitclr_b()
227 __lsx_vbitclr_h(__m128i _1, __m128i _2) { in __lsx_vbitclr_h()
233 __lsx_vbitclr_w(__m128i _1, __m128i _2) { in __lsx_vbitclr_w()
239 __lsx_vbitclr_d(__m128i _1, __m128i _2) { in __lsx_vbitclr_d()
243 #define __lsx_vbitclri_b(/*__m128i*/ _1, /*ui3*/ _2) \ argument
246 #define __lsx_vbitclri_h(/*__m128i*/ _1, /*ui4*/ _2) \ argument
249 #define __lsx_vbitclri_w(/*__m128i*/ _1, /*ui5*/ _2) \ argument
252 #define __lsx_vbitclri_d(/*__m128i*/ _1, /*ui6*/ _2) \ argument
257 __lsx_vbitset_b(__m128i _1, __m128i _2) { in __lsx_vbitset_b()
263 __lsx_vbitset_h(__m128i _1, __m128i _2) { in __lsx_vbitset_h()
269 __lsx_vbitset_w(__m128i _1, __m128i _2) { in __lsx_vbitset_w()
275 __lsx_vbitset_d(__m128i _1, __m128i _2) { in __lsx_vbitset_d()
279 #define __lsx_vbitseti_b(/*__m128i*/ _1, /*ui3*/ _2) \ argument
282 #define __lsx_vbitseti_h(/*__m128i*/ _1, /*ui4*/ _2) \ argument
285 #define __lsx_vbitseti_w(/*__m128i*/ _1, /*ui5*/ _2) \ argument
288 #define __lsx_vbitseti_d(/*__m128i*/ _1, /*ui6*/ _2) \ argument
293 __lsx_vbitrev_b(__m128i _1, __m128i _2) { in __lsx_vbitrev_b()
299 __lsx_vbitrev_h(__m128i _1, __m128i _2) { in __lsx_vbitrev_h()
305 __lsx_vbitrev_w(__m128i _1, __m128i _2) { in __lsx_vbitrev_w()
311 __lsx_vbitrev_d(__m128i _1, __m128i _2) { in __lsx_vbitrev_d()
315 #define __lsx_vbitrevi_b(/*__m128i*/ _1, /*ui3*/ _2) \ argument
318 #define __lsx_vbitrevi_h(/*__m128i*/ _1, /*ui4*/ _2) \ argument
321 #define __lsx_vbitrevi_w(/*__m128i*/ _1, /*ui5*/ _2) \ argument
324 #define __lsx_vbitrevi_d(/*__m128i*/ _1, /*ui6*/ _2) \ argument
329 __lsx_vadd_b(__m128i _1, __m128i _2) { in __lsx_vadd_b()
335 __lsx_vadd_h(__m128i _1, __m128i _2) { in __lsx_vadd_h()
341 __lsx_vadd_w(__m128i _1, __m128i _2) { in __lsx_vadd_w()
347 __lsx_vadd_d(__m128i _1, __m128i _2) { in __lsx_vadd_d()
351 #define __lsx_vaddi_bu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
354 #define __lsx_vaddi_hu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
357 #define __lsx_vaddi_wu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
360 #define __lsx_vaddi_du(/*__m128i*/ _1, /*ui5*/ _2) \ argument
365 __lsx_vsub_b(__m128i _1, __m128i _2) { in __lsx_vsub_b()
371 __lsx_vsub_h(__m128i _1, __m128i _2) { in __lsx_vsub_h()
377 __lsx_vsub_w(__m128i _1, __m128i _2) { in __lsx_vsub_w()
383 __lsx_vsub_d(__m128i _1, __m128i _2) { in __lsx_vsub_d()
387 #define __lsx_vsubi_bu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
390 #define __lsx_vsubi_hu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
393 #define __lsx_vsubi_wu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
396 #define __lsx_vsubi_du(/*__m128i*/ _1, /*ui5*/ _2) \ argument
401 __lsx_vmax_b(__m128i _1, __m128i _2) { in __lsx_vmax_b()
407 __lsx_vmax_h(__m128i _1, __m128i _2) { in __lsx_vmax_h()
413 __lsx_vmax_w(__m128i _1, __m128i _2) { in __lsx_vmax_w()
419 __lsx_vmax_d(__m128i _1, __m128i _2) { in __lsx_vmax_d()
423 #define __lsx_vmaxi_b(/*__m128i*/ _1, /*si5*/ _2) \ argument
426 #define __lsx_vmaxi_h(/*__m128i*/ _1, /*si5*/ _2) \ argument
429 #define __lsx_vmaxi_w(/*__m128i*/ _1, /*si5*/ _2) \ argument
432 #define __lsx_vmaxi_d(/*__m128i*/ _1, /*si5*/ _2) \ argument
437 __lsx_vmax_bu(__m128i _1, __m128i _2) { in __lsx_vmax_bu()
443 __lsx_vmax_hu(__m128i _1, __m128i _2) { in __lsx_vmax_hu()
449 __lsx_vmax_wu(__m128i _1, __m128i _2) { in __lsx_vmax_wu()
455 __lsx_vmax_du(__m128i _1, __m128i _2) { in __lsx_vmax_du()
459 #define __lsx_vmaxi_bu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
462 #define __lsx_vmaxi_hu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
465 #define __lsx_vmaxi_wu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
468 #define __lsx_vmaxi_du(/*__m128i*/ _1, /*ui5*/ _2) \ argument
473 __lsx_vmin_b(__m128i _1, __m128i _2) { in __lsx_vmin_b()
479 __lsx_vmin_h(__m128i _1, __m128i _2) { in __lsx_vmin_h()
485 __lsx_vmin_w(__m128i _1, __m128i _2) { in __lsx_vmin_w()
491 __lsx_vmin_d(__m128i _1, __m128i _2) { in __lsx_vmin_d()
495 #define __lsx_vmini_b(/*__m128i*/ _1, /*si5*/ _2) \ argument
498 #define __lsx_vmini_h(/*__m128i*/ _1, /*si5*/ _2) \ argument
501 #define __lsx_vmini_w(/*__m128i*/ _1, /*si5*/ _2) \ argument
504 #define __lsx_vmini_d(/*__m128i*/ _1, /*si5*/ _2) \ argument
509 __lsx_vmin_bu(__m128i _1, __m128i _2) { in __lsx_vmin_bu()
515 __lsx_vmin_hu(__m128i _1, __m128i _2) { in __lsx_vmin_hu()
521 __lsx_vmin_wu(__m128i _1, __m128i _2) { in __lsx_vmin_wu()
527 __lsx_vmin_du(__m128i _1, __m128i _2) { in __lsx_vmin_du()
531 #define __lsx_vmini_bu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
534 #define __lsx_vmini_hu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
537 #define __lsx_vmini_wu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
540 #define __lsx_vmini_du(/*__m128i*/ _1, /*ui5*/ _2) \ argument
545 __lsx_vseq_b(__m128i _1, __m128i _2) { in __lsx_vseq_b()
551 __lsx_vseq_h(__m128i _1, __m128i _2) { in __lsx_vseq_h()
557 __lsx_vseq_w(__m128i _1, __m128i _2) { in __lsx_vseq_w()
563 __lsx_vseq_d(__m128i _1, __m128i _2) { in __lsx_vseq_d()
567 #define __lsx_vseqi_b(/*__m128i*/ _1, /*si5*/ _2) \ argument
570 #define __lsx_vseqi_h(/*__m128i*/ _1, /*si5*/ _2) \ argument
573 #define __lsx_vseqi_w(/*__m128i*/ _1, /*si5*/ _2) \ argument
576 #define __lsx_vseqi_d(/*__m128i*/ _1, /*si5*/ _2) \ argument
579 #define __lsx_vslti_b(/*__m128i*/ _1, /*si5*/ _2) \ argument
584 __lsx_vslt_b(__m128i _1, __m128i _2) { in __lsx_vslt_b()
590 __lsx_vslt_h(__m128i _1, __m128i _2) { in __lsx_vslt_h()
596 __lsx_vslt_w(__m128i _1, __m128i _2) { in __lsx_vslt_w()
602 __lsx_vslt_d(__m128i _1, __m128i _2) { in __lsx_vslt_d()
606 #define __lsx_vslti_h(/*__m128i*/ _1, /*si5*/ _2) \ argument
609 #define __lsx_vslti_w(/*__m128i*/ _1, /*si5*/ _2) \ argument
612 #define __lsx_vslti_d(/*__m128i*/ _1, /*si5*/ _2) \ argument
617 __lsx_vslt_bu(__m128i _1, __m128i _2) { in __lsx_vslt_bu()
623 __lsx_vslt_hu(__m128i _1, __m128i _2) { in __lsx_vslt_hu()
629 __lsx_vslt_wu(__m128i _1, __m128i _2) { in __lsx_vslt_wu()
635 __lsx_vslt_du(__m128i _1, __m128i _2) { in __lsx_vslt_du()
639 #define __lsx_vslti_bu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
642 #define __lsx_vslti_hu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
645 #define __lsx_vslti_wu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
648 #define __lsx_vslti_du(/*__m128i*/ _1, /*ui5*/ _2) \ argument
653 __lsx_vsle_b(__m128i _1, __m128i _2) { in __lsx_vsle_b()
659 __lsx_vsle_h(__m128i _1, __m128i _2) { in __lsx_vsle_h()
665 __lsx_vsle_w(__m128i _1, __m128i _2) { in __lsx_vsle_w()
671 __lsx_vsle_d(__m128i _1, __m128i _2) { in __lsx_vsle_d()
675 #define __lsx_vslei_b(/*__m128i*/ _1, /*si5*/ _2) \ argument
678 #define __lsx_vslei_h(/*__m128i*/ _1, /*si5*/ _2) \ argument
681 #define __lsx_vslei_w(/*__m128i*/ _1, /*si5*/ _2) \ argument
684 #define __lsx_vslei_d(/*__m128i*/ _1, /*si5*/ _2) \ argument
689 __lsx_vsle_bu(__m128i _1, __m128i _2) { in __lsx_vsle_bu()
695 __lsx_vsle_hu(__m128i _1, __m128i _2) { in __lsx_vsle_hu()
701 __lsx_vsle_wu(__m128i _1, __m128i _2) { in __lsx_vsle_wu()
707 __lsx_vsle_du(__m128i _1, __m128i _2) { in __lsx_vsle_du()
711 #define __lsx_vslei_bu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
714 #define __lsx_vslei_hu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
717 #define __lsx_vslei_wu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
720 #define __lsx_vslei_du(/*__m128i*/ _1, /*ui5*/ _2) \ argument
723 #define __lsx_vsat_b(/*__m128i*/ _1, /*ui3*/ _2) \ argument
726 #define __lsx_vsat_h(/*__m128i*/ _1, /*ui4*/ _2) \ argument
729 #define __lsx_vsat_w(/*__m128i*/ _1, /*ui5*/ _2) \ argument
732 #define __lsx_vsat_d(/*__m128i*/ _1, /*ui6*/ _2) \ argument
735 #define __lsx_vsat_bu(/*__m128i*/ _1, /*ui3*/ _2) \ argument
738 #define __lsx_vsat_hu(/*__m128i*/ _1, /*ui4*/ _2) \ argument
741 #define __lsx_vsat_wu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
744 #define __lsx_vsat_du(/*__m128i*/ _1, /*ui6*/ _2) \ argument
749 __lsx_vadda_b(__m128i _1, __m128i _2) { in __lsx_vadda_b()
755 __lsx_vadda_h(__m128i _1, __m128i _2) { in __lsx_vadda_h()
761 __lsx_vadda_w(__m128i _1, __m128i _2) { in __lsx_vadda_w()
767 __lsx_vadda_d(__m128i _1, __m128i _2) { in __lsx_vadda_d()
773 __lsx_vsadd_b(__m128i _1, __m128i _2) { in __lsx_vsadd_b()
779 __lsx_vsadd_h(__m128i _1, __m128i _2) { in __lsx_vsadd_h()
785 __lsx_vsadd_w(__m128i _1, __m128i _2) { in __lsx_vsadd_w()
791 __lsx_vsadd_d(__m128i _1, __m128i _2) { in __lsx_vsadd_d()
797 __lsx_vsadd_bu(__m128i _1, __m128i _2) { in __lsx_vsadd_bu()
803 __lsx_vsadd_hu(__m128i _1, __m128i _2) { in __lsx_vsadd_hu()
809 __lsx_vsadd_wu(__m128i _1, __m128i _2) { in __lsx_vsadd_wu()
815 __lsx_vsadd_du(__m128i _1, __m128i _2) { in __lsx_vsadd_du()
821 __lsx_vavg_b(__m128i _1, __m128i _2) { in __lsx_vavg_b()
827 __lsx_vavg_h(__m128i _1, __m128i _2) { in __lsx_vavg_h()
833 __lsx_vavg_w(__m128i _1, __m128i _2) { in __lsx_vavg_w()
839 __lsx_vavg_d(__m128i _1, __m128i _2) { in __lsx_vavg_d()
845 __lsx_vavg_bu(__m128i _1, __m128i _2) { in __lsx_vavg_bu()
851 __lsx_vavg_hu(__m128i _1, __m128i _2) { in __lsx_vavg_hu()
857 __lsx_vavg_wu(__m128i _1, __m128i _2) { in __lsx_vavg_wu()
863 __lsx_vavg_du(__m128i _1, __m128i _2) { in __lsx_vavg_du()
869 __lsx_vavgr_b(__m128i _1, __m128i _2) { in __lsx_vavgr_b()
875 __lsx_vavgr_h(__m128i _1, __m128i _2) { in __lsx_vavgr_h()
881 __lsx_vavgr_w(__m128i _1, __m128i _2) { in __lsx_vavgr_w()
887 __lsx_vavgr_d(__m128i _1, __m128i _2) { in __lsx_vavgr_d()
893 __lsx_vavgr_bu(__m128i _1, __m128i _2) { in __lsx_vavgr_bu()
899 __lsx_vavgr_hu(__m128i _1, __m128i _2) { in __lsx_vavgr_hu()
905 __lsx_vavgr_wu(__m128i _1, __m128i _2) { in __lsx_vavgr_wu()
911 __lsx_vavgr_du(__m128i _1, __m128i _2) { in __lsx_vavgr_du()
917 __lsx_vssub_b(__m128i _1, __m128i _2) { in __lsx_vssub_b()
923 __lsx_vssub_h(__m128i _1, __m128i _2) { in __lsx_vssub_h()
929 __lsx_vssub_w(__m128i _1, __m128i _2) { in __lsx_vssub_w()
935 __lsx_vssub_d(__m128i _1, __m128i _2) { in __lsx_vssub_d()
941 __lsx_vssub_bu(__m128i _1, __m128i _2) { in __lsx_vssub_bu()
947 __lsx_vssub_hu(__m128i _1, __m128i _2) { in __lsx_vssub_hu()
953 __lsx_vssub_wu(__m128i _1, __m128i _2) { in __lsx_vssub_wu()
959 __lsx_vssub_du(__m128i _1, __m128i _2) { in __lsx_vssub_du()
965 __lsx_vabsd_b(__m128i _1, __m128i _2) { in __lsx_vabsd_b()
971 __lsx_vabsd_h(__m128i _1, __m128i _2) { in __lsx_vabsd_h()
977 __lsx_vabsd_w(__m128i _1, __m128i _2) { in __lsx_vabsd_w()
983 __lsx_vabsd_d(__m128i _1, __m128i _2) { in __lsx_vabsd_d()
989 __lsx_vabsd_bu(__m128i _1, __m128i _2) { in __lsx_vabsd_bu()
995 __lsx_vabsd_hu(__m128i _1, __m128i _2) { in __lsx_vabsd_hu()
1001 __lsx_vabsd_wu(__m128i _1, __m128i _2) { in __lsx_vabsd_wu()
1007 __lsx_vabsd_du(__m128i _1, __m128i _2) { in __lsx_vabsd_du()
1013 __lsx_vmul_b(__m128i _1, __m128i _2) { in __lsx_vmul_b()
1019 __lsx_vmul_h(__m128i _1, __m128i _2) { in __lsx_vmul_h()
1025 __lsx_vmul_w(__m128i _1, __m128i _2) { in __lsx_vmul_w()
1031 __lsx_vmul_d(__m128i _1, __m128i _2) { in __lsx_vmul_d()
1037 __lsx_vmadd_b(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmadd_b()
1043 __lsx_vmadd_h(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmadd_h()
1049 __lsx_vmadd_w(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmadd_w()
1055 __lsx_vmadd_d(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmadd_d()
1061 __lsx_vmsub_b(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmsub_b()
1067 __lsx_vmsub_h(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmsub_h()
1073 __lsx_vmsub_w(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmsub_w()
1079 __lsx_vmsub_d(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmsub_d()
1085 __lsx_vdiv_b(__m128i _1, __m128i _2) { in __lsx_vdiv_b()
1091 __lsx_vdiv_h(__m128i _1, __m128i _2) { in __lsx_vdiv_h()
1097 __lsx_vdiv_w(__m128i _1, __m128i _2) { in __lsx_vdiv_w()
1103 __lsx_vdiv_d(__m128i _1, __m128i _2) { in __lsx_vdiv_d()
1109 __lsx_vdiv_bu(__m128i _1, __m128i _2) { in __lsx_vdiv_bu()
1115 __lsx_vdiv_hu(__m128i _1, __m128i _2) { in __lsx_vdiv_hu()
1121 __lsx_vdiv_wu(__m128i _1, __m128i _2) { in __lsx_vdiv_wu()
1127 __lsx_vdiv_du(__m128i _1, __m128i _2) { in __lsx_vdiv_du()
1133 __lsx_vhaddw_h_b(__m128i _1, __m128i _2) { in __lsx_vhaddw_h_b()
1139 __lsx_vhaddw_w_h(__m128i _1, __m128i _2) { in __lsx_vhaddw_w_h()
1145 __lsx_vhaddw_d_w(__m128i _1, __m128i _2) { in __lsx_vhaddw_d_w()
1151 __lsx_vhaddw_hu_bu(__m128i _1, __m128i _2) { in __lsx_vhaddw_hu_bu()
1157 __lsx_vhaddw_wu_hu(__m128i _1, __m128i _2) { in __lsx_vhaddw_wu_hu()
1163 __lsx_vhaddw_du_wu(__m128i _1, __m128i _2) { in __lsx_vhaddw_du_wu()
1169 __lsx_vhsubw_h_b(__m128i _1, __m128i _2) { in __lsx_vhsubw_h_b()
1175 __lsx_vhsubw_w_h(__m128i _1, __m128i _2) { in __lsx_vhsubw_w_h()
1181 __lsx_vhsubw_d_w(__m128i _1, __m128i _2) { in __lsx_vhsubw_d_w()
1187 __lsx_vhsubw_hu_bu(__m128i _1, __m128i _2) { in __lsx_vhsubw_hu_bu()
1193 __lsx_vhsubw_wu_hu(__m128i _1, __m128i _2) { in __lsx_vhsubw_wu_hu()
1199 __lsx_vhsubw_du_wu(__m128i _1, __m128i _2) { in __lsx_vhsubw_du_wu()
1205 __lsx_vmod_b(__m128i _1, __m128i _2) { in __lsx_vmod_b()
1211 __lsx_vmod_h(__m128i _1, __m128i _2) { in __lsx_vmod_h()
1217 __lsx_vmod_w(__m128i _1, __m128i _2) { in __lsx_vmod_w()
1223 __lsx_vmod_d(__m128i _1, __m128i _2) { in __lsx_vmod_d()
1229 __lsx_vmod_bu(__m128i _1, __m128i _2) { in __lsx_vmod_bu()
1235 __lsx_vmod_hu(__m128i _1, __m128i _2) { in __lsx_vmod_hu()
1241 __lsx_vmod_wu(__m128i _1, __m128i _2) { in __lsx_vmod_wu()
1247 __lsx_vmod_du(__m128i _1, __m128i _2) { in __lsx_vmod_du()
1253 __lsx_vreplve_b(__m128i _1, int _2) { in __lsx_vreplve_b()
1259 __lsx_vreplve_h(__m128i _1, int _2) { in __lsx_vreplve_h()
1265 __lsx_vreplve_w(__m128i _1, int _2) { in __lsx_vreplve_w()
1271 __lsx_vreplve_d(__m128i _1, int _2) { in __lsx_vreplve_d()
1275 #define __lsx_vreplvei_b(/*__m128i*/ _1, /*ui4*/ _2) \ argument
1278 #define __lsx_vreplvei_h(/*__m128i*/ _1, /*ui3*/ _2) \ argument
1281 #define __lsx_vreplvei_w(/*__m128i*/ _1, /*ui2*/ _2) \ argument
1284 #define __lsx_vreplvei_d(/*__m128i*/ _1, /*ui1*/ _2) \ argument
1289 __lsx_vpickev_b(__m128i _1, __m128i _2) { in __lsx_vpickev_b()
1295 __lsx_vpickev_h(__m128i _1, __m128i _2) { in __lsx_vpickev_h()
1301 __lsx_vpickev_w(__m128i _1, __m128i _2) { in __lsx_vpickev_w()
1307 __lsx_vpickev_d(__m128i _1, __m128i _2) { in __lsx_vpickev_d()
1313 __lsx_vpickod_b(__m128i _1, __m128i _2) { in __lsx_vpickod_b()
1319 __lsx_vpickod_h(__m128i _1, __m128i _2) { in __lsx_vpickod_h()
1325 __lsx_vpickod_w(__m128i _1, __m128i _2) { in __lsx_vpickod_w()
1331 __lsx_vpickod_d(__m128i _1, __m128i _2) { in __lsx_vpickod_d()
1337 __lsx_vilvh_b(__m128i _1, __m128i _2) { in __lsx_vilvh_b()
1343 __lsx_vilvh_h(__m128i _1, __m128i _2) { in __lsx_vilvh_h()
1349 __lsx_vilvh_w(__m128i _1, __m128i _2) { in __lsx_vilvh_w()
1355 __lsx_vilvh_d(__m128i _1, __m128i _2) { in __lsx_vilvh_d()
1361 __lsx_vilvl_b(__m128i _1, __m128i _2) { in __lsx_vilvl_b()
1367 __lsx_vilvl_h(__m128i _1, __m128i _2) { in __lsx_vilvl_h()
1373 __lsx_vilvl_w(__m128i _1, __m128i _2) { in __lsx_vilvl_w()
1379 __lsx_vilvl_d(__m128i _1, __m128i _2) { in __lsx_vilvl_d()
1385 __lsx_vpackev_b(__m128i _1, __m128i _2) { in __lsx_vpackev_b()
1391 __lsx_vpackev_h(__m128i _1, __m128i _2) { in __lsx_vpackev_h()
1397 __lsx_vpackev_w(__m128i _1, __m128i _2) { in __lsx_vpackev_w()
1403 __lsx_vpackev_d(__m128i _1, __m128i _2) { in __lsx_vpackev_d()
1409 __lsx_vpackod_b(__m128i _1, __m128i _2) { in __lsx_vpackod_b()
1415 __lsx_vpackod_h(__m128i _1, __m128i _2) { in __lsx_vpackod_h()
1421 __lsx_vpackod_w(__m128i _1, __m128i _2) { in __lsx_vpackod_w()
1427 __lsx_vpackod_d(__m128i _1, __m128i _2) { in __lsx_vpackod_d()
1433 __lsx_vshuf_h(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vshuf_h()
1439 __lsx_vshuf_w(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vshuf_w()
1445 __lsx_vshuf_d(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vshuf_d()
1451 __lsx_vand_v(__m128i _1, __m128i _2) { in __lsx_vand_v()
1455 #define __lsx_vandi_b(/*__m128i*/ _1, /*ui8*/ _2) \ argument
1460 __lsx_vor_v(__m128i _1, __m128i _2) { in __lsx_vor_v()
1464 #define __lsx_vori_b(/*__m128i*/ _1, /*ui8*/ _2) \ argument
1469 __lsx_vnor_v(__m128i _1, __m128i _2) { in __lsx_vnor_v()
1473 #define __lsx_vnori_b(/*__m128i*/ _1, /*ui8*/ _2) \ argument
1478 __lsx_vxor_v(__m128i _1, __m128i _2) { in __lsx_vxor_v()
1482 #define __lsx_vxori_b(/*__m128i*/ _1, /*ui8*/ _2) \ argument
1487 __lsx_vbitsel_v(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vbitsel_v()
1491 #define __lsx_vbitseli_b(/*__m128i*/ _1, /*__m128i*/ _2, /*ui8*/ _3) \ argument
1494 #define __lsx_vshuf4i_b(/*__m128i*/ _1, /*ui8*/ _2) \ argument
1497 #define __lsx_vshuf4i_h(/*__m128i*/ _1, /*ui8*/ _2) \ argument
1500 #define __lsx_vshuf4i_w(/*__m128i*/ _1, /*ui8*/ _2) \ argument
1505 __lsx_vreplgr2vr_b(int _1) { in __lsx_vreplgr2vr_b()
1511 __lsx_vreplgr2vr_h(int _1) { in __lsx_vreplgr2vr_h()
1517 __lsx_vreplgr2vr_w(int _1) { in __lsx_vreplgr2vr_w()
1523 __lsx_vreplgr2vr_d(long int _1) { in __lsx_vreplgr2vr_d()
1529 __lsx_vpcnt_b(__m128i _1) { in __lsx_vpcnt_b()
1535 __lsx_vpcnt_h(__m128i _1) { in __lsx_vpcnt_h()
1541 __lsx_vpcnt_w(__m128i _1) { in __lsx_vpcnt_w()
1547 __lsx_vpcnt_d(__m128i _1) { in __lsx_vpcnt_d()
1553 __lsx_vclo_b(__m128i _1) { in __lsx_vclo_b()
1559 __lsx_vclo_h(__m128i _1) { in __lsx_vclo_h()
1565 __lsx_vclo_w(__m128i _1) { in __lsx_vclo_w()
1571 __lsx_vclo_d(__m128i _1) { in __lsx_vclo_d()
1577 __lsx_vclz_b(__m128i _1) { in __lsx_vclz_b()
1583 __lsx_vclz_h(__m128i _1) { in __lsx_vclz_h()
1589 __lsx_vclz_w(__m128i _1) { in __lsx_vclz_w()
1595 __lsx_vclz_d(__m128i _1) { in __lsx_vclz_d()
1599 #define __lsx_vpickve2gr_b(/*__m128i*/ _1, /*ui4*/ _2) \ argument
1602 #define __lsx_vpickve2gr_h(/*__m128i*/ _1, /*ui3*/ _2) \ argument
1605 #define __lsx_vpickve2gr_w(/*__m128i*/ _1, /*ui2*/ _2) \ argument
1608 #define __lsx_vpickve2gr_d(/*__m128i*/ _1, /*ui1*/ _2) \ argument
1611 #define __lsx_vpickve2gr_bu(/*__m128i*/ _1, /*ui4*/ _2) \ argument
1614 #define __lsx_vpickve2gr_hu(/*__m128i*/ _1, /*ui3*/ _2) \ argument
1617 #define __lsx_vpickve2gr_wu(/*__m128i*/ _1, /*ui2*/ _2) \ argument
1620 #define __lsx_vpickve2gr_du(/*__m128i*/ _1, /*ui1*/ _2) \ argument
1623 #define __lsx_vinsgr2vr_b(/*__m128i*/ _1, /*int*/ _2, /*ui4*/ _3) \ argument
1626 #define __lsx_vinsgr2vr_h(/*__m128i*/ _1, /*int*/ _2, /*ui3*/ _3) \ argument
1629 #define __lsx_vinsgr2vr_w(/*__m128i*/ _1, /*int*/ _2, /*ui2*/ _3) \ argument
1632 #define __lsx_vinsgr2vr_d(/*__m128i*/ _1, /*long int*/ _2, /*ui1*/ _3) \ argument
1637 __lsx_vfadd_s(__m128 _1, __m128 _2) { in __lsx_vfadd_s()
1643 __lsx_vfadd_d(__m128d _1, __m128d _2) { in __lsx_vfadd_d()
1649 __lsx_vfsub_s(__m128 _1, __m128 _2) { in __lsx_vfsub_s()
1655 __lsx_vfsub_d(__m128d _1, __m128d _2) { in __lsx_vfsub_d()
1661 __lsx_vfmul_s(__m128 _1, __m128 _2) { in __lsx_vfmul_s()
1667 __lsx_vfmul_d(__m128d _1, __m128d _2) { in __lsx_vfmul_d()
1673 __lsx_vfdiv_s(__m128 _1, __m128 _2) { in __lsx_vfdiv_s()
1679 __lsx_vfdiv_d(__m128d _1, __m128d _2) { in __lsx_vfdiv_d()
1685 __lsx_vfcvt_h_s(__m128 _1, __m128 _2) { in __lsx_vfcvt_h_s()
1691 __lsx_vfcvt_s_d(__m128d _1, __m128d _2) { in __lsx_vfcvt_s_d()
1697 __lsx_vfmin_s(__m128 _1, __m128 _2) { in __lsx_vfmin_s()
1703 __lsx_vfmin_d(__m128d _1, __m128d _2) { in __lsx_vfmin_d()
1709 __lsx_vfmina_s(__m128 _1, __m128 _2) { in __lsx_vfmina_s()
1715 __lsx_vfmina_d(__m128d _1, __m128d _2) { in __lsx_vfmina_d()
1721 __lsx_vfmax_s(__m128 _1, __m128 _2) { in __lsx_vfmax_s()
1727 __lsx_vfmax_d(__m128d _1, __m128d _2) { in __lsx_vfmax_d()
1733 __lsx_vfmaxa_s(__m128 _1, __m128 _2) { in __lsx_vfmaxa_s()
1739 __lsx_vfmaxa_d(__m128d _1, __m128d _2) { in __lsx_vfmaxa_d()
1745 __lsx_vfclass_s(__m128 _1) { in __lsx_vfclass_s()
1751 __lsx_vfclass_d(__m128d _1) { in __lsx_vfclass_d()
1757 __lsx_vfsqrt_s(__m128 _1) { in __lsx_vfsqrt_s()
1763 __lsx_vfsqrt_d(__m128d _1) { in __lsx_vfsqrt_d()
1769 __lsx_vfrecip_s(__m128 _1) { in __lsx_vfrecip_s()
1775 __lsx_vfrecip_d(__m128d _1) { in __lsx_vfrecip_d()
1781 __lsx_vfrecipe_s(__m128 _1) { in __lsx_vfrecipe_s()
1787 __lsx_vfrecipe_d(__m128d _1) { in __lsx_vfrecipe_d()
1793 __lsx_vfrint_s(__m128 _1) { in __lsx_vfrint_s()
1799 __lsx_vfrint_d(__m128d _1) { in __lsx_vfrint_d()
1805 __lsx_vfrsqrt_s(__m128 _1) { in __lsx_vfrsqrt_s()
1811 __lsx_vfrsqrt_d(__m128d _1) { in __lsx_vfrsqrt_d()
1817 __lsx_vfrsqrte_s(__m128 _1) { in __lsx_vfrsqrte_s()
1823 __lsx_vfrsqrte_d(__m128d _1) { in __lsx_vfrsqrte_d()
1829 __lsx_vflogb_s(__m128 _1) { in __lsx_vflogb_s()
1835 __lsx_vflogb_d(__m128d _1) { in __lsx_vflogb_d()
1841 __lsx_vfcvth_s_h(__m128i _1) { in __lsx_vfcvth_s_h()
1847 __lsx_vfcvth_d_s(__m128 _1) { in __lsx_vfcvth_d_s()
1853 __lsx_vfcvtl_s_h(__m128i _1) { in __lsx_vfcvtl_s_h()
1859 __lsx_vfcvtl_d_s(__m128 _1) { in __lsx_vfcvtl_d_s()
1865 __lsx_vftint_w_s(__m128 _1) { in __lsx_vftint_w_s()
1871 __lsx_vftint_l_d(__m128d _1) { in __lsx_vftint_l_d()
1877 __lsx_vftint_wu_s(__m128 _1) { in __lsx_vftint_wu_s()
1883 __lsx_vftint_lu_d(__m128d _1) { in __lsx_vftint_lu_d()
1889 __lsx_vftintrz_w_s(__m128 _1) { in __lsx_vftintrz_w_s()
1895 __lsx_vftintrz_l_d(__m128d _1) { in __lsx_vftintrz_l_d()
1901 __lsx_vftintrz_wu_s(__m128 _1) { in __lsx_vftintrz_wu_s()
1907 __lsx_vftintrz_lu_d(__m128d _1) { in __lsx_vftintrz_lu_d()
1913 __lsx_vffint_s_w(__m128i _1) { in __lsx_vffint_s_w()
1919 __lsx_vffint_d_l(__m128i _1) { in __lsx_vffint_d_l()
1925 __lsx_vffint_s_wu(__m128i _1) { in __lsx_vffint_s_wu()
1931 __lsx_vffint_d_lu(__m128i _1) { in __lsx_vffint_d_lu()
1937 __lsx_vandn_v(__m128i _1, __m128i _2) { in __lsx_vandn_v()
1943 __lsx_vneg_b(__m128i _1) { in __lsx_vneg_b()
1949 __lsx_vneg_h(__m128i _1) { in __lsx_vneg_h()
1955 __lsx_vneg_w(__m128i _1) { in __lsx_vneg_w()
1961 __lsx_vneg_d(__m128i _1) { in __lsx_vneg_d()
1967 __lsx_vmuh_b(__m128i _1, __m128i _2) { in __lsx_vmuh_b()
1973 __lsx_vmuh_h(__m128i _1, __m128i _2) { in __lsx_vmuh_h()
1979 __lsx_vmuh_w(__m128i _1, __m128i _2) { in __lsx_vmuh_w()
1985 __lsx_vmuh_d(__m128i _1, __m128i _2) { in __lsx_vmuh_d()
1991 __lsx_vmuh_bu(__m128i _1, __m128i _2) { in __lsx_vmuh_bu()
1997 __lsx_vmuh_hu(__m128i _1, __m128i _2) { in __lsx_vmuh_hu()
2003 __lsx_vmuh_wu(__m128i _1, __m128i _2) { in __lsx_vmuh_wu()
2009 __lsx_vmuh_du(__m128i _1, __m128i _2) { in __lsx_vmuh_du()
2013 #define __lsx_vsllwil_h_b(/*__m128i*/ _1, /*ui3*/ _2) \ argument
2016 #define __lsx_vsllwil_w_h(/*__m128i*/ _1, /*ui4*/ _2) \ argument
2019 #define __lsx_vsllwil_d_w(/*__m128i*/ _1, /*ui5*/ _2) \ argument
2022 #define __lsx_vsllwil_hu_bu(/*__m128i*/ _1, /*ui3*/ _2) \ argument
2025 #define __lsx_vsllwil_wu_hu(/*__m128i*/ _1, /*ui4*/ _2) \ argument
2028 #define __lsx_vsllwil_du_wu(/*__m128i*/ _1, /*ui5*/ _2) \ argument
2033 __lsx_vsran_b_h(__m128i _1, __m128i _2) { in __lsx_vsran_b_h()
2039 __lsx_vsran_h_w(__m128i _1, __m128i _2) { in __lsx_vsran_h_w()
2045 __lsx_vsran_w_d(__m128i _1, __m128i _2) { in __lsx_vsran_w_d()
2051 __lsx_vssran_b_h(__m128i _1, __m128i _2) { in __lsx_vssran_b_h()
2057 __lsx_vssran_h_w(__m128i _1, __m128i _2) { in __lsx_vssran_h_w()
2063 __lsx_vssran_w_d(__m128i _1, __m128i _2) { in __lsx_vssran_w_d()
2069 __lsx_vssran_bu_h(__m128i _1, __m128i _2) { in __lsx_vssran_bu_h()
2075 __lsx_vssran_hu_w(__m128i _1, __m128i _2) { in __lsx_vssran_hu_w()
2081 __lsx_vssran_wu_d(__m128i _1, __m128i _2) { in __lsx_vssran_wu_d()
2087 __lsx_vsrarn_b_h(__m128i _1, __m128i _2) { in __lsx_vsrarn_b_h()
2093 __lsx_vsrarn_h_w(__m128i _1, __m128i _2) { in __lsx_vsrarn_h_w()
2099 __lsx_vsrarn_w_d(__m128i _1, __m128i _2) { in __lsx_vsrarn_w_d()
2105 __lsx_vssrarn_b_h(__m128i _1, __m128i _2) { in __lsx_vssrarn_b_h()
2111 __lsx_vssrarn_h_w(__m128i _1, __m128i _2) { in __lsx_vssrarn_h_w()
2117 __lsx_vssrarn_w_d(__m128i _1, __m128i _2) { in __lsx_vssrarn_w_d()
2123 __lsx_vssrarn_bu_h(__m128i _1, __m128i _2) { in __lsx_vssrarn_bu_h()
2129 __lsx_vssrarn_hu_w(__m128i _1, __m128i _2) { in __lsx_vssrarn_hu_w()
2135 __lsx_vssrarn_wu_d(__m128i _1, __m128i _2) { in __lsx_vssrarn_wu_d()
2141 __lsx_vsrln_b_h(__m128i _1, __m128i _2) { in __lsx_vsrln_b_h()
2147 __lsx_vsrln_h_w(__m128i _1, __m128i _2) { in __lsx_vsrln_h_w()
2153 __lsx_vsrln_w_d(__m128i _1, __m128i _2) { in __lsx_vsrln_w_d()
2159 __lsx_vssrln_bu_h(__m128i _1, __m128i _2) { in __lsx_vssrln_bu_h()
2165 __lsx_vssrln_hu_w(__m128i _1, __m128i _2) { in __lsx_vssrln_hu_w()
2171 __lsx_vssrln_wu_d(__m128i _1, __m128i _2) { in __lsx_vssrln_wu_d()
2177 __lsx_vsrlrn_b_h(__m128i _1, __m128i _2) { in __lsx_vsrlrn_b_h()
2183 __lsx_vsrlrn_h_w(__m128i _1, __m128i _2) { in __lsx_vsrlrn_h_w()
2189 __lsx_vsrlrn_w_d(__m128i _1, __m128i _2) { in __lsx_vsrlrn_w_d()
2195 __lsx_vssrlrn_bu_h(__m128i _1, __m128i _2) { in __lsx_vssrlrn_bu_h()
2201 __lsx_vssrlrn_hu_w(__m128i _1, __m128i _2) { in __lsx_vssrlrn_hu_w()
2207 __lsx_vssrlrn_wu_d(__m128i _1, __m128i _2) { in __lsx_vssrlrn_wu_d()
2211 #define __lsx_vfrstpi_b(/*__m128i*/ _1, /*__m128i*/ _2, /*ui5*/ _3) \ argument
2214 #define __lsx_vfrstpi_h(/*__m128i*/ _1, /*__m128i*/ _2, /*ui5*/ _3) \ argument
2219 __lsx_vfrstp_b(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vfrstp_b()
2225 __lsx_vfrstp_h(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vfrstp_h()
2229 #define __lsx_vshuf4i_d(/*__m128i*/ _1, /*__m128i*/ _2, /*ui8*/ _3) \ argument
2232 #define __lsx_vbsrl_v(/*__m128i*/ _1, /*ui5*/ _2) \ argument
2235 #define __lsx_vbsll_v(/*__m128i*/ _1, /*ui5*/ _2) \ argument
2238 #define __lsx_vextrins_b(/*__m128i*/ _1, /*__m128i*/ _2, /*ui8*/ _3) \ argument
2241 #define __lsx_vextrins_h(/*__m128i*/ _1, /*__m128i*/ _2, /*ui8*/ _3) \ argument
2244 #define __lsx_vextrins_w(/*__m128i*/ _1, /*__m128i*/ _2, /*ui8*/ _3) \ argument
2247 #define __lsx_vextrins_d(/*__m128i*/ _1, /*__m128i*/ _2, /*ui8*/ _3) \ argument
2252 __lsx_vmskltz_b(__m128i _1) { in __lsx_vmskltz_b()
2258 __lsx_vmskltz_h(__m128i _1) { in __lsx_vmskltz_h()
2264 __lsx_vmskltz_w(__m128i _1) { in __lsx_vmskltz_w()
2270 __lsx_vmskltz_d(__m128i _1) { in __lsx_vmskltz_d()
2276 __lsx_vsigncov_b(__m128i _1, __m128i _2) { in __lsx_vsigncov_b()
2282 __lsx_vsigncov_h(__m128i _1, __m128i _2) { in __lsx_vsigncov_h()
2288 __lsx_vsigncov_w(__m128i _1, __m128i _2) { in __lsx_vsigncov_w()
2294 __lsx_vsigncov_d(__m128i _1, __m128i _2) { in __lsx_vsigncov_d()
2300 __lsx_vfmadd_s(__m128 _1, __m128 _2, __m128 _3) { in __lsx_vfmadd_s()
2306 __lsx_vfmadd_d(__m128d _1, __m128d _2, __m128d _3) { in __lsx_vfmadd_d()
2312 __lsx_vfmsub_s(__m128 _1, __m128 _2, __m128 _3) { in __lsx_vfmsub_s()
2318 __lsx_vfmsub_d(__m128d _1, __m128d _2, __m128d _3) { in __lsx_vfmsub_d()
2324 __lsx_vfnmadd_s(__m128 _1, __m128 _2, __m128 _3) { in __lsx_vfnmadd_s()
2330 __lsx_vfnmadd_d(__m128d _1, __m128d _2, __m128d _3) { in __lsx_vfnmadd_d()
2336 __lsx_vfnmsub_s(__m128 _1, __m128 _2, __m128 _3) { in __lsx_vfnmsub_s()
2342 __lsx_vfnmsub_d(__m128d _1, __m128d _2, __m128d _3) { in __lsx_vfnmsub_d()
2348 __lsx_vftintrne_w_s(__m128 _1) { in __lsx_vftintrne_w_s()
2354 __lsx_vftintrne_l_d(__m128d _1) { in __lsx_vftintrne_l_d()
2360 __lsx_vftintrp_w_s(__m128 _1) { in __lsx_vftintrp_w_s()
2366 __lsx_vftintrp_l_d(__m128d _1) { in __lsx_vftintrp_l_d()
2372 __lsx_vftintrm_w_s(__m128 _1) { in __lsx_vftintrm_w_s()
2378 __lsx_vftintrm_l_d(__m128d _1) { in __lsx_vftintrm_l_d()
2384 __lsx_vftint_w_d(__m128d _1, __m128d _2) { in __lsx_vftint_w_d()
2390 __lsx_vffint_s_l(__m128i _1, __m128i _2) { in __lsx_vffint_s_l()
2396 __lsx_vftintrz_w_d(__m128d _1, __m128d _2) { in __lsx_vftintrz_w_d()
2402 __lsx_vftintrp_w_d(__m128d _1, __m128d _2) { in __lsx_vftintrp_w_d()
2408 __lsx_vftintrm_w_d(__m128d _1, __m128d _2) { in __lsx_vftintrm_w_d()
2414 __lsx_vftintrne_w_d(__m128d _1, __m128d _2) { in __lsx_vftintrne_w_d()
2420 __lsx_vftintl_l_s(__m128 _1) { in __lsx_vftintl_l_s()
2426 __lsx_vftinth_l_s(__m128 _1) { in __lsx_vftinth_l_s()
2432 __lsx_vffinth_d_w(__m128i _1) { in __lsx_vffinth_d_w()
2438 __lsx_vffintl_d_w(__m128i _1) { in __lsx_vffintl_d_w()
2444 __lsx_vftintrzl_l_s(__m128 _1) { in __lsx_vftintrzl_l_s()
2450 __lsx_vftintrzh_l_s(__m128 _1) { in __lsx_vftintrzh_l_s()
2456 __lsx_vftintrpl_l_s(__m128 _1) { in __lsx_vftintrpl_l_s()
2462 __lsx_vftintrph_l_s(__m128 _1) { in __lsx_vftintrph_l_s()
2468 __lsx_vftintrml_l_s(__m128 _1) { in __lsx_vftintrml_l_s()
2474 __lsx_vftintrmh_l_s(__m128 _1) { in __lsx_vftintrmh_l_s()
2480 __lsx_vftintrnel_l_s(__m128 _1) { in __lsx_vftintrnel_l_s()
2486 __lsx_vftintrneh_l_s(__m128 _1) { in __lsx_vftintrneh_l_s()
2492 __lsx_vfrintrne_s(__m128 _1) { in __lsx_vfrintrne_s()
2498 __lsx_vfrintrne_d(__m128d _1) { in __lsx_vfrintrne_d()
2504 __lsx_vfrintrz_s(__m128 _1) { in __lsx_vfrintrz_s()
2510 __lsx_vfrintrz_d(__m128d _1) { in __lsx_vfrintrz_d()
2516 __lsx_vfrintrp_s(__m128 _1) { in __lsx_vfrintrp_s()
2522 __lsx_vfrintrp_d(__m128d _1) { in __lsx_vfrintrp_d()
2528 __lsx_vfrintrm_s(__m128 _1) { in __lsx_vfrintrm_s()
2534 __lsx_vfrintrm_d(__m128d _1) { in __lsx_vfrintrm_d()
2538 #define __lsx_vstelm_b(/*__m128i*/ _1, /*void **/ _2, /*si8*/ _3, /*idx*/ _4) \ argument
2541 #define __lsx_vstelm_h(/*__m128i*/ _1, /*void **/ _2, /*si8*/ _3, /*idx*/ _4) \ argument
2544 #define __lsx_vstelm_w(/*__m128i*/ _1, /*void **/ _2, /*si8*/ _3, /*idx*/ _4) \ argument
2547 #define __lsx_vstelm_d(/*__m128i*/ _1, /*void **/ _2, /*si8*/ _3, /*idx*/ _4) \ argument
2552 __lsx_vaddwev_d_w(__m128i _1, __m128i _2) { in __lsx_vaddwev_d_w()
2558 __lsx_vaddwev_w_h(__m128i _1, __m128i _2) { in __lsx_vaddwev_w_h()
2564 __lsx_vaddwev_h_b(__m128i _1, __m128i _2) { in __lsx_vaddwev_h_b()
2570 __lsx_vaddwod_d_w(__m128i _1, __m128i _2) { in __lsx_vaddwod_d_w()
2576 __lsx_vaddwod_w_h(__m128i _1, __m128i _2) { in __lsx_vaddwod_w_h()
2582 __lsx_vaddwod_h_b(__m128i _1, __m128i _2) { in __lsx_vaddwod_h_b()
2588 __lsx_vaddwev_d_wu(__m128i _1, __m128i _2) { in __lsx_vaddwev_d_wu()
2594 __lsx_vaddwev_w_hu(__m128i _1, __m128i _2) { in __lsx_vaddwev_w_hu()
2600 __lsx_vaddwev_h_bu(__m128i _1, __m128i _2) { in __lsx_vaddwev_h_bu()
2606 __lsx_vaddwod_d_wu(__m128i _1, __m128i _2) { in __lsx_vaddwod_d_wu()
2612 __lsx_vaddwod_w_hu(__m128i _1, __m128i _2) { in __lsx_vaddwod_w_hu()
2618 __lsx_vaddwod_h_bu(__m128i _1, __m128i _2) { in __lsx_vaddwod_h_bu()
2624 __lsx_vaddwev_d_wu_w(__m128i _1, __m128i _2) { in __lsx_vaddwev_d_wu_w()
2630 __lsx_vaddwev_w_hu_h(__m128i _1, __m128i _2) { in __lsx_vaddwev_w_hu_h()
2636 __lsx_vaddwev_h_bu_b(__m128i _1, __m128i _2) { in __lsx_vaddwev_h_bu_b()
2642 __lsx_vaddwod_d_wu_w(__m128i _1, __m128i _2) { in __lsx_vaddwod_d_wu_w()
2648 __lsx_vaddwod_w_hu_h(__m128i _1, __m128i _2) { in __lsx_vaddwod_w_hu_h()
2654 __lsx_vaddwod_h_bu_b(__m128i _1, __m128i _2) { in __lsx_vaddwod_h_bu_b()
2660 __lsx_vsubwev_d_w(__m128i _1, __m128i _2) { in __lsx_vsubwev_d_w()
2666 __lsx_vsubwev_w_h(__m128i _1, __m128i _2) { in __lsx_vsubwev_w_h()
2672 __lsx_vsubwev_h_b(__m128i _1, __m128i _2) { in __lsx_vsubwev_h_b()
2678 __lsx_vsubwod_d_w(__m128i _1, __m128i _2) { in __lsx_vsubwod_d_w()
2684 __lsx_vsubwod_w_h(__m128i _1, __m128i _2) { in __lsx_vsubwod_w_h()
2690 __lsx_vsubwod_h_b(__m128i _1, __m128i _2) { in __lsx_vsubwod_h_b()
2696 __lsx_vsubwev_d_wu(__m128i _1, __m128i _2) { in __lsx_vsubwev_d_wu()
2702 __lsx_vsubwev_w_hu(__m128i _1, __m128i _2) { in __lsx_vsubwev_w_hu()
2708 __lsx_vsubwev_h_bu(__m128i _1, __m128i _2) { in __lsx_vsubwev_h_bu()
2714 __lsx_vsubwod_d_wu(__m128i _1, __m128i _2) { in __lsx_vsubwod_d_wu()
2720 __lsx_vsubwod_w_hu(__m128i _1, __m128i _2) { in __lsx_vsubwod_w_hu()
2726 __lsx_vsubwod_h_bu(__m128i _1, __m128i _2) { in __lsx_vsubwod_h_bu()
2732 __lsx_vaddwev_q_d(__m128i _1, __m128i _2) { in __lsx_vaddwev_q_d()
2738 __lsx_vaddwod_q_d(__m128i _1, __m128i _2) { in __lsx_vaddwod_q_d()
2744 __lsx_vaddwev_q_du(__m128i _1, __m128i _2) { in __lsx_vaddwev_q_du()
2750 __lsx_vaddwod_q_du(__m128i _1, __m128i _2) { in __lsx_vaddwod_q_du()
2756 __lsx_vsubwev_q_d(__m128i _1, __m128i _2) { in __lsx_vsubwev_q_d()
2762 __lsx_vsubwod_q_d(__m128i _1, __m128i _2) { in __lsx_vsubwod_q_d()
2768 __lsx_vsubwev_q_du(__m128i _1, __m128i _2) { in __lsx_vsubwev_q_du()
2774 __lsx_vsubwod_q_du(__m128i _1, __m128i _2) { in __lsx_vsubwod_q_du()
2780 __lsx_vaddwev_q_du_d(__m128i _1, __m128i _2) { in __lsx_vaddwev_q_du_d()
2786 __lsx_vaddwod_q_du_d(__m128i _1, __m128i _2) { in __lsx_vaddwod_q_du_d()
2792 __lsx_vmulwev_d_w(__m128i _1, __m128i _2) { in __lsx_vmulwev_d_w()
2798 __lsx_vmulwev_w_h(__m128i _1, __m128i _2) { in __lsx_vmulwev_w_h()
2804 __lsx_vmulwev_h_b(__m128i _1, __m128i _2) { in __lsx_vmulwev_h_b()
2810 __lsx_vmulwod_d_w(__m128i _1, __m128i _2) { in __lsx_vmulwod_d_w()
2816 __lsx_vmulwod_w_h(__m128i _1, __m128i _2) { in __lsx_vmulwod_w_h()
2822 __lsx_vmulwod_h_b(__m128i _1, __m128i _2) { in __lsx_vmulwod_h_b()
2828 __lsx_vmulwev_d_wu(__m128i _1, __m128i _2) { in __lsx_vmulwev_d_wu()
2834 __lsx_vmulwev_w_hu(__m128i _1, __m128i _2) { in __lsx_vmulwev_w_hu()
2840 __lsx_vmulwev_h_bu(__m128i _1, __m128i _2) { in __lsx_vmulwev_h_bu()
2846 __lsx_vmulwod_d_wu(__m128i _1, __m128i _2) { in __lsx_vmulwod_d_wu()
2852 __lsx_vmulwod_w_hu(__m128i _1, __m128i _2) { in __lsx_vmulwod_w_hu()
2858 __lsx_vmulwod_h_bu(__m128i _1, __m128i _2) { in __lsx_vmulwod_h_bu()
2864 __lsx_vmulwev_d_wu_w(__m128i _1, __m128i _2) { in __lsx_vmulwev_d_wu_w()
2870 __lsx_vmulwev_w_hu_h(__m128i _1, __m128i _2) { in __lsx_vmulwev_w_hu_h()
2876 __lsx_vmulwev_h_bu_b(__m128i _1, __m128i _2) { in __lsx_vmulwev_h_bu_b()
2882 __lsx_vmulwod_d_wu_w(__m128i _1, __m128i _2) { in __lsx_vmulwod_d_wu_w()
2888 __lsx_vmulwod_w_hu_h(__m128i _1, __m128i _2) { in __lsx_vmulwod_w_hu_h()
2894 __lsx_vmulwod_h_bu_b(__m128i _1, __m128i _2) { in __lsx_vmulwod_h_bu_b()
2900 __lsx_vmulwev_q_d(__m128i _1, __m128i _2) { in __lsx_vmulwev_q_d()
2906 __lsx_vmulwod_q_d(__m128i _1, __m128i _2) { in __lsx_vmulwod_q_d()
2912 __lsx_vmulwev_q_du(__m128i _1, __m128i _2) { in __lsx_vmulwev_q_du()
2918 __lsx_vmulwod_q_du(__m128i _1, __m128i _2) { in __lsx_vmulwod_q_du()
2924 __lsx_vmulwev_q_du_d(__m128i _1, __m128i _2) { in __lsx_vmulwev_q_du_d()
2930 __lsx_vmulwod_q_du_d(__m128i _1, __m128i _2) { in __lsx_vmulwod_q_du_d()
2936 __lsx_vhaddw_q_d(__m128i _1, __m128i _2) { in __lsx_vhaddw_q_d()
2942 __lsx_vhaddw_qu_du(__m128i _1, __m128i _2) { in __lsx_vhaddw_qu_du()
2948 __lsx_vhsubw_q_d(__m128i _1, __m128i _2) { in __lsx_vhsubw_q_d()
2954 __lsx_vhsubw_qu_du(__m128i _1, __m128i _2) { in __lsx_vhsubw_qu_du()
2960 __lsx_vmaddwev_d_w(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwev_d_w()
2966 __lsx_vmaddwev_w_h(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwev_w_h()
2972 __lsx_vmaddwev_h_b(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwev_h_b()
2978 __lsx_vmaddwev_d_wu(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwev_d_wu()
2984 __lsx_vmaddwev_w_hu(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwev_w_hu()
2990 __lsx_vmaddwev_h_bu(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwev_h_bu()
2996 __lsx_vmaddwod_d_w(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwod_d_w()
3002 __lsx_vmaddwod_w_h(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwod_w_h()
3008 __lsx_vmaddwod_h_b(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwod_h_b()
3014 __lsx_vmaddwod_d_wu(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwod_d_wu()
3020 __lsx_vmaddwod_w_hu(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwod_w_hu()
3026 __lsx_vmaddwod_h_bu(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwod_h_bu()
3032 __lsx_vmaddwev_d_wu_w(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwev_d_wu_w()
3039 __lsx_vmaddwev_w_hu_h(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwev_w_hu_h()
3046 __lsx_vmaddwev_h_bu_b(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwev_h_bu_b()
3053 __lsx_vmaddwod_d_wu_w(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwod_d_wu_w()
3060 __lsx_vmaddwod_w_hu_h(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwod_w_hu_h()
3067 __lsx_vmaddwod_h_bu_b(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwod_h_bu_b()
3074 __lsx_vmaddwev_q_d(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwev_q_d()
3080 __lsx_vmaddwod_q_d(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwod_q_d()
3086 __lsx_vmaddwev_q_du(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwev_q_du()
3092 __lsx_vmaddwod_q_du(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwod_q_du()
3098 __lsx_vmaddwev_q_du_d(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwev_q_du_d()
3105 __lsx_vmaddwod_q_du_d(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vmaddwod_q_du_d()
3112 __lsx_vrotr_b(__m128i _1, __m128i _2) { in __lsx_vrotr_b()
3118 __lsx_vrotr_h(__m128i _1, __m128i _2) { in __lsx_vrotr_h()
3124 __lsx_vrotr_w(__m128i _1, __m128i _2) { in __lsx_vrotr_w()
3130 __lsx_vrotr_d(__m128i _1, __m128i _2) { in __lsx_vrotr_d()
3136 __lsx_vadd_q(__m128i _1, __m128i _2) { in __lsx_vadd_q()
3142 __lsx_vsub_q(__m128i _1, __m128i _2) { in __lsx_vsub_q()
3146 #define __lsx_vldrepl_b(/*void **/ _1, /*si12*/ _2) \ argument
3149 #define __lsx_vldrepl_h(/*void **/ _1, /*si11*/ _2) \ argument
3152 #define __lsx_vldrepl_w(/*void **/ _1, /*si10*/ _2) \ argument
3155 #define __lsx_vldrepl_d(/*void **/ _1, /*si9*/ _2) \ argument
3160 __lsx_vmskgez_b(__m128i _1) { in __lsx_vmskgez_b()
3166 __lsx_vmsknz_b(__m128i _1) { in __lsx_vmsknz_b()
3172 __lsx_vexth_h_b(__m128i _1) { in __lsx_vexth_h_b()
3178 __lsx_vexth_w_h(__m128i _1) { in __lsx_vexth_w_h()
3184 __lsx_vexth_d_w(__m128i _1) { in __lsx_vexth_d_w()
3190 __lsx_vexth_q_d(__m128i _1) { in __lsx_vexth_q_d()
3196 __lsx_vexth_hu_bu(__m128i _1) { in __lsx_vexth_hu_bu()
3202 __lsx_vexth_wu_hu(__m128i _1) { in __lsx_vexth_wu_hu()
3208 __lsx_vexth_du_wu(__m128i _1) { in __lsx_vexth_du_wu()
3214 __lsx_vexth_qu_du(__m128i _1) { in __lsx_vexth_qu_du()
3218 #define __lsx_vrotri_b(/*__m128i*/ _1, /*ui3*/ _2) \ argument
3221 #define __lsx_vrotri_h(/*__m128i*/ _1, /*ui4*/ _2) \ argument
3224 #define __lsx_vrotri_w(/*__m128i*/ _1, /*ui5*/ _2) \ argument
3227 #define __lsx_vrotri_d(/*__m128i*/ _1, /*ui6*/ _2) \ argument
3232 __lsx_vextl_q_d(__m128i _1) { in __lsx_vextl_q_d()
3236 #define __lsx_vsrlni_b_h(/*__m128i*/ _1, /*__m128i*/ _2, /*ui4*/ _3) \ argument
3239 #define __lsx_vsrlni_h_w(/*__m128i*/ _1, /*__m128i*/ _2, /*ui5*/ _3) \ argument
3242 #define __lsx_vsrlni_w_d(/*__m128i*/ _1, /*__m128i*/ _2, /*ui6*/ _3) \ argument
3245 #define __lsx_vsrlni_d_q(/*__m128i*/ _1, /*__m128i*/ _2, /*ui7*/ _3) \ argument
3248 #define __lsx_vsrlrni_b_h(/*__m128i*/ _1, /*__m128i*/ _2, /*ui4*/ _3) \ argument
3251 #define __lsx_vsrlrni_h_w(/*__m128i*/ _1, /*__m128i*/ _2, /*ui5*/ _3) \ argument
3254 #define __lsx_vsrlrni_w_d(/*__m128i*/ _1, /*__m128i*/ _2, /*ui6*/ _3) \ argument
3257 #define __lsx_vsrlrni_d_q(/*__m128i*/ _1, /*__m128i*/ _2, /*ui7*/ _3) \ argument
3260 #define __lsx_vssrlni_b_h(/*__m128i*/ _1, /*__m128i*/ _2, /*ui4*/ _3) \ argument
3263 #define __lsx_vssrlni_h_w(/*__m128i*/ _1, /*__m128i*/ _2, /*ui5*/ _3) \ argument
3266 #define __lsx_vssrlni_w_d(/*__m128i*/ _1, /*__m128i*/ _2, /*ui6*/ _3) \ argument
3269 #define __lsx_vssrlni_d_q(/*__m128i*/ _1, /*__m128i*/ _2, /*ui7*/ _3) \ argument
3272 #define __lsx_vssrlni_bu_h(/*__m128i*/ _1, /*__m128i*/ _2, /*ui4*/ _3) \ argument
3275 #define __lsx_vssrlni_hu_w(/*__m128i*/ _1, /*__m128i*/ _2, /*ui5*/ _3) \ argument
3278 #define __lsx_vssrlni_wu_d(/*__m128i*/ _1, /*__m128i*/ _2, /*ui6*/ _3) \ argument
3281 #define __lsx_vssrlni_du_q(/*__m128i*/ _1, /*__m128i*/ _2, /*ui7*/ _3) \ argument
3284 #define __lsx_vssrlrni_b_h(/*__m128i*/ _1, /*__m128i*/ _2, /*ui4*/ _3) \ argument
3287 #define __lsx_vssrlrni_h_w(/*__m128i*/ _1, /*__m128i*/ _2, /*ui5*/ _3) \ argument
3290 #define __lsx_vssrlrni_w_d(/*__m128i*/ _1, /*__m128i*/ _2, /*ui6*/ _3) \ argument
3293 #define __lsx_vssrlrni_d_q(/*__m128i*/ _1, /*__m128i*/ _2, /*ui7*/ _3) \ argument
3296 #define __lsx_vssrlrni_bu_h(/*__m128i*/ _1, /*__m128i*/ _2, /*ui4*/ _3) \ argument
3299 #define __lsx_vssrlrni_hu_w(/*__m128i*/ _1, /*__m128i*/ _2, /*ui5*/ _3) \ argument
3302 #define __lsx_vssrlrni_wu_d(/*__m128i*/ _1, /*__m128i*/ _2, /*ui6*/ _3) \ argument
3305 #define __lsx_vssrlrni_du_q(/*__m128i*/ _1, /*__m128i*/ _2, /*ui7*/ _3) \ argument
3308 #define __lsx_vsrani_b_h(/*__m128i*/ _1, /*__m128i*/ _2, /*ui4*/ _3) \ argument
3311 #define __lsx_vsrani_h_w(/*__m128i*/ _1, /*__m128i*/ _2, /*ui5*/ _3) \ argument
3314 #define __lsx_vsrani_w_d(/*__m128i*/ _1, /*__m128i*/ _2, /*ui6*/ _3) \ argument
3317 #define __lsx_vsrani_d_q(/*__m128i*/ _1, /*__m128i*/ _2, /*ui7*/ _3) \ argument
3320 #define __lsx_vsrarni_b_h(/*__m128i*/ _1, /*__m128i*/ _2, /*ui4*/ _3) \ argument
3323 #define __lsx_vsrarni_h_w(/*__m128i*/ _1, /*__m128i*/ _2, /*ui5*/ _3) \ argument
3326 #define __lsx_vsrarni_w_d(/*__m128i*/ _1, /*__m128i*/ _2, /*ui6*/ _3) \ argument
3329 #define __lsx_vsrarni_d_q(/*__m128i*/ _1, /*__m128i*/ _2, /*ui7*/ _3) \ argument
3332 #define __lsx_vssrani_b_h(/*__m128i*/ _1, /*__m128i*/ _2, /*ui4*/ _3) \ argument
3335 #define __lsx_vssrani_h_w(/*__m128i*/ _1, /*__m128i*/ _2, /*ui5*/ _3) \ argument
3338 #define __lsx_vssrani_w_d(/*__m128i*/ _1, /*__m128i*/ _2, /*ui6*/ _3) \ argument
3341 #define __lsx_vssrani_d_q(/*__m128i*/ _1, /*__m128i*/ _2, /*ui7*/ _3) \ argument
3344 #define __lsx_vssrani_bu_h(/*__m128i*/ _1, /*__m128i*/ _2, /*ui4*/ _3) \ argument
3347 #define __lsx_vssrani_hu_w(/*__m128i*/ _1, /*__m128i*/ _2, /*ui5*/ _3) \ argument
3350 #define __lsx_vssrani_wu_d(/*__m128i*/ _1, /*__m128i*/ _2, /*ui6*/ _3) \ argument
3353 #define __lsx_vssrani_du_q(/*__m128i*/ _1, /*__m128i*/ _2, /*ui7*/ _3) \ argument
3356 #define __lsx_vssrarni_b_h(/*__m128i*/ _1, /*__m128i*/ _2, /*ui4*/ _3) \ argument
3359 #define __lsx_vssrarni_h_w(/*__m128i*/ _1, /*__m128i*/ _2, /*ui5*/ _3) \ argument
3362 #define __lsx_vssrarni_w_d(/*__m128i*/ _1, /*__m128i*/ _2, /*ui6*/ _3) \ argument
3365 #define __lsx_vssrarni_d_q(/*__m128i*/ _1, /*__m128i*/ _2, /*ui7*/ _3) \ argument
3368 #define __lsx_vssrarni_bu_h(/*__m128i*/ _1, /*__m128i*/ _2, /*ui4*/ _3) \ argument
3371 #define __lsx_vssrarni_hu_w(/*__m128i*/ _1, /*__m128i*/ _2, /*ui5*/ _3) \ argument
3374 #define __lsx_vssrarni_wu_d(/*__m128i*/ _1, /*__m128i*/ _2, /*ui6*/ _3) \ argument
3377 #define __lsx_vssrarni_du_q(/*__m128i*/ _1, /*__m128i*/ _2, /*ui7*/ _3) \ argument
3380 #define __lsx_vpermi_w(/*__m128i*/ _1, /*__m128i*/ _2, /*ui8*/ _3) \ argument
3383 #define __lsx_vld(/*void **/ _1, /*si12*/ _2) \ argument
3386 #define __lsx_vst(/*__m128i*/ _1, /*void **/ _2, /*si12*/ _3) \ argument
3391 __lsx_vssrlrn_b_h(__m128i _1, __m128i _2) { in __lsx_vssrlrn_b_h()
3397 __lsx_vssrlrn_h_w(__m128i _1, __m128i _2) { in __lsx_vssrlrn_h_w()
3403 __lsx_vssrlrn_w_d(__m128i _1, __m128i _2) { in __lsx_vssrlrn_w_d()
3409 __lsx_vssrln_b_h(__m128i _1, __m128i _2) { in __lsx_vssrln_b_h()
3415 __lsx_vssrln_h_w(__m128i _1, __m128i _2) { in __lsx_vssrln_h_w()
3421 __lsx_vssrln_w_d(__m128i _1, __m128i _2) { in __lsx_vssrln_w_d()
3427 __lsx_vorn_v(__m128i _1, __m128i _2) { in __lsx_vorn_v()
3431 #define __lsx_vldi(/*i13*/ _1) ((__m128i)__builtin_lsx_vldi((_1))) argument
3435 __lsx_vshuf_b(__m128i _1, __m128i _2, __m128i _3) { in __lsx_vshuf_b()
3441 __lsx_vldx(void const *_1, long int _2) { in __lsx_vldx()
3447 __lsx_vstx(__m128i _1, void *_2, long int _3) { in __lsx_vstx()
3453 __lsx_vextl_qu_du(__m128i _1) { in __lsx_vextl_qu_du()
3457 #define __lsx_bnz_b(/*__m128i*/ _1) ((int)__builtin_lsx_bnz_b((v16u8)(_1))) argument
3459 #define __lsx_bnz_d(/*__m128i*/ _1) ((int)__builtin_lsx_bnz_d((v2u64)(_1))) argument
3461 #define __lsx_bnz_h(/*__m128i*/ _1) ((int)__builtin_lsx_bnz_h((v8u16)(_1))) argument
3463 #define __lsx_bnz_v(/*__m128i*/ _1) ((int)__builtin_lsx_bnz_v((v16u8)(_1))) argument
3465 #define __lsx_bnz_w(/*__m128i*/ _1) ((int)__builtin_lsx_bnz_w((v4u32)(_1))) argument
3467 #define __lsx_bz_b(/*__m128i*/ _1) ((int)__builtin_lsx_bz_b((v16u8)(_1))) argument
3469 #define __lsx_bz_d(/*__m128i*/ _1) ((int)__builtin_lsx_bz_d((v2u64)(_1))) argument
3471 #define __lsx_bz_h(/*__m128i*/ _1) ((int)__builtin_lsx_bz_h((v8u16)(_1))) argument
3473 #define __lsx_bz_v(/*__m128i*/ _1) ((int)__builtin_lsx_bz_v((v16u8)(_1))) argument
3475 #define __lsx_bz_w(/*__m128i*/ _1) ((int)__builtin_lsx_bz_w((v4u32)(_1))) argument
3479 __lsx_vfcmp_caf_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_caf_d()
3485 __lsx_vfcmp_caf_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_caf_s()
3491 __lsx_vfcmp_ceq_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_ceq_d()
3497 __lsx_vfcmp_ceq_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_ceq_s()
3503 __lsx_vfcmp_cle_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_cle_d()
3509 __lsx_vfcmp_cle_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_cle_s()
3515 __lsx_vfcmp_clt_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_clt_d()
3521 __lsx_vfcmp_clt_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_clt_s()
3527 __lsx_vfcmp_cne_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_cne_d()
3533 __lsx_vfcmp_cne_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_cne_s()
3539 __lsx_vfcmp_cor_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_cor_d()
3545 __lsx_vfcmp_cor_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_cor_s()
3551 __lsx_vfcmp_cueq_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_cueq_d()
3557 __lsx_vfcmp_cueq_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_cueq_s()
3563 __lsx_vfcmp_cule_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_cule_d()
3569 __lsx_vfcmp_cule_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_cule_s()
3575 __lsx_vfcmp_cult_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_cult_d()
3581 __lsx_vfcmp_cult_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_cult_s()
3587 __lsx_vfcmp_cun_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_cun_d()
3593 __lsx_vfcmp_cune_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_cune_d()
3599 __lsx_vfcmp_cune_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_cune_s()
3605 __lsx_vfcmp_cun_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_cun_s()
3611 __lsx_vfcmp_saf_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_saf_d()
3617 __lsx_vfcmp_saf_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_saf_s()
3623 __lsx_vfcmp_seq_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_seq_d()
3629 __lsx_vfcmp_seq_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_seq_s()
3635 __lsx_vfcmp_sle_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_sle_d()
3641 __lsx_vfcmp_sle_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_sle_s()
3647 __lsx_vfcmp_slt_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_slt_d()
3653 __lsx_vfcmp_slt_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_slt_s()
3659 __lsx_vfcmp_sne_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_sne_d()
3665 __lsx_vfcmp_sne_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_sne_s()
3671 __lsx_vfcmp_sor_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_sor_d()
3677 __lsx_vfcmp_sor_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_sor_s()
3683 __lsx_vfcmp_sueq_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_sueq_d()
3689 __lsx_vfcmp_sueq_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_sueq_s()
3695 __lsx_vfcmp_sule_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_sule_d()
3701 __lsx_vfcmp_sule_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_sule_s()
3707 __lsx_vfcmp_sult_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_sult_d()
3713 __lsx_vfcmp_sult_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_sult_s()
3719 __lsx_vfcmp_sun_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_sun_d()
3725 __lsx_vfcmp_sune_d(__m128d _1, __m128d _2) { in __lsx_vfcmp_sune_d()
3731 __lsx_vfcmp_sune_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_sune_s()
3737 __lsx_vfcmp_sun_s(__m128 _1, __m128 _2) { in __lsx_vfcmp_sun_s()
3741 #define __lsx_vrepli_b(/*si10*/ _1) ((__m128i)__builtin_lsx_vrepli_b((_1))) argument
3743 #define __lsx_vrepli_d(/*si10*/ _1) ((__m128i)__builtin_lsx_vrepli_d((_1))) argument
3745 #define __lsx_vrepli_h(/*si10*/ _1) ((__m128i)__builtin_lsx_vrepli_h((_1))) argument
3747 #define __lsx_vrepli_w(/*si10*/ _1) ((__m128i)__builtin_lsx_vrepli_w((_1))) argument