Lines Matching defs:B
279 #define _mm_comi_round_sh(A, B, P, R) \
280 __builtin_ia32_vcomish((__v8hf)A, (__v8hf)B, (int)(P), (int)(R))
282 #define _mm_comi_sh(A, B, pred) \
283 _mm_comi_round_sh((A), (B), (pred), _MM_FROUND_CUR_DIRECTION)
375 #define _mm512_add_round_ph(A, B, R) \
377 (__v32hf)(__m512h)(B), (int)(R)))
379 #define _mm512_mask_add_round_ph(W, U, A, B, R) \
381 (__mmask32)(U), (__v32hf)_mm512_add_round_ph((A), (B), (R)), \
384 #define _mm512_maskz_add_round_ph(U, A, B, R) \
386 (__mmask32)(U), (__v32hf)_mm512_add_round_ph((A), (B), (R)), \
407 #define _mm512_sub_round_ph(A, B, R) \
409 (__v32hf)(__m512h)(B), (int)(R)))
411 #define _mm512_mask_sub_round_ph(W, U, A, B, R) \
413 (__mmask32)(U), (__v32hf)_mm512_sub_round_ph((A), (B), (R)), \
416 #define _mm512_maskz_sub_round_ph(U, A, B, R) \
418 (__mmask32)(U), (__v32hf)_mm512_sub_round_ph((A), (B), (R)), \
439 #define _mm512_mul_round_ph(A, B, R) \
441 (__v32hf)(__m512h)(B), (int)(R)))
443 #define _mm512_mask_mul_round_ph(W, U, A, B, R) \
445 (__mmask32)(U), (__v32hf)_mm512_mul_round_ph((A), (B), (R)), \
448 #define _mm512_maskz_mul_round_ph(U, A, B, R) \
450 (__mmask32)(U), (__v32hf)_mm512_mul_round_ph((A), (B), (R)), \
471 #define _mm512_div_round_ph(A, B, R) \
473 (__v32hf)(__m512h)(B), (int)(R)))
475 #define _mm512_mask_div_round_ph(W, U, A, B, R) \
477 (__mmask32)(U), (__v32hf)_mm512_div_round_ph((A), (B), (R)), \
480 #define _mm512_maskz_div_round_ph(U, A, B, R) \
482 (__mmask32)(U), (__v32hf)_mm512_div_round_ph((A), (B), (R)), \
504 #define _mm512_min_round_ph(A, B, R) \
506 (__v32hf)(__m512h)(B), (int)(R)))
508 #define _mm512_mask_min_round_ph(W, U, A, B, R) \
510 (__mmask32)(U), (__v32hf)_mm512_min_round_ph((A), (B), (R)), \
513 #define _mm512_maskz_min_round_ph(U, A, B, R) \
515 (__mmask32)(U), (__v32hf)_mm512_min_round_ph((A), (B), (R)), \
537 #define _mm512_max_round_ph(A, B, R) \
539 (__v32hf)(__m512h)(B), (int)(R)))
541 #define _mm512_mask_max_round_ph(W, U, A, B, R) \
543 (__mmask32)(U), (__v32hf)_mm512_max_round_ph((A), (B), (R)), \
546 #define _mm512_maskz_max_round_ph(U, A, B, R) \
548 (__mmask32)(U), (__v32hf)_mm512_max_round_ph((A), (B), (R)), \
593 #define _mm_add_round_sh(A, B, R) \
595 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
598 #define _mm_mask_add_round_sh(W, U, A, B, R) \
600 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)(__m128h)(W), \
603 #define _mm_maskz_add_round_sh(U, A, B, R) \
605 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
629 #define _mm_sub_round_sh(A, B, R) \
631 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
634 #define _mm_mask_sub_round_sh(W, U, A, B, R) \
636 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)(__m128h)(W), \
639 #define _mm_maskz_sub_round_sh(U, A, B, R) \
641 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
665 #define _mm_mul_round_sh(A, B, R) \
667 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
670 #define _mm_mask_mul_round_sh(W, U, A, B, R) \
672 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)(__m128h)(W), \
675 #define _mm_maskz_mul_round_sh(U, A, B, R) \
677 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
701 #define _mm_div_round_sh(A, B, R) \
703 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
706 #define _mm_mask_div_round_sh(W, U, A, B, R) \
708 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)(__m128h)(W), \
711 #define _mm_maskz_div_round_sh(U, A, B, R) \
713 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
740 #define _mm_min_round_sh(A, B, R) \
742 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
745 #define _mm_mask_min_round_sh(W, U, A, B, R) \
747 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)(__m128h)(W), \
750 #define _mm_maskz_min_round_sh(U, A, B, R) \
752 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
779 #define _mm_max_round_sh(A, B, R) \
781 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
784 #define _mm_mask_max_round_sh(W, U, A, B, R) \
786 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)(__m128h)(W), \
789 #define _mm_maskz_max_round_sh(U, A, B, R) \
791 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
794 #define _mm512_cmp_round_ph_mask(A, B, P, R) \
796 (__v32hf)(__m512h)(B), (int)(P), \
799 #define _mm512_mask_cmp_round_ph_mask(U, A, B, P, R) \
801 (__v32hf)(__m512h)(B), (int)(P), \
804 #define _mm512_cmp_ph_mask(A, B, P) \
805 _mm512_cmp_round_ph_mask((A), (B), (P), _MM_FROUND_CUR_DIRECTION)
807 #define _mm512_mask_cmp_ph_mask(U, A, B, P) \
808 _mm512_mask_cmp_round_ph_mask((U), (A), (B), (P), _MM_FROUND_CUR_DIRECTION)
1008 #define _mm512_getmant_ph(A, B, C) \
1010 (__v32hf)(__m512h)(A), (int)(((C) << 2) | (B)), \
1014 #define _mm512_mask_getmant_ph(W, U, A, B, C) \
1016 (__v32hf)(__m512h)(A), (int)(((C) << 2) | (B)), (__v32hf)(__m512h)(W), \
1019 #define _mm512_maskz_getmant_ph(U, A, B, C) \
1021 (__v32hf)(__m512h)(A), (int)(((C) << 2) | (B)), \
1024 #define _mm512_getmant_round_ph(A, B, C, R) \
1026 (__v32hf)(__m512h)(A), (int)(((C) << 2) | (B)), \
1029 #define _mm512_mask_getmant_round_ph(W, U, A, B, C, R) \
1031 (__v32hf)(__m512h)(A), (int)(((C) << 2) | (B)), (__v32hf)(__m512h)(W), \
1034 #define _mm512_maskz_getmant_round_ph(U, A, B, C, R) \
1036 (__v32hf)(__m512h)(A), (int)(((C) << 2) | (B)), \
1093 #define _mm512_scalef_round_ph(A, B, R) \
1095 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), \
1098 #define _mm512_mask_scalef_round_ph(W, U, A, B, R) \
1100 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), (__v32hf)(__m512h)(W), \
1103 #define _mm512_maskz_scalef_round_ph(U, A, B, R) \
1105 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), \
1108 #define _mm512_roundscale_ph(A, B) \
1110 (__v32hf)(__m512h)(A), (int)(B), (__v32hf)(__m512h)(A), (__mmask32)-1, \
1113 #define _mm512_mask_roundscale_ph(A, B, C, imm) \
1116 (__mmask32)(B), _MM_FROUND_CUR_DIRECTION))
1118 #define _mm512_maskz_roundscale_ph(A, B, imm) \
1120 (__v32hf)(__m512h)(B), (int)(imm), (__v32hf)_mm512_setzero_ph(), \
1123 #define _mm512_mask_roundscale_round_ph(A, B, C, imm, R) \
1126 (__mmask32)(B), (int)(R)))
1128 #define _mm512_maskz_roundscale_round_ph(A, B, imm, R) \
1129 ((__m512h)__builtin_ia32_rndscaleph_mask((__v32hf)(__m512h)(B), (int)(imm), \
1209 #define _mm_getmant_round_sh(A, B, C, D, R) \
1211 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (int)(((D) << 2) | (C)), \
1214 #define _mm_getmant_sh(A, B, C, D) \
1216 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (int)(((D) << 2) | (C)), \
1219 #define _mm_mask_getmant_sh(W, U, A, B, C, D) \
1221 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (int)(((D) << 2) | (C)), \
1224 #define _mm_mask_getmant_round_sh(W, U, A, B, C, D, R) \
1226 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (int)(((D) << 2) | (C)), \
1229 #define _mm_maskz_getmant_sh(U, A, B, C, D) \
1231 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (int)(((D) << 2) | (C)), \
1234 #define _mm_maskz_getmant_round_sh(U, A, B, C, D, R) \
1236 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (int)(((D) << 2) | (C)), \
1239 #define _mm_getexp_round_sh(A, B, R) \
1241 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
1258 #define _mm_mask_getexp_round_sh(W, U, A, B, R) \
1260 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)(__m128h)(W), \
1270 #define _mm_maskz_getexp_round_sh(U, A, B, R) \
1272 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
1275 #define _mm_scalef_round_sh(A, B, R) \
1277 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
1294 #define _mm_mask_scalef_round_sh(W, U, A, B, R) \
1296 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)(__m128h)(W), \
1306 #define _mm_maskz_scalef_round_sh(U, A, B, R) \
1308 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
1311 #define _mm_roundscale_round_sh(A, B, imm, R) \
1313 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
1316 #define _mm_roundscale_sh(A, B, imm) \
1318 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
1321 #define _mm_mask_roundscale_sh(W, U, A, B, I) \
1323 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)(__m128h)(W), \
1326 #define _mm_mask_roundscale_round_sh(W, U, A, B, I, R) \
1328 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)(__m128h)(W), \
1331 #define _mm_maskz_roundscale_sh(U, A, B, I) \
1333 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
1336 #define _mm_maskz_roundscale_round_sh(U, A, B, I, R) \
1338 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
1341 #define _mm_reduce_sh(A, B, C) \
1343 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
1346 #define _mm_mask_reduce_sh(W, U, A, B, C) \
1348 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)(__m128h)(W), \
1351 #define _mm_maskz_reduce_sh(U, A, B, C) \
1353 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
1356 #define _mm_reduce_round_sh(A, B, C, R) \
1358 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
1361 #define _mm_mask_reduce_round_sh(W, U, A, B, C, R) \
1363 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)(__m128h)(W), \
1366 #define _mm_maskz_reduce_round_sh(U, A, B, C, R) \
1368 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
1405 #define _mm_sqrt_round_sh(A, B, R) \
1407 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
1410 #define _mm_mask_sqrt_round_sh(W, U, A, B, R) \
1412 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)(__m128h)(W), \
1415 #define _mm_maskz_sqrt_round_sh(U, A, B, R) \
1417 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)_mm_setzero_ph(), \
1522 #define _mm_cvt_roundsh_ss(A, B, R) \
1523 ((__m128)__builtin_ia32_vcvtsh2ss_round_mask((__v4sf)(A), (__v8hf)(B), \
1527 #define _mm_mask_cvt_roundsh_ss(W, U, A, B, R) \
1529 (__v4sf)(A), (__v8hf)(B), (__v4sf)(W), (__mmask8)(U), (int)(R)))
1531 #define _mm_maskz_cvt_roundsh_ss(U, A, B, R) \
1532 ((__m128)__builtin_ia32_vcvtsh2ss_round_mask((__v4sf)(A), (__v8hf)(B), \
1560 #define _mm_cvt_roundss_sh(A, B, R) \
1561 ((__m128h)__builtin_ia32_vcvtss2sh_round_mask((__v8hf)(A), (__v4sf)(B), \
1565 #define _mm_mask_cvt_roundss_sh(W, U, A, B, R) \
1567 (__v8hf)(A), (__v4sf)(B), (__v8hf)(W), (__mmask8)(U), (int)(R)))
1569 #define _mm_maskz_cvt_roundss_sh(U, A, B, R) \
1570 ((__m128h)__builtin_ia32_vcvtss2sh_round_mask((__v8hf)(A), (__v4sf)(B), \
1598 #define _mm_cvt_roundsd_sh(A, B, R) \
1599 ((__m128h)__builtin_ia32_vcvtsd2sh_round_mask((__v8hf)(A), (__v2df)(B), \
1603 #define _mm_mask_cvt_roundsd_sh(W, U, A, B, R) \
1605 (__v8hf)(A), (__v2df)(B), (__v8hf)(W), (__mmask8)(U), (int)(R)))
1607 #define _mm_maskz_cvt_roundsd_sh(U, A, B, R) \
1608 ((__m128h)__builtin_ia32_vcvtsd2sh_round_mask((__v8hf)(A), (__v2df)(B), \
1635 #define _mm_cvt_roundsh_sd(A, B, R) \
1636 ((__m128d)__builtin_ia32_vcvtsh2sd_round_mask((__v2df)(A), (__v8hf)(B), \
1640 #define _mm_mask_cvt_roundsh_sd(W, U, A, B, R) \
1642 (__v2df)(A), (__v8hf)(B), (__v2df)(W), (__mmask8)(U), (int)(R)))
1644 #define _mm_maskz_cvt_roundsh_sd(U, A, B, R) \
1645 ((__m128d)__builtin_ia32_vcvtsh2sd_round_mask((__v2df)(A), (__v8hf)(B), \
2307 #define _mm_cvt_roundu32_sh(A, B, R) \
2308 ((__m128h)__builtin_ia32_vcvtusi2sh((__v8hf)(A), (unsigned int)(B), (int)(R)))
2317 #define _mm_cvt_roundu64_sh(A, B, R) \
2318 ((__m128h)__builtin_ia32_vcvtusi642sh((__v8hf)(A), (unsigned long long)(B), \
2328 #define _mm_cvt_roundi32_sh(A, B, R) \
2329 ((__m128h)__builtin_ia32_vcvtsi2sh((__v8hf)(A), (int)(B), (int)(R)))
2338 #define _mm_cvt_roundi64_sh(A, B, R) \
2339 ((__m128h)__builtin_ia32_vcvtsi642sh((__v8hf)(A), (long long)(B), (int)(R)))
2450 #define _mm512_fmadd_round_ph(A, B, C, R) \
2452 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), (__v32hf)(__m512h)(C), \
2455 #define _mm512_mask_fmadd_round_ph(A, U, B, C, R) \
2457 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), (__v32hf)(__m512h)(C), \
2460 #define _mm512_mask3_fmadd_round_ph(A, B, C, U, R) \
2462 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), (__v32hf)(__m512h)(C), \
2465 #define _mm512_maskz_fmadd_round_ph(U, A, B, C, R) \
2467 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), (__v32hf)(__m512h)(C), \
2470 #define _mm512_fmsub_round_ph(A, B, C, R) \
2472 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), -(__v32hf)(__m512h)(C), \
2475 #define _mm512_mask_fmsub_round_ph(A, U, B, C, R) \
2477 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), -(__v32hf)(__m512h)(C), \
2480 #define _mm512_maskz_fmsub_round_ph(U, A, B, C, R) \
2482 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), -(__v32hf)(__m512h)(C), \
2485 #define _mm512_fnmadd_round_ph(A, B, C, R) \
2487 (__v32hf)(__m512h)(A), -(__v32hf)(__m512h)(B), (__v32hf)(__m512h)(C), \
2490 #define _mm512_mask3_fnmadd_round_ph(A, B, C, U, R) \
2492 -(__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), (__v32hf)(__m512h)(C), \
2495 #define _mm512_maskz_fnmadd_round_ph(U, A, B, C, R) \
2497 -(__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), (__v32hf)(__m512h)(C), \
2500 #define _mm512_fnmsub_round_ph(A, B, C, R) \
2502 (__v32hf)(__m512h)(A), -(__v32hf)(__m512h)(B), -(__v32hf)(__m512h)(C), \
2505 #define _mm512_maskz_fnmsub_round_ph(U, A, B, C, R) \
2507 -(__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), -(__v32hf)(__m512h)(C), \
2598 #define _mm512_fmaddsub_round_ph(A, B, C, R) \
2600 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), (__v32hf)(__m512h)(C), \
2603 #define _mm512_mask_fmaddsub_round_ph(A, U, B, C, R) \
2605 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), (__v32hf)(__m512h)(C), \
2608 #define _mm512_mask3_fmaddsub_round_ph(A, B, C, U, R) \
2610 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), (__v32hf)(__m512h)(C), \
2613 #define _mm512_maskz_fmaddsub_round_ph(U, A, B, C, R) \
2615 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), (__v32hf)(__m512h)(C), \
2618 #define _mm512_fmsubadd_round_ph(A, B, C, R) \
2620 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), -(__v32hf)(__m512h)(C), \
2623 #define _mm512_mask_fmsubadd_round_ph(A, U, B, C, R) \
2625 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), -(__v32hf)(__m512h)(C), \
2628 #define _mm512_maskz_fmsubadd_round_ph(U, A, B, C, R) \
2630 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), -(__v32hf)(__m512h)(C), \
2682 #define _mm512_mask3_fmsub_round_ph(A, B, C, U, R) \
2684 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), (__v32hf)(__m512h)(C), \
2694 #define _mm512_mask3_fmsubadd_round_ph(A, B, C, U, R) \
2696 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), (__v32hf)(__m512h)(C), \
2706 #define _mm512_mask_fnmadd_round_ph(A, U, B, C, R) \
2708 (__v32hf)(__m512h)(A), -(__v32hf)(__m512h)(B), (__v32hf)(__m512h)(C), \
2718 #define _mm512_mask_fnmsub_round_ph(A, U, B, C, R) \
2720 (__v32hf)(__m512h)(A), -(__v32hf)(__m512h)(B), -(__v32hf)(__m512h)(C), \
2723 #define _mm512_mask3_fnmsub_round_ph(A, B, C, U, R) \
2725 -(__v32hf)(__m512h)(A), (__v32hf)(__m512h)(B), (__v32hf)(__m512h)(C), \
2757 #define _mm_fmadd_round_sh(A, B, C, R) \
2759 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)(__m128h)(C), \
2762 #define _mm_mask_fmadd_round_sh(W, U, A, B, R) \
2764 (__v8hf)(__m128h)(W), (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), \
2774 #define _mm_maskz_fmadd_round_sh(U, A, B, C, R) \
2776 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), (__v8hf)(__m128h)(C), \
2808 #define _mm_fmsub_round_sh(A, B, C, R) \
2810 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), -(__v8hf)(__m128h)(C), \
2813 #define _mm_mask_fmsub_round_sh(W, U, A, B, R) \
2815 (__v8hf)(__m128h)(W), (__v8hf)(__m128h)(A), -(__v8hf)(__m128h)(B), \
2825 #define _mm_maskz_fmsub_round_sh(U, A, B, C, R) \
2827 (__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), -(__v8hf)(__m128h)(C), \
2855 #define _mm_fnmadd_round_sh(A, B, C, R) \
2857 (__v8hf)(__m128h)(A), -(__v8hf)(__m128h)(B), (__v8hf)(__m128h)(C), \
2860 #define _mm_mask_fnmadd_round_sh(W, U, A, B, R) \
2862 (__v8hf)(__m128h)(W), -(__v8hf)(__m128h)(A), (__v8hf)(__m128h)(B), \
2872 #define _mm_maskz_fnmadd_round_sh(U, A, B, C, R) \
2874 (__v8hf)(__m128h)(A), -(__v8hf)(__m128h)(B), (__v8hf)(__m128h)(C), \
2902 #define _mm_fnmsub_round_sh(A, B, C, R) \
2904 (__v8hf)(__m128h)(A), -(__v8hf)(__m128h)(B), -(__v8hf)(__m128h)(C), \
2907 #define _mm_mask_fnmsub_round_sh(W, U, A, B, R) \
2909 (__v8hf)(__m128h)(W), -(__v8hf)(__m128h)(A), -(__v8hf)(__m128h)(B), \
2919 #define _mm_maskz_fnmsub_round_sh(U, A, B, C, R) \
2921 (__v8hf)(__m128h)(A), -(__v8hf)(__m128h)(B), -(__v8hf)(__m128h)(C), \
2963 #define _mm_fcmadd_round_sch(A, B, C, R) \
2965 (__v4sf)(__m128h)(A), (__v4sf)(__m128h)(B), (__v4sf)(__m128h)(C), \
2968 #define _mm_mask_fcmadd_round_sch(A, U, B, C, R) \
2970 (__v4sf)(__m128h)(A), (__v4sf)(__m128h)(B), (__v4sf)(__m128h)(C), \
2973 #define _mm_maskz_fcmadd_round_sch(U, A, B, C, R) \
2975 (__v4sf)(__m128h)(A), (__v4sf)(__m128h)(B), (__v4sf)(__m128h)(C), \
2978 #define _mm_mask3_fcmadd_round_sch(A, B, C, U, R) \
2980 (__v4sf)(__m128h)(A), (__v4sf)(__m128h)(B), (__v4sf)(__m128h)(C), \
3010 #define _mm_fmadd_round_sch(A, B, C, R) \
3012 (__v4sf)(__m128h)(A), (__v4sf)(__m128h)(B), (__v4sf)(__m128h)(C), \
3015 #define _mm_mask_fmadd_round_sch(A, U, B, C, R) \
3017 (__v4sf)(__m128h)(A), (__v4sf)(__m128h)(B), (__v4sf)(__m128h)(C), \
3020 #define _mm_maskz_fmadd_round_sch(U, A, B, C, R) \
3022 (__v4sf)(__m128h)(A), (__v4sf)(__m128h)(B), (__v4sf)(__m128h)(C), \
3025 #define _mm_mask3_fmadd_round_sch(A, B, C, U, R) \
3027 (__v4sf)(__m128h)(A), (__v4sf)(__m128h)(B), (__v4sf)(__m128h)(C), \
3051 #define _mm_fcmul_round_sch(A, B, R) \
3053 (__v4sf)(__m128h)(A), (__v4sf)(__m128h)(B), \
3056 #define _mm_mask_fcmul_round_sch(W, U, A, B, R) \
3058 (__v4sf)(__m128h)(A), (__v4sf)(__m128h)(B), (__v4sf)(__m128h)(W), \
3061 #define _mm_maskz_fcmul_round_sch(U, A, B, R) \
3063 (__v4sf)(__m128h)(A), (__v4sf)(__m128h)(B), \
3089 #define _mm_fmul_round_sch(A, B, R) \
3091 (__v4sf)(__m128h)(A), (__v4sf)(__m128h)(B), \
3094 #define _mm_mask_fmul_round_sch(W, U, A, B, R) \
3096 (__v4sf)(__m128h)(A), (__v4sf)(__m128h)(B), (__v4sf)(__m128h)(W), \
3099 #define _mm_maskz_fmul_round_sch(U, A, B, R) \
3101 (__v4sf)(__m128h)(A), (__v4sf)(__m128h)(B), \
3125 #define _mm512_fcmul_round_pch(A, B, R) \
3127 (__v16sf)(__m512h)(A), (__v16sf)(__m512h)(B), \
3130 #define _mm512_mask_fcmul_round_pch(W, U, A, B, R) \
3132 (__v16sf)(__m512h)(A), (__v16sf)(__m512h)(B), (__v16sf)(__m512h)(W), \
3135 #define _mm512_maskz_fcmul_round_pch(U, A, B, R) \
3137 (__v16sf)(__m512h)(A), (__v16sf)(__m512h)(B), \
3161 #define _mm512_fmul_round_pch(A, B, R) \
3163 (__v16sf)(__m512h)(A), (__v16sf)(__m512h)(B), \
3166 #define _mm512_mask_fmul_round_pch(W, U, A, B, R) \
3168 (__v16sf)(__m512h)(A), (__v16sf)(__m512h)(B), (__v16sf)(__m512h)(W), \
3171 #define _mm512_maskz_fmul_round_pch(U, A, B, R) \
3173 (__v16sf)(__m512h)(A), (__v16sf)(__m512h)(B), \
3205 #define _mm512_fcmadd_round_pch(A, B, C, R) \
3207 (__v16sf)(__m512h)(A), (__v16sf)(__m512h)(B), (__v16sf)(__m512h)(C), \
3210 #define _mm512_mask_fcmadd_round_pch(A, U, B, C, R) \
3212 (__v16sf)(__m512h)(A), (__v16sf)(__m512h)(B), (__v16sf)(__m512h)(C), \
3215 #define _mm512_mask3_fcmadd_round_pch(A, B, C, U, R) \
3217 (__v16sf)(__m512h)(A), (__v16sf)(__m512h)(B), (__v16sf)(__m512h)(C), \
3220 #define _mm512_maskz_fcmadd_round_pch(U, A, B, C, R) \
3222 (__v16sf)(__m512h)(A), (__v16sf)(__m512h)(B), (__v16sf)(__m512h)(C), \
3254 #define _mm512_fmadd_round_pch(A, B, C, R) \
3256 (__v16sf)(__m512h)(A), (__v16sf)(__m512h)(B), (__v16sf)(__m512h)(C), \
3259 #define _mm512_mask_fmadd_round_pch(A, U, B, C, R) \
3261 (__v16sf)(__m512h)(A), (__v16sf)(__m512h)(B), (__v16sf)(__m512h)(C), \
3264 #define _mm512_mask3_fmadd_round_pch(A, B, C, U, R) \
3266 (__v16sf)(__m512h)(A), (__v16sf)(__m512h)(B), (__v16sf)(__m512h)(C), \
3269 #define _mm512_maskz_fmadd_round_pch(U, A, B, C, R) \
3271 (__v16sf)(__m512h)(A), (__v16sf)(__m512h)(B), (__v16sf)(__m512h)(C), \
3312 #define _mm512_mul_pch(A, B) _mm512_fmul_pch(A, B)
3313 #define _mm512_mask_mul_pch(W, U, A, B) _mm512_mask_fmul_pch(W, U, A, B)
3314 #define _mm512_maskz_mul_pch(U, A, B) _mm512_maskz_fmul_pch(U, A, B)
3315 #define _mm512_mul_round_pch(A, B, R) _mm512_fmul_round_pch(A, B, R)
3316 #define _mm512_mask_mul_round_pch(W, U, A, B, R) \
3317 _mm512_mask_fmul_round_pch(W, U, A, B, R)
3318 #define _mm512_maskz_mul_round_pch(U, A, B, R) \
3319 _mm512_maskz_fmul_round_pch(U, A, B, R)
3321 #define _mm512_cmul_pch(A, B) _mm512_fcmul_pch(A, B)
3322 #define _mm512_mask_cmul_pch(W, U, A, B) _mm512_mask_fcmul_pch(W, U, A, B)
3323 #define _mm512_maskz_cmul_pch(U, A, B) _mm512_maskz_fcmul_pch(U, A, B)
3324 #define _mm512_cmul_round_pch(A, B, R) _mm512_fcmul_round_pch(A, B, R)
3325 #define _mm512_mask_cmul_round_pch(W, U, A, B, R) \
3326 _mm512_mask_fcmul_round_pch(W, U, A, B, R)
3327 #define _mm512_maskz_cmul_round_pch(U, A, B, R) \
3328 _mm512_maskz_fcmul_round_pch(U, A, B, R)
3330 #define _mm_mul_sch(A, B) _mm_fmul_sch(A, B)
3331 #define _mm_mask_mul_sch(W, U, A, B) _mm_mask_fmul_sch(W, U, A, B)
3332 #define _mm_maskz_mul_sch(U, A, B) _mm_maskz_fmul_sch(U, A, B)
3333 #define _mm_mul_round_sch(A, B, R) _mm_fmul_round_sch(A, B, R)
3334 #define _mm_mask_mul_round_sch(W, U, A, B, R) \
3335 _mm_mask_fmul_round_sch(W, U, A, B, R)
3336 #define _mm_maskz_mul_round_sch(U, A, B, R) _mm_maskz_fmul_round_sch(U, A, B, R)
3338 #define _mm_cmul_sch(A, B) _mm_fcmul_sch(A, B)
3339 #define _mm_mask_cmul_sch(W, U, A, B) _mm_mask_fcmul_sch(W, U, A, B)
3340 #define _mm_maskz_cmul_sch(U, A, B) _mm_maskz_fcmul_sch(U, A, B)
3341 #define _mm_cmul_round_sch(A, B, R) _mm_fcmul_round_sch(A, B, R)
3342 #define _mm_mask_cmul_round_sch(W, U, A, B, R) \
3343 _mm_mask_fcmul_round_sch(W, U, A, B, R)
3344 #define _mm_maskz_cmul_round_sch(U, A, B, R) \
3345 _mm_maskz_fcmul_round_sch(U, A, B, R)