Lines Matching defs:R

279 #define _mm_comi_round_sh(A, B, P, R)                                          \
280 __builtin_ia32_vcomish((__v8hf)A, (__v8hf)B, (int)(P), (int)(R))
375 #define _mm512_add_round_ph(A, B, R) \
377 (__v32hf)(__m512h)(B), (int)(R)))
379 #define _mm512_mask_add_round_ph(W, U, A, B, R) \
381 (__mmask32)(U), (__v32hf)_mm512_add_round_ph((A), (B), (R)), \
384 #define _mm512_maskz_add_round_ph(U, A, B, R) \
386 (__mmask32)(U), (__v32hf)_mm512_add_round_ph((A), (B), (R)), \
407 #define _mm512_sub_round_ph(A, B, R) \
409 (__v32hf)(__m512h)(B), (int)(R)))
411 #define _mm512_mask_sub_round_ph(W, U, A, B, R) \
413 (__mmask32)(U), (__v32hf)_mm512_sub_round_ph((A), (B), (R)), \
416 #define _mm512_maskz_sub_round_ph(U, A, B, R) \
418 (__mmask32)(U), (__v32hf)_mm512_sub_round_ph((A), (B), (R)), \
439 #define _mm512_mul_round_ph(A, B, R) \
441 (__v32hf)(__m512h)(B), (int)(R)))
443 #define _mm512_mask_mul_round_ph(W, U, A, B, R) \
445 (__mmask32)(U), (__v32hf)_mm512_mul_round_ph((A), (B), (R)), \
448 #define _mm512_maskz_mul_round_ph(U, A, B, R) \
450 (__mmask32)(U), (__v32hf)_mm512_mul_round_ph((A), (B), (R)), \
471 #define _mm512_div_round_ph(A, B, R) \
473 (__v32hf)(__m512h)(B), (int)(R)))
475 #define _mm512_mask_div_round_ph(W, U, A, B, R) \
477 (__mmask32)(U), (__v32hf)_mm512_div_round_ph((A), (B), (R)), \
480 #define _mm512_maskz_div_round_ph(U, A, B, R) \
482 (__mmask32)(U), (__v32hf)_mm512_div_round_ph((A), (B), (R)), \
504 #define _mm512_min_round_ph(A, B, R) \
506 (__v32hf)(__m512h)(B), (int)(R)))
508 #define _mm512_mask_min_round_ph(W, U, A, B, R) \
510 (__mmask32)(U), (__v32hf)_mm512_min_round_ph((A), (B), (R)), \
513 #define _mm512_maskz_min_round_ph(U, A, B, R) \
515 (__mmask32)(U), (__v32hf)_mm512_min_round_ph((A), (B), (R)), \
537 #define _mm512_max_round_ph(A, B, R) \
539 (__v32hf)(__m512h)(B), (int)(R)))
541 #define _mm512_mask_max_round_ph(W, U, A, B, R) \
543 (__mmask32)(U), (__v32hf)_mm512_max_round_ph((A), (B), (R)), \
546 #define _mm512_maskz_max_round_ph(U, A, B, R) \
548 (__mmask32)(U), (__v32hf)_mm512_max_round_ph((A), (B), (R)), \
593 #define _mm_add_round_sh(A, B, R) \
596 (__mmask8)-1, (int)(R)))
598 #define _mm_mask_add_round_sh(W, U, A, B, R) \
601 (__mmask8)(U), (int)(R)))
603 #define _mm_maskz_add_round_sh(U, A, B, R) \
606 (__mmask8)(U), (int)(R)))
629 #define _mm_sub_round_sh(A, B, R) \
632 (__mmask8)-1, (int)(R)))
634 #define _mm_mask_sub_round_sh(W, U, A, B, R) \
637 (__mmask8)(U), (int)(R)))
639 #define _mm_maskz_sub_round_sh(U, A, B, R) \
642 (__mmask8)(U), (int)(R)))
665 #define _mm_mul_round_sh(A, B, R) \
668 (__mmask8)-1, (int)(R)))
670 #define _mm_mask_mul_round_sh(W, U, A, B, R) \
673 (__mmask8)(U), (int)(R)))
675 #define _mm_maskz_mul_round_sh(U, A, B, R) \
678 (__mmask8)(U), (int)(R)))
701 #define _mm_div_round_sh(A, B, R) \
704 (__mmask8)-1, (int)(R)))
706 #define _mm_mask_div_round_sh(W, U, A, B, R) \
709 (__mmask8)(U), (int)(R)))
711 #define _mm_maskz_div_round_sh(U, A, B, R) \
714 (__mmask8)(U), (int)(R)))
740 #define _mm_min_round_sh(A, B, R) \
743 (__mmask8)-1, (int)(R)))
745 #define _mm_mask_min_round_sh(W, U, A, B, R) \
748 (__mmask8)(U), (int)(R)))
750 #define _mm_maskz_min_round_sh(U, A, B, R) \
753 (__mmask8)(U), (int)(R)))
779 #define _mm_max_round_sh(A, B, R) \
782 (__mmask8)-1, (int)(R)))
784 #define _mm_mask_max_round_sh(W, U, A, B, R) \
787 (__mmask8)(U), (int)(R)))
789 #define _mm_maskz_max_round_sh(U, A, B, R) \
792 (__mmask8)(U), (int)(R)))
794 #define _mm512_cmp_round_ph_mask(A, B, P, R) \
797 (__mmask32)-1, (int)(R)))
799 #define _mm512_mask_cmp_round_ph_mask(U, A, B, P, R) \
802 (__mmask32)(U), (int)(R)))
810 #define _mm_cmp_round_sh_mask(X, Y, P, R) \
813 (__mmask8)-1, (int)(R)))
815 #define _mm_mask_cmp_round_sh_mask(M, X, Y, P, R) \
818 (__mmask8)(M), (int)(R)))
1024 #define _mm512_getmant_round_ph(A, B, C, R) \
1027 (__v32hf)_mm512_undefined_ph(), (__mmask32)-1, (int)(R)))
1029 #define _mm512_mask_getmant_round_ph(W, U, A, B, C, R) \
1032 (__mmask32)(U), (int)(R)))
1034 #define _mm512_maskz_getmant_round_ph(U, A, B, C, R) \
1037 (__v32hf)_mm512_setzero_ph(), (__mmask32)(U), (int)(R)))
1058 #define _mm512_getexp_round_ph(A, R) \
1061 (__mmask32)-1, (int)(R)))
1063 #define _mm512_mask_getexp_round_ph(W, U, A, R) \
1065 (__v32hf)(__m512h)(A), (__v32hf)(__m512h)(W), (__mmask32)(U), (int)(R)))
1067 #define _mm512_maskz_getexp_round_ph(U, A, R) \
1070 (__mmask32)(U), (int)(R)))
1093 #define _mm512_scalef_round_ph(A, B, R) \
1096 (__v32hf)_mm512_undefined_ph(), (__mmask32)-1, (int)(R)))
1098 #define _mm512_mask_scalef_round_ph(W, U, A, B, R) \
1101 (__mmask32)(U), (int)(R)))
1103 #define _mm512_maskz_scalef_round_ph(U, A, B, R) \
1106 (__v32hf)_mm512_setzero_ph(), (__mmask32)(U), (int)(R)))
1123 #define _mm512_mask_roundscale_round_ph(A, B, C, imm, R) \
1126 (__mmask32)(B), (int)(R)))
1128 #define _mm512_maskz_roundscale_round_ph(A, B, imm, R) \
1131 (__mmask32)(A), (int)(R)))
1133 #define _mm512_roundscale_round_ph(A, imm, R) \
1136 (__mmask32)-1, (int)(R)))
1153 #define _mm512_mask_reduce_round_ph(W, U, A, imm, R) \
1156 (__mmask32)(U), (int)(R)))
1158 #define _mm512_maskz_reduce_round_ph(U, A, imm, R) \
1161 (__mmask32)(U), (int)(R)))
1163 #define _mm512_reduce_round_ph(A, imm, R) \
1166 (__mmask32)-1, (int)(R)))
1209 #define _mm_getmant_round_sh(A, B, C, D, R) \
1212 (__v8hf)_mm_setzero_ph(), (__mmask8)-1, (int)(R)))
1224 #define _mm_mask_getmant_round_sh(W, U, A, B, C, D, R) \
1227 (__v8hf)(__m128h)(W), (__mmask8)(U), (int)(R)))
1234 #define _mm_maskz_getmant_round_sh(U, A, B, C, D, R) \
1237 (__v8hf)_mm_setzero_ph(), (__mmask8)(U), (int)(R)))
1239 #define _mm_getexp_round_sh(A, B, R) \
1242 (__mmask8)-1, (int)(R)))
1258 #define _mm_mask_getexp_round_sh(W, U, A, B, R) \
1261 (__mmask8)(U), (int)(R)))
1270 #define _mm_maskz_getexp_round_sh(U, A, B, R) \
1273 (__mmask8)(U), (int)(R)))
1275 #define _mm_scalef_round_sh(A, B, R) \
1278 (__mmask8)-1, (int)(R)))
1294 #define _mm_mask_scalef_round_sh(W, U, A, B, R) \
1297 (__mmask8)(U), (int)(R)))
1306 #define _mm_maskz_scalef_round_sh(U, A, B, R) \
1309 (__mmask8)(U), (int)(R)))
1311 #define _mm_roundscale_round_sh(A, B, imm, R) \
1314 (__mmask8)-1, (int)(imm), (int)(R)))
1326 #define _mm_mask_roundscale_round_sh(W, U, A, B, I, R) \
1329 (__mmask8)(U), (int)(I), (int)(R)))
1336 #define _mm_maskz_roundscale_round_sh(U, A, B, I, R) \
1339 (__mmask8)(U), (int)(I), (int)(R)))
1356 #define _mm_reduce_round_sh(A, B, C, R) \
1359 (__mmask8)-1, (int)(C), (int)(R)))
1361 #define _mm_mask_reduce_round_sh(W, U, A, B, C, R) \
1364 (__mmask8)(U), (int)(C), (int)(R)))
1366 #define _mm_maskz_reduce_round_sh(U, A, B, C, R) \
1369 (__mmask8)(U), (int)(C), (int)(R)))
1371 #define _mm512_sqrt_round_ph(A, R) \
1372 ((__m512h)__builtin_ia32_sqrtph512((__v32hf)(__m512h)(A), (int)(R)))
1374 #define _mm512_mask_sqrt_round_ph(W, U, A, R) \
1376 (__mmask32)(U), (__v32hf)_mm512_sqrt_round_ph((A), (R)), \
1379 #define _mm512_maskz_sqrt_round_ph(U, A, R) \
1381 (__mmask32)(U), (__v32hf)_mm512_sqrt_round_ph((A), (R)), \
1405 #define _mm_sqrt_round_sh(A, B, R) \
1408 (__mmask8)-1, (int)(R)))
1410 #define _mm_mask_sqrt_round_sh(W, U, A, B, R) \
1413 (__mmask8)(U), (int)(R)))
1415 #define _mm_maskz_sqrt_round_sh(U, A, B, R) \
1418 (__mmask8)(U), (int)(R)))
1460 #define _mm512_cvt_roundpd_ph(A, R) \
1462 (__v8df)(A), (__v8hf)_mm_undefined_ph(), (__mmask8)(-1), (int)(R)))
1464 #define _mm512_mask_cvt_roundpd_ph(W, U, A, R) \
1466 (__mmask8)(U), (int)(R)))
1468 #define _mm512_maskz_cvt_roundpd_ph(U, A, R) \
1470 (__v8df)(A), (__v8hf)_mm_setzero_ph(), (__mmask8)(U), (int)(R)))
1491 #define _mm512_cvt_roundph_pd(A, R) \
1493 (__v8hf)(A), (__v8df)_mm512_undefined_pd(), (__mmask8)(-1), (int)(R)))
1495 #define _mm512_mask_cvt_roundph_pd(W, U, A, R) \
1497 (__mmask8)(U), (int)(R)))
1499 #define _mm512_maskz_cvt_roundph_pd(U, A, R) \
1501 (__v8hf)(A), (__v8df)_mm512_setzero_pd(), (__mmask8)(U), (int)(R)))
1522 #define _mm_cvt_roundsh_ss(A, B, R) \
1525 (__mmask8)(-1), (int)(R)))
1527 #define _mm_mask_cvt_roundsh_ss(W, U, A, B, R) \
1529 (__v4sf)(A), (__v8hf)(B), (__v4sf)(W), (__mmask8)(U), (int)(R)))
1531 #define _mm_maskz_cvt_roundsh_ss(U, A, B, R) \
1534 (__mmask8)(U), (int)(R)))
1560 #define _mm_cvt_roundss_sh(A, B, R) \
1563 (__mmask8)(-1), (int)(R)))
1565 #define _mm_mask_cvt_roundss_sh(W, U, A, B, R) \
1567 (__v8hf)(A), (__v4sf)(B), (__v8hf)(W), (__mmask8)(U), (int)(R)))
1569 #define _mm_maskz_cvt_roundss_sh(U, A, B, R) \
1572 (__mmask8)(U), (int)(R)))
1598 #define _mm_cvt_roundsd_sh(A, B, R) \
1601 (__mmask8)(-1), (int)(R)))
1603 #define _mm_mask_cvt_roundsd_sh(W, U, A, B, R) \
1605 (__v8hf)(A), (__v2df)(B), (__v8hf)(W), (__mmask8)(U), (int)(R)))
1607 #define _mm_maskz_cvt_roundsd_sh(U, A, B, R) \
1610 (__mmask8)(U), (int)(R)))
1635 #define _mm_cvt_roundsh_sd(A, B, R) \
1638 (__mmask8)(-1), (int)(R)))
1640 #define _mm_mask_cvt_roundsh_sd(W, U, A, B, R) \
1642 (__v2df)(A), (__v8hf)(B), (__v2df)(W), (__mmask8)(U), (int)(R)))
1644 #define _mm_maskz_cvt_roundsh_sd(U, A, B, R) \
1647 (__mmask8)(U), (int)(R)))
1672 #define _mm512_cvt_roundph_epi16(A, R) \
1675 (__mmask32)(-1), (int)(R)))
1677 #define _mm512_mask_cvt_roundph_epi16(W, U, A, R) \
1679 (__mmask32)(U), (int)(R)))
1681 #define _mm512_maskz_cvt_roundph_epi16(U, A, R) \
1684 (__mmask32)(U), (int)(R)))
1706 #define _mm512_cvtt_roundph_epi16(A, R) \
1709 (int)(R)))
1711 #define _mm512_mask_cvtt_roundph_epi16(W, U, A, R) \
1713 (__mmask32)(U), (int)(R)))
1715 #define _mm512_maskz_cvtt_roundph_epi16(U, A, R) \
1718 (__mmask32)(U), (int)(R)))
1740 #define _mm512_cvt_roundepi16_ph(A, R) \
1743 (__mmask32)(-1), (int)(R)))
1745 #define _mm512_mask_cvt_roundepi16_ph(W, U, A, R) \
1747 (__mmask32)(U), (int)(R)))
1749 #define _mm512_maskz_cvt_roundepi16_ph(U, A, R) \
1751 (__v32hi)(A), (__v32hf)_mm512_setzero_ph(), (__mmask32)(U), (int)(R)))
1773 #define _mm512_cvt_roundph_epu16(A, R) \
1776 (int)(R)))
1778 #define _mm512_mask_cvt_roundph_epu16(W, U, A, R) \
1780 (__mmask32)(U), (int)(R)))
1782 #define _mm512_maskz_cvt_roundph_epu16(U, A, R) \
1785 (__mmask32)(U), (int)(R)))
1807 #define _mm512_cvtt_roundph_epu16(A, R) \
1810 (int)(R)))
1812 #define _mm512_mask_cvtt_roundph_epu16(W, U, A, R) \
1814 (__mmask32)(U), (int)(R)))
1816 #define _mm512_maskz_cvtt_roundph_epu16(U, A, R) \
1819 (__mmask32)(U), (int)(R)))
1841 #define _mm512_cvt_roundepu16_ph(A, R) \
1844 (__mmask32)(-1), (int)(R)))
1846 #define _mm512_mask_cvt_roundepu16_ph(W, U, A, R) \
1848 (__mmask32)(U), (int)(R)))
1850 #define _mm512_maskz_cvt_roundepu16_ph(U, A, R) \
1852 (__v32hu)(A), (__v32hf)_mm512_setzero_ph(), (__mmask32)(U), (int)(R)))
1874 #define _mm512_cvt_roundph_epi32(A, R) \
1877 (int)(R)))
1879 #define _mm512_mask_cvt_roundph_epi32(W, U, A, R) \
1881 (__mmask16)(U), (int)(R)))
1883 #define _mm512_maskz_cvt_roundph_epi32(U, A, R) \
1886 (__mmask16)(U), (int)(R)))
1908 #define _mm512_cvt_roundph_epu32(A, R) \
1911 (int)(R)))
1913 #define _mm512_mask_cvt_roundph_epu32(W, U, A, R) \
1915 (__mmask16)(U), (int)(R)))
1917 #define _mm512_maskz_cvt_roundph_epu32(U, A, R) \
1920 (__mmask16)(U), (int)(R)))
1942 #define _mm512_cvt_roundepi32_ph(A, R) \
1945 (__mmask16)(-1), (int)(R)))
1947 #define _mm512_mask_cvt_roundepi32_ph(W, U, A, R) \
1949 (__mmask16)(U), (int)(R)))
1951 #define _mm512_maskz_cvt_roundepi32_ph(U, A, R) \
1953 (__v16si)(A), (__v16hf)_mm256_setzero_ph(), (__mmask16)(U), (int)(R)))
1975 #define _mm512_cvt_roundepu32_ph(A, R) \
1978 (__mmask16)(-1), (int)(R)))
1980 #define _mm512_mask_cvt_roundepu32_ph(W, U, A, R) \
1982 (__mmask16)(U), (int)(R)))
1984 #define _mm512_maskz_cvt_roundepu32_ph(U, A, R) \
1986 (__v16su)(A), (__v16hf)_mm256_setzero_ph(), (__mmask16)(U), (int)(R)))
2008 #define _mm512_cvtt_roundph_epi32(A, R) \
2011 (int)(R)))
2013 #define _mm512_mask_cvtt_roundph_epi32(W, U, A, R) \
2015 (__mmask16)(U), (int)(R)))
2017 #define _mm512_maskz_cvtt_roundph_epi32(U, A, R) \
2020 (__mmask16)(U), (int)(R)))
2042 #define _mm512_cvtt_roundph_epu32(A, R) \
2045 (int)(R)))
2047 #define _mm512_mask_cvtt_roundph_epu32(W, U, A, R) \
2049 (__mmask16)(U), (int)(R)))
2051 #define _mm512_maskz_cvtt_roundph_epu32(U, A, R) \
2054 (int)(R)))
2076 #define _mm512_cvt_roundepi64_ph(A, R) \
2078 (__v8di)(A), (__v8hf)_mm_undefined_ph(), (__mmask8)(-1), (int)(R)))
2080 #define _mm512_mask_cvt_roundepi64_ph(W, U, A, R) \
2082 (__mmask8)(U), (int)(R)))
2084 #define _mm512_maskz_cvt_roundepi64_ph(U, A, R) \
2086 (__v8di)(A), (__v8hf)_mm_setzero_ph(), (__mmask8)(U), (int)(R)))
2108 #define _mm512_cvt_roundph_epi64(A, R) \
2111 (__mmask8)(-1), (int)(R)))
2113 #define _mm512_mask_cvt_roundph_epi64(W, U, A, R) \
2115 (__mmask8)(U), (int)(R)))
2117 #define _mm512_maskz_cvt_roundph_epi64(U, A, R) \
2119 (__v8hf)(A), (__v8di)_mm512_setzero_epi32(), (__mmask8)(U), (int)(R)))
2141 #define _mm512_cvt_roundepu64_ph(A, R) \
2143 (__v8du)(A), (__v8hf)_mm_undefined_ph(), (__mmask8)(-1), (int)(R)))
2145 #define _mm512_mask_cvt_roundepu64_ph(W, U, A, R) \
2147 (__mmask8)(U), (int)(R)))
2149 #define _mm512_maskz_cvt_roundepu64_ph(U, A, R) \
2151 (__v8du)(A), (__v8hf)_mm_setzero_ph(), (__mmask8)(U), (int)(R)))
2173 #define _mm512_cvt_roundph_epu64(A, R) \
2176 (int)(R)))
2178 #define _mm512_mask_cvt_roundph_epu64(W, U, A, R) \
2180 (__mmask8)(U), (int)(R)))
2182 #define _mm512_maskz_cvt_roundph_epu64(U, A, R) \
2184 (__v8hf)(A), (__v8du)_mm512_setzero_epi32(), (__mmask8)(U), (int)(R)))
2206 #define _mm512_cvtt_roundph_epi64(A, R) \
2209 (int)(R)))
2211 #define _mm512_mask_cvtt_roundph_epi64(W, U, A, R) \
2213 (__mmask8)(U), (int)(R)))
2215 #define _mm512_maskz_cvtt_roundph_epi64(U, A, R) \
2217 (__v8hf)(A), (__v8di)_mm512_setzero_epi32(), (__mmask8)(U), (int)(R)))
2239 #define _mm512_cvtt_roundph_epu64(A, R) \
2242 (int)(R)))
2244 #define _mm512_mask_cvtt_roundph_epu64(W, U, A, R) \
2246 (__mmask8)(U), (int)(R)))
2248 #define _mm512_maskz_cvtt_roundph_epu64(U, A, R) \
2250 (__v8hf)(A), (__v8du)_mm512_setzero_epi32(), (__mmask8)(U), (int)(R)))
2272 #define _mm_cvt_roundsh_i32(A, R) \
2273 ((int)__builtin_ia32_vcvtsh2si32((__v8hf)(A), (int)(R)))
2279 #define _mm_cvt_roundsh_u32(A, R) \
2280 ((unsigned int)__builtin_ia32_vcvtsh2usi32((__v8hf)(A), (int)(R)))
2289 #define _mm_cvt_roundsh_i64(A, R) \
2290 ((long long)__builtin_ia32_vcvtsh2si64((__v8hf)(A), (int)(R)))
2297 #define _mm_cvt_roundsh_u64(A, R) \
2298 ((unsigned long long)__builtin_ia32_vcvtsh2usi64((__v8hf)(A), (int)(R)))
2307 #define _mm_cvt_roundu32_sh(A, B, R) \
2308 ((__m128h)__builtin_ia32_vcvtusi2sh((__v8hf)(A), (unsigned int)(B), (int)(R)))
2317 #define _mm_cvt_roundu64_sh(A, B, R) \
2319 (int)(R)))
2328 #define _mm_cvt_roundi32_sh(A, B, R) \
2329 ((__m128h)__builtin_ia32_vcvtsi2sh((__v8hf)(A), (int)(B), (int)(R)))
2338 #define _mm_cvt_roundi64_sh(A, B, R) \
2339 ((__m128h)__builtin_ia32_vcvtsi642sh((__v8hf)(A), (long long)(B), (int)(R)))
2348 #define _mm_cvtt_roundsh_i32(A, R) \
2349 ((int)__builtin_ia32_vcvttsh2si32((__v8hf)(A), (int)(R)))
2357 #define _mm_cvtt_roundsh_i64(A, R) \
2358 ((long long)__builtin_ia32_vcvttsh2si64((__v8hf)(A), (int)(R)))
2366 #define _mm_cvtt_roundsh_u32(A, R) \
2367 ((unsigned int)__builtin_ia32_vcvttsh2usi32((__v8hf)(A), (int)(R)))
2376 #define _mm_cvtt_roundsh_u64(A, R) \
2377 ((unsigned long long)__builtin_ia32_vcvttsh2usi64((__v8hf)(A), (int)(R)))
2386 #define _mm512_cvtx_roundph_ps(A, R) \
2389 (__mmask16)(-1), (int)(R)))
2391 #define _mm512_mask_cvtx_roundph_ps(W, U, A, R) \
2393 (__mmask16)(U), (int)(R)))
2395 #define _mm512_maskz_cvtx_roundph_ps(U, A, R) \
2397 (__v16hf)(A), (__v16sf)_mm512_setzero_ps(), (__mmask16)(U), (int)(R)))
2418 #define _mm512_cvtx_roundps_ph(A, R) \
2421 (__mmask16)(-1), (int)(R)))
2423 #define _mm512_mask_cvtx_roundps_ph(W, U, A, R) \
2425 (__mmask16)(U), (int)(R)))
2427 #define _mm512_maskz_cvtx_roundps_ph(U, A, R) \
2429 (__v16sf)(A), (__v16hf)_mm256_setzero_ph(), (__mmask16)(U), (int)(R)))
2450 #define _mm512_fmadd_round_ph(A, B, C, R) \
2453 (__mmask32)-1, (int)(R)))
2455 #define _mm512_mask_fmadd_round_ph(A, U, B, C, R) \
2458 (__mmask32)(U), (int)(R)))
2460 #define _mm512_mask3_fmadd_round_ph(A, B, C, U, R) \
2463 (__mmask32)(U), (int)(R)))
2465 #define _mm512_maskz_fmadd_round_ph(U, A, B, C, R) \
2468 (__mmask32)(U), (int)(R)))
2470 #define _mm512_fmsub_round_ph(A, B, C, R) \
2473 (__mmask32)-1, (int)(R)))
2475 #define _mm512_mask_fmsub_round_ph(A, U, B, C, R) \
2478 (__mmask32)(U), (int)(R)))
2480 #define _mm512_maskz_fmsub_round_ph(U, A, B, C, R) \
2483 (__mmask32)(U), (int)(R)))
2485 #define _mm512_fnmadd_round_ph(A, B, C, R) \
2488 (__mmask32)-1, (int)(R)))
2490 #define _mm512_mask3_fnmadd_round_ph(A, B, C, U, R) \
2493 (__mmask32)(U), (int)(R)))
2495 #define _mm512_maskz_fnmadd_round_ph(U, A, B, C, R) \
2498 (__mmask32)(U), (int)(R)))
2500 #define _mm512_fnmsub_round_ph(A, B, C, R) \
2503 (__mmask32)-1, (int)(R)))
2505 #define _mm512_maskz_fnmsub_round_ph(U, A, B, C, R) \
2508 (__mmask32)(U), (int)(R)))
2598 #define _mm512_fmaddsub_round_ph(A, B, C, R) \
2601 (__mmask32)-1, (int)(R)))
2603 #define _mm512_mask_fmaddsub_round_ph(A, U, B, C, R) \
2606 (__mmask32)(U), (int)(R)))
2608 #define _mm512_mask3_fmaddsub_round_ph(A, B, C, U, R) \
2611 (__mmask32)(U), (int)(R)))
2613 #define _mm512_maskz_fmaddsub_round_ph(U, A, B, C, R) \
2616 (__mmask32)(U), (int)(R)))
2618 #define _mm512_fmsubadd_round_ph(A, B, C, R) \
2621 (__mmask32)-1, (int)(R)))
2623 #define _mm512_mask_fmsubadd_round_ph(A, U, B, C, R) \
2626 (__mmask32)(U), (int)(R)))
2628 #define _mm512_maskz_fmsubadd_round_ph(U, A, B, C, R) \
2631 (__mmask32)(U), (int)(R)))
2682 #define _mm512_mask3_fmsub_round_ph(A, B, C, U, R) \
2685 (__mmask32)(U), (int)(R)))
2694 #define _mm512_mask3_fmsubadd_round_ph(A, B, C, U, R) \
2697 (__mmask32)(U), (int)(R)))
2706 #define _mm512_mask_fnmadd_round_ph(A, U, B, C, R) \
2709 (__mmask32)(U), (int)(R)))
2718 #define _mm512_mask_fnmsub_round_ph(A, U, B, C, R) \
2721 (__mmask32)(U), (int)(R)))
2723 #define _mm512_mask3_fnmsub_round_ph(A, B, C, U, R) \
2726 (__mmask32)(U), (int)(R)))
2757 #define _mm_fmadd_round_sh(A, B, C, R) \
2760 (__mmask8)-1, (int)(R)))
2762 #define _mm_mask_fmadd_round_sh(W, U, A, B, R) \
2765 (__mmask8)(U), (int)(R)))
2774 #define _mm_maskz_fmadd_round_sh(U, A, B, C, R) \
2777 (__mmask8)(U), (int)(R)))
2786 #define _mm_mask3_fmadd_round_sh(W, X, Y, U, R) \
2789 (__mmask8)(U), (int)(R)))
2808 #define _mm_fmsub_round_sh(A, B, C, R) \
2811 (__mmask8)-1, (int)(R)))
2813 #define _mm_mask_fmsub_round_sh(W, U, A, B, R) \
2816 (__mmask8)(U), (int)(R)))
2825 #define _mm_maskz_fmsub_round_sh(U, A, B, C, R) \
2828 (__mmask8)(U), (int)R))
2837 #define _mm_mask3_fmsub_round_sh(W, X, Y, U, R) \
2840 (__mmask8)(U), (int)(R)))
2855 #define _mm_fnmadd_round_sh(A, B, C, R) \
2858 (__mmask8)-1, (int)(R)))
2860 #define _mm_mask_fnmadd_round_sh(W, U, A, B, R) \
2863 (__mmask8)(U), (int)(R)))
2872 #define _mm_maskz_fnmadd_round_sh(U, A, B, C, R) \
2875 (__mmask8)(U), (int)(R)))
2884 #define _mm_mask3_fnmadd_round_sh(W, X, Y, U, R) \
2887 (__mmask8)(U), (int)(R)))
2902 #define _mm_fnmsub_round_sh(A, B, C, R) \
2905 (__mmask8)-1, (int)(R)))
2907 #define _mm_mask_fnmsub_round_sh(W, U, A, B, R) \
2910 (__mmask8)(U), (int)(R)))
2919 #define _mm_maskz_fnmsub_round_sh(U, A, B, C, R) \
2922 (__mmask8)(U), (int)(R)))
2931 #define _mm_mask3_fnmsub_round_sh(W, X, Y, U, R) \
2934 (__mmask8)(U), (int)(R)))
2963 #define _mm_fcmadd_round_sch(A, B, C, R) \
2966 (__mmask8)-1, (int)(R)))
2968 #define _mm_mask_fcmadd_round_sch(A, U, B, C, R) \
2971 (__mmask8)(U), (int)(R)))
2973 #define _mm_maskz_fcmadd_round_sch(U, A, B, C, R) \
2976 (__mmask8)(U), (int)(R)))
2978 #define _mm_mask3_fcmadd_round_sch(A, B, C, U, R) \
2981 (__mmask8)(U), (int)(R)))
3010 #define _mm_fmadd_round_sch(A, B, C, R) \
3013 (__mmask8)-1, (int)(R)))
3015 #define _mm_mask_fmadd_round_sch(A, U, B, C, R) \
3018 (__mmask8)(U), (int)(R)))
3020 #define _mm_maskz_fmadd_round_sch(U, A, B, C, R) \
3023 (__mmask8)(U), (int)(R)))
3025 #define _mm_mask3_fmadd_round_sch(A, B, C, U, R) \
3028 (__mmask8)(U), (int)(R)))
3051 #define _mm_fcmul_round_sch(A, B, R) \
3054 (__v4sf)(__m128h)_mm_undefined_ph(), (__mmask8)-1, (int)(R)))
3056 #define _mm_mask_fcmul_round_sch(W, U, A, B, R) \
3059 (__mmask8)(U), (int)(R)))
3061 #define _mm_maskz_fcmul_round_sch(U, A, B, R) \
3064 (__v4sf)(__m128h)_mm_setzero_ph(), (__mmask8)(U), (int)(R)))
3089 #define _mm_fmul_round_sch(A, B, R) \
3092 (__v4sf)(__m128h)_mm_undefined_ph(), (__mmask8)-1, (int)(R)))
3094 #define _mm_mask_fmul_round_sch(W, U, A, B, R) \
3097 (__mmask8)(U), (int)(R)))
3099 #define _mm_maskz_fmul_round_sch(U, A, B, R) \
3102 (__v4sf)(__m128h)_mm_setzero_ph(), (__mmask8)(U), (int)(R)))
3125 #define _mm512_fcmul_round_pch(A, B, R) \
3128 (__v16sf)(__m512h)_mm512_undefined_ph(), (__mmask16)-1, (int)(R)))
3130 #define _mm512_mask_fcmul_round_pch(W, U, A, B, R) \
3133 (__mmask16)(U), (int)(R)))
3135 #define _mm512_maskz_fcmul_round_pch(U, A, B, R) \
3138 (__v16sf)(__m512h)_mm512_setzero_ph(), (__mmask16)(U), (int)(R)))
3161 #define _mm512_fmul_round_pch(A, B, R) \
3164 (__v16sf)(__m512h)_mm512_undefined_ph(), (__mmask16)-1, (int)(R)))
3166 #define _mm512_mask_fmul_round_pch(W, U, A, B, R) \
3169 (__mmask16)(U), (int)(R)))
3171 #define _mm512_maskz_fmul_round_pch(U, A, B, R) \
3174 (__v16sf)(__m512h)_mm512_setzero_ph(), (__mmask16)(U), (int)(R)))
3205 #define _mm512_fcmadd_round_pch(A, B, C, R) \
3208 (__mmask16)-1, (int)(R)))
3210 #define _mm512_mask_fcmadd_round_pch(A, U, B, C, R) \
3213 (__mmask16)(U), (int)(R)))
3215 #define _mm512_mask3_fcmadd_round_pch(A, B, C, U, R) \
3218 (__mmask16)(U), (int)(R)))
3220 #define _mm512_maskz_fcmadd_round_pch(U, A, B, C, R) \
3223 (__mmask16)(U), (int)(R)))
3254 #define _mm512_fmadd_round_pch(A, B, C, R) \
3257 (__mmask16)-1, (int)(R)))
3259 #define _mm512_mask_fmadd_round_pch(A, U, B, C, R) \
3262 (__mmask16)(U), (int)(R)))
3264 #define _mm512_mask3_fmadd_round_pch(A, B, C, U, R) \
3267 (__mmask16)(U), (int)(R)))
3269 #define _mm512_maskz_fmadd_round_pch(U, A, B, C, R) \
3272 (__mmask16)(U), (int)(R)))
3315 #define _mm512_mul_round_pch(A, B, R) _mm512_fmul_round_pch(A, B, R)
3316 #define _mm512_mask_mul_round_pch(W, U, A, B, R) \
3317 _mm512_mask_fmul_round_pch(W, U, A, B, R)
3318 #define _mm512_maskz_mul_round_pch(U, A, B, R) \
3319 _mm512_maskz_fmul_round_pch(U, A, B, R)
3324 #define _mm512_cmul_round_pch(A, B, R) _mm512_fcmul_round_pch(A, B, R)
3325 #define _mm512_mask_cmul_round_pch(W, U, A, B, R) \
3326 _mm512_mask_fcmul_round_pch(W, U, A, B, R)
3327 #define _mm512_maskz_cmul_round_pch(U, A, B, R) \
3328 _mm512_maskz_fcmul_round_pch(U, A, B, R)
3333 #define _mm_mul_round_sch(A, B, R) _mm_fmul_round_sch(A, B, R)
3334 #define _mm_mask_mul_round_sch(W, U, A, B, R) \
3335 _mm_mask_fmul_round_sch(W, U, A, B, R)
3336 #define _mm_maskz_mul_round_sch(U, A, B, R) _mm_maskz_fmul_round_sch(U, A, B, R)
3341 #define _mm_cmul_round_sch(A, B, R) _mm_fcmul_round_sch(A, B, R)
3342 #define _mm_mask_cmul_round_sch(W, U, A, B, R) \
3343 _mm_mask_fcmul_round_sch(W, U, A, B, R)
3344 #define _mm_maskz_cmul_round_sch(U, A, B, R) \
3345 _mm_maskz_fcmul_round_sch(U, A, B, R)