Lines Matching defs:__b

84 /// \param __b
90 __m128d __b) {
91 __a[0] += __b[0];
103 /// \param __b
108 __m128d __b) {
109 return (__m128d)((__v2df)__a + (__v2df)__b);
124 /// \param __b
130 __m128d __b) {
131 __a[0] -= __b[0];
143 /// \param __b
148 __m128d __b) {
149 return (__m128d)((__v2df)__a - (__v2df)__b);
163 /// \param __b
169 __m128d __b) {
170 __a[0] *= __b[0];
182 /// \param __b
187 __m128d __b) {
188 return (__m128d)((__v2df)__a * (__v2df)__b);
203 /// \param __b
209 __m128d __b) {
210 __a[0] /= __b[0];
223 /// \param __b
228 __m128d __b) {
229 return (__m128d)((__v2df)__a / (__v2df)__b);
245 /// \param __b
249 /// square root of the lower 64 bits of operand \a __b, and whose upper 64
252 __m128d __b) {
253 __m128d __c = __builtin_ia32_sqrtsd((__v2df)__b);
277 /// If either value in a comparison is NaN, returns the value from \a __b.
286 /// \param __b
293 __m128d __b) {
294 return __builtin_ia32_minsd((__v2df)__a, (__v2df)__b);
301 /// If either value in a comparison is NaN, returns the value from \a __b.
309 /// \param __b
314 __m128d __b) {
315 return __builtin_ia32_minpd((__v2df)__a, (__v2df)__b);
323 /// If either value in a comparison is NaN, returns the value from \a __b.
332 /// \param __b
339 __m128d __b) {
340 return __builtin_ia32_maxsd((__v2df)__a, (__v2df)__b);
347 /// If either value in a comparison is NaN, returns the value from \a __b.
355 /// \param __b
360 __m128d __b) {
361 return __builtin_ia32_maxpd((__v2df)__a, (__v2df)__b);
372 /// \param __b
377 __m128d __b) {
378 return (__m128d)((__v2du)__a & (__v2du)__b);
391 /// \param __b
397 _mm_andnot_pd(__m128d __a, __m128d __b) {
398 return (__m128d)(~(__v2du)__a & (__v2du)__b);
409 /// \param __b
414 __m128d __b) {
415 return (__m128d)((__v2du)__a | (__v2du)__b);
426 /// \param __b
431 __m128d __b) {
432 return (__m128d)((__v2du)__a ^ (__v2du)__b);
447 /// \param __b
451 __m128d __b) {
452 return (__m128d)__builtin_ia32_cmpeqpd((__v2df)__a, (__v2df)__b);
468 /// \param __b
472 __m128d __b) {
473 return (__m128d)__builtin_ia32_cmpltpd((__v2df)__a, (__v2df)__b);
489 /// \param __b
493 __m128d __b) {
494 return (__m128d)__builtin_ia32_cmplepd((__v2df)__a, (__v2df)__b);
510 /// \param __b
514 __m128d __b) {
515 return (__m128d)__builtin_ia32_cmpltpd((__v2df)__b, (__v2df)__a);
531 /// \param __b
535 __m128d __b) {
536 return (__m128d)__builtin_ia32_cmplepd((__v2df)__b, (__v2df)__a);
553 /// \param __b
557 __m128d __b) {
558 return (__m128d)__builtin_ia32_cmpordpd((__v2df)__a, (__v2df)__b);
576 /// \param __b
580 __m128d __b) {
581 return (__m128d)__builtin_ia32_cmpunordpd((__v2df)__a, (__v2df)__b);
597 /// \param __b
601 __m128d __b) {
602 return (__m128d)__builtin_ia32_cmpneqpd((__v2df)__a, (__v2df)__b);
618 /// \param __b
622 __m128d __b) {
623 return (__m128d)__builtin_ia32_cmpnltpd((__v2df)__a, (__v2df)__b);
639 /// \param __b
643 __m128d __b) {
644 return (__m128d)__builtin_ia32_cmpnlepd((__v2df)__a, (__v2df)__b);
660 /// \param __b
664 __m128d __b) {
665 return (__m128d)__builtin_ia32_cmpnltpd((__v2df)__b, (__v2df)__a);
681 /// \param __b
685 __m128d __b) {
686 return (__m128d)__builtin_ia32_cmpnlepd((__v2df)__b, (__v2df)__a);
701 /// compared to the lower double-precision value of \a __b.
702 /// \param __b
708 __m128d __b) {
709 return (__m128d)__builtin_ia32_cmpeqsd((__v2df)__a, (__v2df)__b);
726 /// compared to the lower double-precision value of \a __b.
727 /// \param __b
733 __m128d __b) {
734 return (__m128d)__builtin_ia32_cmpltsd((__v2df)__a, (__v2df)__b);
751 /// compared to the lower double-precision value of \a __b.
752 /// \param __b
758 __m128d __b) {
759 return (__m128d)__builtin_ia32_cmplesd((__v2df)__a, (__v2df)__b);
776 /// compared to the lower double-precision value of \a __b.
777 /// \param __b
783 __m128d __b) {
784 __m128d __c = __builtin_ia32_cmpltsd((__v2df)__b, (__v2df)__a);
802 /// compared to the lower double-precision value of \a __b.
803 /// \param __b
809 __m128d __b) {
810 __m128d __c = __builtin_ia32_cmplesd((__v2df)__b, (__v2df)__a);
829 /// compared to the lower double-precision value of \a __b.
830 /// \param __b
836 __m128d __b) {
837 return (__m128d)__builtin_ia32_cmpordsd((__v2df)__a, (__v2df)__b);
856 /// compared to the lower double-precision value of \a __b.
857 /// \param __b
863 __m128d __b) {
864 return (__m128d)__builtin_ia32_cmpunordsd((__v2df)__a, (__v2df)__b);
881 /// compared to the lower double-precision value of \a __b.
882 /// \param __b
888 __m128d __b) {
889 return (__m128d)__builtin_ia32_cmpneqsd((__v2df)__a, (__v2df)__b);
906 /// compared to the lower double-precision value of \a __b.
907 /// \param __b
913 __m128d __b) {
914 return (__m128d)__builtin_ia32_cmpnltsd((__v2df)__a, (__v2df)__b);
931 /// compared to the lower double-precision value of \a __b.
932 /// \param __b
938 __m128d __b) {
939 return (__m128d)__builtin_ia32_cmpnlesd((__v2df)__a, (__v2df)__b);
956 /// compared to the lower double-precision value of \a __b.
957 /// \param __b
963 __m128d __b) {
964 __m128d __c = __builtin_ia32_cmpnltsd((__v2df)__b, (__v2df)__a);
982 /// compared to the lower double-precision value of \a __b.
983 /// \param __b
989 __m128d __b) {
990 __m128d __c = __builtin_ia32_cmpnlesd((__v2df)__b, (__v2df)__a);
1006 /// compared to the lower double-precision value of \a __b.
1007 /// \param __b
1012 __m128d __b) {
1013 return __builtin_ia32_comisdeq((__v2df)__a, (__v2df)__b);
1030 /// compared to the lower double-precision value of \a __b.
1031 /// \param __b
1036 __m128d __b) {
1037 return __builtin_ia32_comisdlt((__v2df)__a, (__v2df)__b);
1054 /// compared to the lower double-precision value of \a __b.
1055 /// \param __b
1060 __m128d __b) {
1061 return __builtin_ia32_comisdle((__v2df)__a, (__v2df)__b);
1078 /// compared to the lower double-precision value of \a __b.
1079 /// \param __b
1084 __m128d __b) {
1085 return __builtin_ia32_comisdgt((__v2df)__a, (__v2df)__b);
1102 /// compared to the lower double-precision value of \a __b.
1103 /// \param __b
1108 __m128d __b) {
1109 return __builtin_ia32_comisdge((__v2df)__a, (__v2df)__b);
1126 /// compared to the lower double-precision value of \a __b.
1127 /// \param __b
1132 __m128d __b) {
1133 return __builtin_ia32_comisdneq((__v2df)__a, (__v2df)__b);
1148 /// compared to the lower double-precision value of \a __b.
1149 /// \param __b
1154 __m128d __b) {
1155 return __builtin_ia32_ucomisdeq((__v2df)__a, (__v2df)__b);
1172 /// compared to the lower double-precision value of \a __b.
1173 /// \param __b
1178 __m128d __b) {
1179 return __builtin_ia32_ucomisdlt((__v2df)__a, (__v2df)__b);
1196 /// compared to the lower double-precision value of \a __b.
1197 /// \param __b
1202 __m128d __b) {
1203 return __builtin_ia32_ucomisdle((__v2df)__a, (__v2df)__b);
1220 /// compared to the lower double-precision value of \a __b.
1221 /// \param __b
1226 __m128d __b) {
1227 return __builtin_ia32_ucomisdgt((__v2df)__a, (__v2df)__b);
1244 /// compared to the lower double-precision value of \a __b.
1245 /// \param __b
1250 __m128d __b) {
1251 return __builtin_ia32_ucomisdge((__v2df)__a, (__v2df)__b);
1268 /// compared to the lower double-precision value of \a __b.
1269 /// \param __b
1274 __m128d __b) {
1275 return __builtin_ia32_ucomisdneq((__v2df)__a, (__v2df)__b);
1390 /// \param __b
1397 __m128d __b) {
1398 return (__m128)__builtin_ia32_cvtsd2ss((__v4sf)__a, (__v2df)__b);
1413 /// \param __b
1419 _mm_cvtsi32_sd(__m128d __a, int __b) {
1420 __a[0] = __b;
1437 /// \param __b
1444 _mm_cvtss_sd(__m128d __a, __m128 __b) {
1445 __a[0] = __b[0];
1891 /// \param __b
1896 _mm_move_sd(__m128d __a, __m128d __b) {
1897 __a[0] = __b[0];
2067 /// \param __b
2072 __m128i __b) {
2073 return (__m128i)((__v16qu)__a + (__v16qu)__b);
2088 /// \param __b
2093 __m128i __b) {
2094 return (__m128i)((__v8hu)__a + (__v8hu)__b);
2109 /// \param __b
2114 _mm_add_epi32(__m128i __a, __m128i __b) {
2115 return (__m128i)((__v4su)__a + (__v4su)__b);
2127 /// \param __b
2130 static __inline__ __m64 __DEFAULT_FN_ATTRS _mm_add_si64(__m64 __a, __m64 __b) {
2131 return (__m64)(((unsigned long long)__a) + ((unsigned long long)__b));
2146 /// \param __b
2151 _mm_add_epi64(__m128i __a, __m128i __b) {
2152 return (__m128i)((__v2du)__a + (__v2du)__b);
2168 /// \param __b
2173 __m128i __b) {
2174 return (__m128i)__builtin_elementwise_add_sat((__v16qs)__a, (__v16qs)__b);
2190 /// \param __b
2195 __m128i __b) {
2196 return (__m128i)__builtin_elementwise_add_sat((__v8hi)__a, (__v8hi)__b);
2212 /// \param __b
2217 __m128i __b) {
2218 return (__m128i)__builtin_elementwise_add_sat((__v16qu)__a, (__v16qu)__b);
2234 /// \param __b
2239 __m128i __b) {
2240 return (__m128i)__builtin_elementwise_add_sat((__v8hu)__a, (__v8hu)__b);
2253 /// \param __b
2258 __m128i __b) {
2259 return (__m128i)__builtin_ia32_pavgb128((__v16qi)__a, (__v16qi)__b);
2272 /// \param __b
2277 __m128i __b) {
2278 return (__m128i)__builtin_ia32_pavgw128((__v8hi)__a, (__v8hi)__b);
2297 /// \param __b
2302 __m128i __b) {
2303 return (__m128i)__builtin_ia32_pmaddwd128((__v8hi)__a, (__v8hi)__b);
2316 /// \param __b
2321 __m128i __b) {
2322 return (__m128i)__builtin_elementwise_max((__v8hi)__a, (__v8hi)__b);
2335 /// \param __b
2340 __m128i __b) {
2341 return (__m128i)__builtin_elementwise_max((__v16qu)__a, (__v16qu)__b);
2354 /// \param __b
2359 __m128i __b) {
2360 return (__m128i)__builtin_elementwise_min((__v8hi)__a, (__v8hi)__b);
2373 /// \param __b
2378 __m128i __b) {
2379 return (__m128i)__builtin_elementwise_min((__v16qu)__a, (__v16qu)__b);
2392 /// \param __b
2397 __m128i __b) {
2398 return (__m128i)__builtin_ia32_pmulhw128((__v8hi)__a, (__v8hi)__b);
2411 /// \param __b
2416 __m128i __b) {
2417 return (__m128i)__builtin_ia32_pmulhuw128((__v8hi)__a, (__v8hi)__b);
2430 /// \param __b
2435 __m128i __b) {
2436 return (__m128i)((__v8hu)__a * (__v8hu)__b);
2449 /// \param __b
2452 static __inline__ __m64 __DEFAULT_FN_ATTRS _mm_mul_su32(__m64 __a, __m64 __b) {
2454 (__v4si)__anyext128(__b)));
2467 /// \param __b
2471 __m128i __b) {
2472 return __builtin_ia32_pmuludq128((__v4si)__a, (__v4si)__b);
2487 /// \param __b
2492 __m128i __b) {
2493 return __builtin_ia32_psadbw128((__v16qi)__a, (__v16qi)__b);
2504 /// \param __b
2509 __m128i __b) {
2510 return (__m128i)((__v16qu)__a - (__v16qu)__b);
2521 /// \param __b
2526 __m128i __b) {
2527 return (__m128i)((__v8hu)__a - (__v8hu)__b);
2538 /// \param __b
2543 _mm_sub_epi32(__m128i __a, __m128i __b) {
2544 return (__m128i)((__v4su)__a - (__v4su)__b);
2556 /// \param __b
2560 static __inline__ __m64 __DEFAULT_FN_ATTRS _mm_sub_si64(__m64 __a, __m64 __b) {
2561 return (__m64)((unsigned long long)__a - (unsigned long long)__b);
2572 /// \param __b
2577 _mm_sub_epi64(__m128i __a, __m128i __b) {
2578 return (__m128i)((__v2du)__a - (__v2du)__b);
2594 /// \param __b
2599 __m128i __b) {
2600 return (__m128i)__builtin_elementwise_sub_sat((__v16qs)__a, (__v16qs)__b);
2616 /// \param __b
2621 __m128i __b) {
2622 return (__m128i)__builtin_elementwise_sub_sat((__v8hi)__a, (__v8hi)__b);
2637 /// \param __b
2642 __m128i __b) {
2643 return (__m128i)__builtin_elementwise_sub_sat((__v16qu)__a, (__v16qu)__b);
2658 /// \param __b
2663 __m128i __b) {
2664 return (__m128i)__builtin_elementwise_sub_sat((__v8hu)__a, (__v8hu)__b);
2675 /// \param __b
2680 __m128i __b) {
2681 return (__m128i)((__v2du)__a & (__v2du)__b);
2694 /// \param __b
2699 __m128i __b) {
2700 return (__m128i)(~(__v2du)__a & (__v2du)__b);
2710 /// \param __b
2715 __m128i __b) {
2716 return (__m128i)((__v2du)__a | (__v2du)__b);
2727 /// \param __b
2732 __m128i __b) {
2733 return (__m128i)((__v2du)__a ^ (__v2du)__b);
3089 /// \param __b
3093 __m128i __b) {
3094 return (__m128i)((__v16qi)__a == (__v16qi)__b);
3108 /// \param __b
3112 __m128i __b) {
3113 return (__m128i)((__v8hi)__a == (__v8hi)__b);
3127 /// \param __b
3131 __m128i __b) {
3132 return (__m128i)((__v4si)__a == (__v4si)__b);
3147 /// \param __b
3151 __m128i __b) {
3154 return (__m128i)((__v16qs)__a > (__v16qs)__b);
3169 /// \param __b
3173 __m128i __b) {
3174 return (__m128i)((__v8hi)__a > (__v8hi)__b);
3189 /// \param __b
3193 __m128i __b) {
3194 return (__m128i)((__v4si)__a > (__v4si)__b);
3209 /// \param __b
3213 __m128i __b) {
3214 return _mm_cmpgt_epi8(__b, __a);
3229 /// \param __b
3233 __m128i __b) {
3234 return _mm_cmpgt_epi16(__b, __a);
3249 /// \param __b
3253 __m128i __b) {
3254 return _mm_cmpgt_epi32(__b, __a);
3270 /// \param __b
3276 _mm_cvtsi64_sd(__m128d __a, long long __b) {
3277 __a[0] = __b;
3413 __v4si __b = (__v4si)__a;
3414 return __b[0];
3741 /// \param __b
3746 static __inline__ __m128i __DEFAULT_FN_ATTRS_CONSTEXPR _mm_set1_epi8(char __b) {
3747 return _mm_set_epi8(__b, __b, __b, __b, __b, __b, __b, __b, __b, __b, __b,
3748 __b, __b, __b, __b, __b);
3894 /// \param __b
3897 __m128i __b) {
3898 *__p = __b;
3909 /// \param __b
3912 __m128i __b) {
3916 ((struct __storeu_si128 *)__p)->__v = __b;
3929 /// \param __b
3932 __m128i __b) {
3936 ((struct __storeu_si64 *)__p)->__v = ((__v2di)__b)[0];
3949 /// \param __b
3952 __m128i __b) {
3956 ((struct __storeu_si32 *)__p)->__v = ((__v4si)__b)[0];
3969 /// \param __b
3972 __m128i __b) {
3976 ((struct __storeu_si16 *)__p)->__v = ((__v8hi)__b)[0];
4160 /// \param __b
4165 __m128i __b) {
4166 return (__m128i)__builtin_ia32_packsswb128((__v8hi)__a, (__v8hi)__b);
4183 /// \param __b
4188 __m128i __b) {
4189 return (__m128i)__builtin_ia32_packssdw128((__v4si)__a, (__v4si)__b);
4206 /// \param __b
4211 __m128i __b) {
4212 return (__m128i)__builtin_ia32_packuswb128((__v8hi)__a, (__v8hi)__b);
4407 /// \param __b
4419 __m128i __b) {
4421 (__v16qi)__a, (__v16qi)__b, 8, 16 + 8, 9, 16 + 9, 10, 16 + 10, 11,
4439 /// \param __b
4447 __m128i __b) {
4448 return (__m128i)__builtin_shufflevector((__v8hi)__a, (__v8hi)__b, 4, 8 + 4, 5,
4464 /// \param __b
4470 __m128i __b) {
4471 return (__m128i)__builtin_shufflevector((__v4si)__a, (__v4si)__b, 2, 4 + 2, 3,
4486 /// \param __b
4491 __m128i __b) {
4492 return (__m128i)__builtin_shufflevector((__v2di)__a, (__v2di)__b, 1, 2 + 1);
4513 /// \param __b
4525 __m128i __b) {
4527 (__v16qi)__a, (__v16qi)__b, 0, 16 + 0, 1, 16 + 1, 2, 16 + 2, 3, 16 + 3, 4,
4546 /// \param __b
4554 __m128i __b) {
4555 return (__m128i)__builtin_shufflevector((__v8hi)__a, (__v8hi)__b, 0, 8 + 0, 1,
4571 /// \param __b
4577 __m128i __b) {
4578 return (__m128i)__builtin_shufflevector((__v4si)__a, (__v4si)__b, 0, 4 + 0, 1,
4593 /// \param __b
4598 __m128i __b) {
4599 return (__m128i)__builtin_shufflevector((__v2di)__a, (__v2di)__b, 0, 2 + 0);
4662 /// \param __b
4667 _mm_unpackhi_pd(__m128d __a, __m128d __b) {
4668 return __builtin_shufflevector((__v2df)__a, (__v2df)__b, 1, 2 + 1);
4682 /// \param __b
4687 _mm_unpacklo_pd(__m128d __a, __m128d __b) {
4688 return __builtin_shufflevector((__v2df)__a, (__v2df)__b, 0, 2 + 0);