Lines Matching defs:__b

80 /// \param __b
87 _mm_add_ss(__m128 __a, __m128 __b) {
88 __a[0] += __b[0];
101 /// \param __b
106 _mm_add_ps(__m128 __a, __m128 __b) {
107 return (__m128)((__v4sf)__a + (__v4sf)__b);
120 /// \param __b
127 _mm_sub_ss(__m128 __a, __m128 __b) {
128 __a[0] -= __b[0];
142 /// \param __b
147 _mm_sub_ps(__m128 __a, __m128 __b) {
148 return (__m128)((__v4sf)__a - (__v4sf)__b);
161 /// \param __b
168 _mm_mul_ss(__m128 __a, __m128 __b) {
169 __a[0] *= __b[0];
182 /// \param __b
187 _mm_mul_ps(__m128 __a, __m128 __b) {
188 return (__m128)((__v4sf)__a * (__v4sf)__b);
201 /// \param __b
208 _mm_div_ss(__m128 __a, __m128 __b) {
209 __a[0] /= __b[0];
221 /// \param __b
226 _mm_div_ps(__m128 __a, __m128 __b) {
227 return (__m128)((__v4sf)__a / (__v4sf)__b);
340 /// If either value in a comparison is NaN, returns the value from \a __b.
349 /// \param __b
356 _mm_min_ss(__m128 __a, __m128 __b)
358 return __builtin_ia32_minss((__v4sf)__a, (__v4sf)__b);
364 /// If either value in a comparison is NaN, returns the value from \a __b.
372 /// \param __b
377 _mm_min_ps(__m128 __a, __m128 __b)
379 return __builtin_ia32_minps((__v4sf)__a, (__v4sf)__b);
386 /// If either value in a comparison is NaN, returns the value from \a __b.
395 /// \param __b
402 _mm_max_ss(__m128 __a, __m128 __b)
404 return __builtin_ia32_maxss((__v4sf)__a, (__v4sf)__b);
410 /// If either value in a comparison is NaN, returns the value from \a __b.
418 /// \param __b
423 _mm_max_ps(__m128 __a, __m128 __b)
425 return __builtin_ia32_maxps((__v4sf)__a, (__v4sf)__b);
436 /// \param __b
441 _mm_and_ps(__m128 __a, __m128 __b) {
442 return (__m128)((__v4su)__a & (__v4su)__b);
456 /// \param __b
462 _mm_andnot_ps(__m128 __a, __m128 __b) {
463 return (__m128)(~(__v4su)__a & (__v4su)__b);
474 /// \param __b
479 _mm_or_ps(__m128 __a, __m128 __b) {
480 return (__m128)((__v4su)__a | (__v4su)__b);
492 /// \param __b
497 _mm_xor_ps(__m128 __a, __m128 __b) {
498 return (__m128)((__v4su)__a ^ (__v4su)__b);
515 /// \param __b
521 _mm_cmpeq_ss(__m128 __a, __m128 __b)
523 return (__m128)__builtin_ia32_cmpeqss((__v4sf)__a, (__v4sf)__b);
538 /// \param __b
542 _mm_cmpeq_ps(__m128 __a, __m128 __b)
544 return (__m128)__builtin_ia32_cmpeqps((__v4sf)__a, (__v4sf)__b);
562 /// \param __b
568 _mm_cmplt_ss(__m128 __a, __m128 __b)
570 return (__m128)__builtin_ia32_cmpltss((__v4sf)__a, (__v4sf)__b);
586 /// \param __b
590 _mm_cmplt_ps(__m128 __a, __m128 __b)
592 return (__m128)__builtin_ia32_cmpltps((__v4sf)__a, (__v4sf)__b);
610 /// \param __b
616 _mm_cmple_ss(__m128 __a, __m128 __b)
618 return (__m128)__builtin_ia32_cmpless((__v4sf)__a, (__v4sf)__b);
634 /// \param __b
638 _mm_cmple_ps(__m128 __a, __m128 __b)
640 return (__m128)__builtin_ia32_cmpleps((__v4sf)__a, (__v4sf)__b);
658 /// \param __b
664 _mm_cmpgt_ss(__m128 __a, __m128 __b)
667 (__v4sf)__builtin_ia32_cmpltss((__v4sf)__b, (__v4sf)__a),
684 /// \param __b
688 _mm_cmpgt_ps(__m128 __a, __m128 __b)
690 return (__m128)__builtin_ia32_cmpltps((__v4sf)__b, (__v4sf)__a);
708 /// \param __b
714 _mm_cmpge_ss(__m128 __a, __m128 __b)
717 (__v4sf)__builtin_ia32_cmpless((__v4sf)__b, (__v4sf)__a),
734 /// \param __b
738 _mm_cmpge_ps(__m128 __a, __m128 __b)
740 return (__m128)__builtin_ia32_cmpleps((__v4sf)__b, (__v4sf)__a);
758 /// \param __b
764 _mm_cmpneq_ss(__m128 __a, __m128 __b)
766 return (__m128)__builtin_ia32_cmpneqss((__v4sf)__a, (__v4sf)__b);
782 /// \param __b
786 _mm_cmpneq_ps(__m128 __a, __m128 __b)
788 return (__m128)__builtin_ia32_cmpneqps((__v4sf)__a, (__v4sf)__b);
807 /// \param __b
813 _mm_cmpnlt_ss(__m128 __a, __m128 __b)
815 return (__m128)__builtin_ia32_cmpnltss((__v4sf)__a, (__v4sf)__b);
832 /// \param __b
836 _mm_cmpnlt_ps(__m128 __a, __m128 __b)
838 return (__m128)__builtin_ia32_cmpnltps((__v4sf)__a, (__v4sf)__b);
857 /// \param __b
863 _mm_cmpnle_ss(__m128 __a, __m128 __b)
865 return (__m128)__builtin_ia32_cmpnless((__v4sf)__a, (__v4sf)__b);
882 /// \param __b
886 _mm_cmpnle_ps(__m128 __a, __m128 __b)
888 return (__m128)__builtin_ia32_cmpnleps((__v4sf)__a, (__v4sf)__b);
907 /// \param __b
913 _mm_cmpngt_ss(__m128 __a, __m128 __b)
916 (__v4sf)__builtin_ia32_cmpnltss((__v4sf)__b, (__v4sf)__a),
934 /// \param __b
938 _mm_cmpngt_ps(__m128 __a, __m128 __b)
940 return (__m128)__builtin_ia32_cmpnltps((__v4sf)__b, (__v4sf)__a);
959 /// \param __b
965 _mm_cmpnge_ss(__m128 __a, __m128 __b)
968 (__v4sf)__builtin_ia32_cmpnless((__v4sf)__b, (__v4sf)__a),
986 /// \param __b
990 _mm_cmpnge_ps(__m128 __a, __m128 __b)
992 return (__m128)__builtin_ia32_cmpnleps((__v4sf)__b, (__v4sf)__a);
1011 /// \param __b
1017 _mm_cmpord_ss(__m128 __a, __m128 __b)
1019 return (__m128)__builtin_ia32_cmpordss((__v4sf)__a, (__v4sf)__b);
1037 /// \param __b
1041 _mm_cmpord_ps(__m128 __a, __m128 __b)
1043 return (__m128)__builtin_ia32_cmpordps((__v4sf)__a, (__v4sf)__b);
1062 /// \param __b
1068 _mm_cmpunord_ss(__m128 __a, __m128 __b)
1070 return (__m128)__builtin_ia32_cmpunordss((__v4sf)__a, (__v4sf)__b);
1088 /// \param __b
1092 _mm_cmpunord_ps(__m128 __a, __m128 __b)
1094 return (__m128)__builtin_ia32_cmpunordps((__v4sf)__a, (__v4sf)__b);
1111 /// \param __b
1116 _mm_comieq_ss(__m128 __a, __m128 __b)
1118 return __builtin_ia32_comieq((__v4sf)__a, (__v4sf)__b);
1136 /// \param __b
1141 _mm_comilt_ss(__m128 __a, __m128 __b)
1143 return __builtin_ia32_comilt((__v4sf)__a, (__v4sf)__b);
1160 /// \param __b
1165 _mm_comile_ss(__m128 __a, __m128 __b)
1167 return __builtin_ia32_comile((__v4sf)__a, (__v4sf)__b);
1184 /// \param __b
1189 _mm_comigt_ss(__m128 __a, __m128 __b)
1191 return __builtin_ia32_comigt((__v4sf)__a, (__v4sf)__b);
1208 /// \param __b
1213 _mm_comige_ss(__m128 __a, __m128 __b)
1215 return __builtin_ia32_comige((__v4sf)__a, (__v4sf)__b);
1232 /// \param __b
1237 _mm_comineq_ss(__m128 __a, __m128 __b)
1239 return __builtin_ia32_comineq((__v4sf)__a, (__v4sf)__b);
1255 /// \param __b
1260 _mm_ucomieq_ss(__m128 __a, __m128 __b)
1262 return __builtin_ia32_ucomieq((__v4sf)__a, (__v4sf)__b);
1279 /// \param __b
1284 _mm_ucomilt_ss(__m128 __a, __m128 __b)
1286 return __builtin_ia32_ucomilt((__v4sf)__a, (__v4sf)__b);
1303 /// \param __b
1308 _mm_ucomile_ss(__m128 __a, __m128 __b)
1310 return __builtin_ia32_ucomile((__v4sf)__a, (__v4sf)__b);
1327 /// \param __b
1332 _mm_ucomigt_ss(__m128 __a, __m128 __b)
1334 return __builtin_ia32_ucomigt((__v4sf)__a, (__v4sf)__b);
1351 /// \param __b
1356 _mm_ucomige_ss(__m128 __a, __m128 __b)
1358 return __builtin_ia32_ucomige((__v4sf)__a, (__v4sf)__b);
1374 /// \param __b
1379 _mm_ucomineq_ss(__m128 __a, __m128 __b)
1381 return __builtin_ia32_ucomineq((__v4sf)__a, (__v4sf)__b);
1616 /// \param __b
1622 int __b) {
1623 __a[0] = __b;
1638 /// \param __b
1644 int __b) {
1645 return _mm_cvtsi32_ss(__a, __b);
1661 /// \param __b
1667 _mm_cvtsi64_ss(__m128 __a, long long __b) {
1668 __a[0] = __b;
1685 /// \param __b
1692 _mm_cvtpi32_ps(__m128 __a, __m64 __b)
1696 __builtin_convertvector((__v4si)__zext128(__b), __v4sf),
1711 /// \param __b
1718 _mm_cvt_pi2ps(__m128 __a, __m64 __b)
1720 return _mm_cvtpi32_ps(__a, __b);
1761 __mm_loadh_pi_v2f32 __b = ((const struct __mm_loadh_pi_struct*)__p)->__u;
1762 __m128 __bb = __builtin_shufflevector(__b, __b, 0, 1, 0, 1);
1788 __mm_loadl_pi_v2f32 __b = ((const struct __mm_loadl_pi_struct*)__p)->__u;
1789 __m128 __bb = __builtin_shufflevector(__b, __b, 0, 1, 0, 1);
2352 /// \param __b
2356 _mm_max_pi16(__m64 __a, __m64 __b)
2358 return (__m64)__builtin_elementwise_max((__v4hi)__a, (__v4hi)__b);
2371 /// \param __b
2375 _mm_max_pu8(__m64 __a, __m64 __b)
2377 return (__m64)__builtin_elementwise_max((__v8qu)__a, (__v8qu)__b);
2390 /// \param __b
2394 _mm_min_pi16(__m64 __a, __m64 __b)
2396 return (__m64)__builtin_elementwise_min((__v4hi)__a, (__v4hi)__b);
2409 /// \param __b
2413 _mm_min_pu8(__m64 __a, __m64 __b)
2415 return (__m64)__builtin_elementwise_min((__v8qu)__a, (__v8qu)__b);
2446 /// \param __b
2450 _mm_mulhi_pu16(__m64 __a, __m64 __b)
2453 (__v8hi)__anyext128(__b)));
2549 /// \param __b
2553 _mm_avg_pu8(__m64 __a, __m64 __b)
2556 (__v16qi)__anyext128(__b)));
2569 /// \param __b
2573 _mm_avg_pu16(__m64 __a, __m64 __b)
2576 (__v8hi)__anyext128(__b)));
2590 /// \param __b
2596 _mm_sad_pu8(__m64 __a, __m64 __b)
2599 (__v16qi)__zext128(__b)));
2770 /// \param __b
2776 _mm_unpackhi_ps(__m128 __a, __m128 __b) {
2777 return __builtin_shufflevector((__v4sf)__a, (__v4sf)__b, 2, 6, 3, 7);
2791 /// \param __b
2797 _mm_unpacklo_ps(__m128 __a, __m128 __b) {
2798 return __builtin_shufflevector((__v4sf)__a, (__v4sf)__b, 0, 4, 1, 5);
2813 /// \param __b
2818 _mm_move_ss(__m128 __a, __m128 __b) {
2819 __a[0] = __b[0];
2834 /// \param __b
2839 _mm_movehl_ps(__m128 __a, __m128 __b) {
2840 return __builtin_shufflevector((__v4sf)__a, (__v4sf)__b, 6, 7, 2, 3);
2854 /// \param __b
2859 _mm_movelh_ps(__m128 __a, __m128 __b) {
2860 return __builtin_shufflevector((__v4sf)__a, (__v4sf)__b, 0, 1, 4, 5);
2950 /// \param __b
2957 _mm_cvtpi32x2_ps(__m64 __a, __m64 __b)
2960 __builtin_shufflevector((__v2si)__a, (__v2si)__b,
3009 __m64 __b, __c;
3011 __b = _mm_cvtps_pi16(__a);
3014 return _mm_packs_pi16(__b, __c);