Lines Matching defs:__b
86 /// \param __b
91 _mm256_add_pd(__m256d __a, __m256d __b)
93 return (__m256d)((__v4df)__a+(__v4df)__b);
104 /// \param __b
109 _mm256_add_ps(__m256 __a, __m256 __b)
111 return (__m256)((__v8sf)__a+(__v8sf)__b);
122 /// \param __b
127 _mm256_sub_pd(__m256d __a, __m256d __b)
129 return (__m256d)((__v4df)__a-(__v4df)__b);
140 /// \param __b
145 _mm256_sub_ps(__m256 __a, __m256 __b)
147 return (__m256)((__v8sf)__a-(__v8sf)__b);
159 /// \param __b
164 _mm256_addsub_pd(__m256d __a, __m256d __b)
166 return (__m256d)__builtin_ia32_addsubpd256((__v4df)__a, (__v4df)__b);
178 /// \param __b
183 _mm256_addsub_ps(__m256 __a, __m256 __b)
185 return (__m256)__builtin_ia32_addsubps256((__v8sf)__a, (__v8sf)__b);
196 /// \param __b
201 _mm256_div_pd(__m256d __a, __m256d __b)
203 return (__m256d)((__v4df)__a/(__v4df)__b);
214 /// \param __b
219 _mm256_div_ps(__m256 __a, __m256 __b)
221 return (__m256)((__v8sf)__a/(__v8sf)__b);
227 /// If either value in a comparison is NaN, returns the value from \a __b.
235 /// \param __b
240 _mm256_max_pd(__m256d __a, __m256d __b)
242 return (__m256d)__builtin_ia32_maxpd256((__v4df)__a, (__v4df)__b);
248 /// If either value in a comparison is NaN, returns the value from \a __b.
256 /// \param __b
261 _mm256_max_ps(__m256 __a, __m256 __b)
263 return (__m256)__builtin_ia32_maxps256((__v8sf)__a, (__v8sf)__b);
269 /// If either value in a comparison is NaN, returns the value from \a __b.
277 /// \param __b
282 _mm256_min_pd(__m256d __a, __m256d __b)
284 return (__m256d)__builtin_ia32_minpd256((__v4df)__a, (__v4df)__b);
290 /// If either value in a comparison is NaN, returns the value from \a __b.
298 /// \param __b
303 _mm256_min_ps(__m256 __a, __m256 __b)
305 return (__m256)__builtin_ia32_minps256((__v8sf)__a, (__v8sf)__b);
316 /// \param __b
321 _mm256_mul_pd(__m256d __a, __m256d __b)
323 return (__m256d)((__v4df)__a * (__v4df)__b);
334 /// \param __b
339 _mm256_mul_ps(__m256 __a, __m256 __b)
341 return (__m256)((__v8sf)__a * (__v8sf)__b);
554 /// \param __b
559 _mm256_and_pd(__m256d __a, __m256d __b)
561 return (__m256d)((__v4du)__a & (__v4du)__b);
572 /// \param __b
577 _mm256_and_ps(__m256 __a, __m256 __b)
579 return (__m256)((__v8su)__a & (__v8su)__b);
592 /// \param __b
598 _mm256_andnot_pd(__m256d __a, __m256d __b)
600 return (__m256d)(~(__v4du)__a & (__v4du)__b);
613 /// \param __b
619 _mm256_andnot_ps(__m256 __a, __m256 __b)
621 return (__m256)(~(__v8su)__a & (__v8su)__b);
632 /// \param __b
637 _mm256_or_pd(__m256d __a, __m256d __b)
639 return (__m256d)((__v4du)__a | (__v4du)__b);
650 /// \param __b
655 _mm256_or_ps(__m256 __a, __m256 __b)
657 return (__m256)((__v8su)__a | (__v8su)__b);
668 /// \param __b
673 _mm256_xor_pd(__m256d __a, __m256d __b)
675 return (__m256d)((__v4du)__a ^ (__v4du)__b);
686 /// \param __b
691 _mm256_xor_ps(__m256 __a, __m256 __b)
693 return (__m256)((__v8su)__a ^ (__v8su)__b);
708 /// \param __b
715 _mm256_hadd_pd(__m256d __a, __m256d __b)
717 return (__m256d)__builtin_ia32_haddpd256((__v4df)__a, (__v4df)__b);
731 /// \param __b
738 _mm256_hadd_ps(__m256 __a, __m256 __b)
740 return (__m256)__builtin_ia32_haddps256((__v8sf)__a, (__v8sf)__b);
754 /// \param __b
761 _mm256_hsub_pd(__m256d __a, __m256d __b)
763 return (__m256d)__builtin_ia32_hsubpd256((__v4df)__a, (__v4df)__b);
777 /// \param __b
784 _mm256_hsub_ps(__m256 __a, __m256 __b)
786 return (__m256)__builtin_ia32_hsubps256((__v8sf)__a, (__v8sf)__b);
1411 /// \param __b
1419 /// 64-bit element in operand \a __b is copied to the same position in the
1423 _mm256_blendv_pd(__m256d __a, __m256d __b, __m256d __c)
1426 (__v4df)__a, (__v4df)__b, (__v4df)__c);
1439 /// \param __b
1447 /// corresponding 32-bit element in operand \a __b is copied to the same
1451 _mm256_blendv_ps(__m256 __a, __m256 __b, __m256 __c)
1454 (__v8sf)__a, (__v8sf)__b, (__v8sf)__c);
2355 __v8si __b = (__v8si)__a;
2356 return __b[0];
2460 /// \param __b
2466 _mm256_unpackhi_pd(__m256d __a, __m256d __b)
2468 return __builtin_shufflevector((__v4df)__a, (__v4df)__b, 1, 5, 1+2, 5+2);
2482 /// \param __b
2488 _mm256_unpacklo_pd(__m256d __a, __m256d __b)
2490 return __builtin_shufflevector((__v4df)__a, (__v4df)__b, 0, 4, 0+2, 4+2);
2507 /// \param __b
2515 _mm256_unpackhi_ps(__m256 __a, __m256 __b)
2517 return __builtin_shufflevector((__v8sf)__a, (__v8sf)__b, 2, 10, 2+1, 10+1, 6, 14, 6+1, 14+1);
2534 /// \param __b
2542 _mm256_unpacklo_ps(__m256 __a, __m256 __b)
2544 return __builtin_shufflevector((__v8sf)__a, (__v8sf)__b, 0, 8, 0+1, 8+1, 4, 12, 4+1, 12+1);
2568 /// \param __b
2572 _mm_testz_pd(__m128d __a, __m128d __b)
2574 return __builtin_ia32_vtestzpd((__v2df)__a, (__v2df)__b);
2597 /// \param __b
2601 _mm_testc_pd(__m128d __a, __m128d __b)
2603 return __builtin_ia32_vtestcpd((__v2df)__a, (__v2df)__b);
2627 /// \param __b
2631 _mm_testnzc_pd(__m128d __a, __m128d __b)
2633 return __builtin_ia32_vtestnzcpd((__v2df)__a, (__v2df)__b);
2656 /// \param __b
2660 _mm_testz_ps(__m128 __a, __m128 __b)
2662 return __builtin_ia32_vtestzps((__v4sf)__a, (__v4sf)__b);
2685 /// \param __b
2689 _mm_testc_ps(__m128 __a, __m128 __b)
2691 return __builtin_ia32_vtestcps((__v4sf)__a, (__v4sf)__b);
2715 /// \param __b
2719 _mm_testnzc_ps(__m128 __a, __m128 __b)
2721 return __builtin_ia32_vtestnzcps((__v4sf)__a, (__v4sf)__b);
2744 /// \param __b
2748 _mm256_testz_pd(__m256d __a, __m256d __b)
2750 return __builtin_ia32_vtestzpd256((__v4df)__a, (__v4df)__b);
2773 /// \param __b
2777 _mm256_testc_pd(__m256d __a, __m256d __b)
2779 return __builtin_ia32_vtestcpd256((__v4df)__a, (__v4df)__b);
2803 /// \param __b
2807 _mm256_testnzc_pd(__m256d __a, __m256d __b)
2809 return __builtin_ia32_vtestnzcpd256((__v4df)__a, (__v4df)__b);
2832 /// \param __b
2836 _mm256_testz_ps(__m256 __a, __m256 __b)
2838 return __builtin_ia32_vtestzps256((__v8sf)__a, (__v8sf)__b);
2861 /// \param __b
2865 _mm256_testc_ps(__m256 __a, __m256 __b)
2867 return __builtin_ia32_vtestcps256((__v8sf)__a, (__v8sf)__b);
2891 /// \param __b
2895 _mm256_testnzc_ps(__m256 __a, __m256 __b)
2897 return __builtin_ia32_vtestnzcps256((__v8sf)__a, (__v8sf)__b);
2917 /// \param __b
2921 _mm256_testz_si256(__m256i __a, __m256i __b)
2923 return __builtin_ia32_ptestz256((__v4di)__a, (__v4di)__b);
2943 /// \param __b
2947 _mm256_testc_si256(__m256i __a, __m256i __b)
2949 return __builtin_ia32_ptestc256((__v4di)__a, (__v4di)__b);
2970 /// \param __b
2974 _mm256_testnzc_si256(__m256i __a, __m256i __b)
2976 return __builtin_ia32_ptestnzc256((__v4di)__a, (__v4di)__b);
3121 __m128d __b = _mm_loadu_pd((const double *)__a);
3122 return (__m256d)__builtin_shufflevector((__v2df)__b, (__v2df)__b,
3141 __m128 __b = _mm_loadu_ps((const float *)__a);
3142 return (__m256)__builtin_shufflevector((__v4sf)__b, (__v4sf)__b,
3598 /// \param __b
3601 _mm256_stream_si256(void *__a, __m256i __b)
3604 __builtin_nontemporal_store((__v4di_aligned)__b, (__v4di_aligned*)__a);
3618 /// \param __b
3621 _mm256_stream_pd(void *__a, __m256d __b)
3624 __builtin_nontemporal_store((__v4df_aligned)__b, (__v4df_aligned*)__a);
3699 /// \param __b
3710 _mm256_set_pd(double __a, double __b, double __c, double __d)
3712 return __extension__ (__m256d){ __d, __c, __b, __a };
3726 /// \param __b
3749 _mm256_set_ps(float __a, float __b, float __c, float __d,
3752 return __extension__ (__m256){ __h, __g, __f, __e, __d, __c, __b, __a };
3939 /// \param __b
3947 _mm256_set_epi64x(long long __a, long long __b, long long __c, long long __d)
3949 return __extension__ (__m256i)(__v4di){ __d, __c, __b, __a };
3965 /// \param __b
3976 _mm256_setr_pd(double __a, double __b, double __c, double __d)
3978 return _mm256_set_pd(__d, __c, __b, __a);
3993 /// \param __b
4016 _mm256_setr_ps(float __a, float __b, float __c, float __d,
4019 return _mm256_set_ps(__h, __g, __f, __e, __d, __c, __b, __a);
4206 /// \param __b
4214 _mm256_setr_epi64x(long long __a, long long __b, long long __c, long long __d)
4216 return _mm256_set_epi64x(__d, __c, __b, __a);
4302 /// \param __b
4307 _mm256_set1_epi8(char __b)
4309 return _mm256_set_epi8(__b, __b, __b, __b, __b, __b, __b, __b,
4310 __b, __b, __b, __b, __b, __b, __b, __b,
4311 __b, __b, __b, __b, __b, __b, __b, __b,
4312 __b, __b, __b, __b, __b, __b, __b, __b);