Lines Matching defs:__a

219                                                           v128_t __a) {
224 ((struct __wasm_v128_store_struct *)__mem)->__v = __a;
485 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_splat(int8_t __a) {
486 return (v128_t)(__i8x16){__a, __a, __a, __a, __a, __a, __a, __a,
487 __a, __a, __a, __a, __a, __a, __a, __a};
490 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_splat(uint8_t __a) {
491 return (v128_t)(__u8x16){__a, __a, __a, __a, __a, __a, __a, __a,
492 __a, __a, __a, __a, __a, __a, __a, __a};
495 static __inline__ int8_t __DEFAULT_FN_ATTRS wasm_i8x16_extract_lane(v128_t __a,
498 return ((__i8x16)__a)[__i];
501 static __inline__ uint8_t __DEFAULT_FN_ATTRS wasm_u8x16_extract_lane(v128_t __a,
504 return ((__u8x16)__a)[__i];
507 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_replace_lane(v128_t __a,
511 __i8x16 __v = (__i8x16)__a;
516 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_replace_lane(v128_t __a,
520 __u8x16 __v = (__u8x16)__a;
525 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_splat(int16_t __a) {
526 return (v128_t)(__i16x8){__a, __a, __a, __a, __a, __a, __a, __a};
529 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_splat(uint16_t __a) {
530 return (v128_t)(__u16x8){__a, __a, __a, __a, __a, __a, __a, __a};
533 static __inline__ int16_t __DEFAULT_FN_ATTRS wasm_i16x8_extract_lane(v128_t __a,
536 return ((__i16x8)__a)[__i];
540 wasm_u16x8_extract_lane(v128_t __a, int __i) __REQUIRE_CONSTANT(__i) {
541 return ((__u16x8)__a)[__i];
544 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_replace_lane(v128_t __a,
548 __i16x8 __v = (__i16x8)__a;
554 v128_t __a, int __i, uint16_t __b) __REQUIRE_CONSTANT(__i) {
555 __u16x8 __v = (__u16x8)__a;
560 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_splat(int32_t __a) {
561 return (v128_t)(__i32x4){__a, __a, __a, __a};
564 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u32x4_splat(uint32_t __a) {
565 return (v128_t)(__u32x4){__a, __a, __a, __a};
568 static __inline__ int32_t __DEFAULT_FN_ATTRS wasm_i32x4_extract_lane(v128_t __a,
571 return ((__i32x4)__a)[__i];
575 wasm_u32x4_extract_lane(v128_t __a, int __i) __REQUIRE_CONSTANT(__i) {
576 return ((__u32x4)__a)[__i];
579 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_replace_lane(v128_t __a,
583 __i32x4 __v = (__i32x4)__a;
589 v128_t __a, int __i, uint32_t __b) __REQUIRE_CONSTANT(__i) {
590 __u32x4 __v = (__u32x4)__a;
595 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_splat(int64_t __a) {
596 return (v128_t)(__i64x2){__a, __a};
599 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u64x2_splat(uint64_t __a) {
600 return (v128_t)(__u64x2){__a, __a};
603 static __inline__ int64_t __DEFAULT_FN_ATTRS wasm_i64x2_extract_lane(v128_t __a,
606 return ((__i64x2)__a)[__i];
610 wasm_u64x2_extract_lane(v128_t __a, int __i) __REQUIRE_CONSTANT(__i) {
611 return ((__u64x2)__a)[__i];
614 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_replace_lane(v128_t __a,
618 __i64x2 __v = (__i64x2)__a;
624 v128_t __a, int __i, uint64_t __b) __REQUIRE_CONSTANT(__i) {
625 __u64x2 __v = (__u64x2)__a;
630 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_splat(float __a) {
631 return (v128_t)(__f32x4){__a, __a, __a, __a};
634 static __inline__ float __DEFAULT_FN_ATTRS wasm_f32x4_extract_lane(v128_t __a,
637 return ((__f32x4)__a)[__i];
640 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_replace_lane(v128_t __a,
644 __f32x4 __v = (__f32x4)__a;
649 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_splat(double __a) {
650 return (v128_t)(__f64x2){__a, __a};
653 static __inline__ double __DEFAULT_FN_ATTRS wasm_f64x2_extract_lane(v128_t __a,
656 return ((__f64x2)__a)[__i];
659 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_replace_lane(v128_t __a,
663 __f64x2 __v = (__f64x2)__a;
668 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_eq(v128_t __a,
670 return (v128_t)((__i8x16)__a == (__i8x16)__b);
673 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_ne(v128_t __a,
675 return (v128_t)((__i8x16)__a != (__i8x16)__b);
678 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_lt(v128_t __a,
680 return (v128_t)((__i8x16)__a < (__i8x16)__b);
683 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_lt(v128_t __a,
685 return (v128_t)((__u8x16)__a < (__u8x16)__b);
688 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_gt(v128_t __a,
690 return (v128_t)((__i8x16)__a > (__i8x16)__b);
693 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_gt(v128_t __a,
695 return (v128_t)((__u8x16)__a > (__u8x16)__b);
698 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_le(v128_t __a,
700 return (v128_t)((__i8x16)__a <= (__i8x16)__b);
703 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_le(v128_t __a,
705 return (v128_t)((__u8x16)__a <= (__u8x16)__b);
708 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_ge(v128_t __a,
710 return (v128_t)((__i8x16)__a >= (__i8x16)__b);
713 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_ge(v128_t __a,
715 return (v128_t)((__u8x16)__a >= (__u8x16)__b);
718 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_eq(v128_t __a,
720 return (v128_t)((__i16x8)__a == (__i16x8)__b);
723 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_ne(v128_t __a,
725 return (v128_t)((__u16x8)__a != (__u16x8)__b);
728 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_lt(v128_t __a,
730 return (v128_t)((__i16x8)__a < (__i16x8)__b);
733 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_lt(v128_t __a,
735 return (v128_t)((__u16x8)__a < (__u16x8)__b);
738 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_gt(v128_t __a,
740 return (v128_t)((__i16x8)__a > (__i16x8)__b);
743 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_gt(v128_t __a,
745 return (v128_t)((__u16x8)__a > (__u16x8)__b);
748 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_le(v128_t __a,
750 return (v128_t)((__i16x8)__a <= (__i16x8)__b);
753 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_le(v128_t __a,
755 return (v128_t)((__u16x8)__a <= (__u16x8)__b);
758 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_ge(v128_t __a,
760 return (v128_t)((__i16x8)__a >= (__i16x8)__b);
763 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_ge(v128_t __a,
765 return (v128_t)((__u16x8)__a >= (__u16x8)__b);
768 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_eq(v128_t __a,
770 return (v128_t)((__i32x4)__a == (__i32x4)__b);
773 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_ne(v128_t __a,
775 return (v128_t)((__i32x4)__a != (__i32x4)__b);
778 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_lt(v128_t __a,
780 return (v128_t)((__i32x4)__a < (__i32x4)__b);
783 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u32x4_lt(v128_t __a,
785 return (v128_t)((__u32x4)__a < (__u32x4)__b);
788 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_gt(v128_t __a,
790 return (v128_t)((__i32x4)__a > (__i32x4)__b);
793 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u32x4_gt(v128_t __a,
795 return (v128_t)((__u32x4)__a > (__u32x4)__b);
798 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_le(v128_t __a,
800 return (v128_t)((__i32x4)__a <= (__i32x4)__b);
803 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u32x4_le(v128_t __a,
805 return (v128_t)((__u32x4)__a <= (__u32x4)__b);
808 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_ge(v128_t __a,
810 return (v128_t)((__i32x4)__a >= (__i32x4)__b);
813 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u32x4_ge(v128_t __a,
815 return (v128_t)((__u32x4)__a >= (__u32x4)__b);
818 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_eq(v128_t __a,
820 return (v128_t)((__i64x2)__a == (__i64x2)__b);
823 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_ne(v128_t __a,
825 return (v128_t)((__i64x2)__a != (__i64x2)__b);
828 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_lt(v128_t __a,
830 return (v128_t)((__i64x2)__a < (__i64x2)__b);
833 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_gt(v128_t __a,
835 return (v128_t)((__i64x2)__a > (__i64x2)__b);
838 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_le(v128_t __a,
840 return (v128_t)((__i64x2)__a <= (__i64x2)__b);
843 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_ge(v128_t __a,
845 return (v128_t)((__i64x2)__a >= (__i64x2)__b);
848 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_eq(v128_t __a,
850 return (v128_t)((__f32x4)__a == (__f32x4)__b);
853 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_ne(v128_t __a,
855 return (v128_t)((__f32x4)__a != (__f32x4)__b);
858 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_lt(v128_t __a,
860 return (v128_t)((__f32x4)__a < (__f32x4)__b);
863 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_gt(v128_t __a,
865 return (v128_t)((__f32x4)__a > (__f32x4)__b);
868 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_le(v128_t __a,
870 return (v128_t)((__f32x4)__a <= (__f32x4)__b);
873 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_ge(v128_t __a,
875 return (v128_t)((__f32x4)__a >= (__f32x4)__b);
878 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_eq(v128_t __a,
880 return (v128_t)((__f64x2)__a == (__f64x2)__b);
883 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_ne(v128_t __a,
885 return (v128_t)((__f64x2)__a != (__f64x2)__b);
888 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_lt(v128_t __a,
890 return (v128_t)((__f64x2)__a < (__f64x2)__b);
893 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_gt(v128_t __a,
895 return (v128_t)((__f64x2)__a > (__f64x2)__b);
898 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_le(v128_t __a,
900 return (v128_t)((__f64x2)__a <= (__f64x2)__b);
903 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_ge(v128_t __a,
905 return (v128_t)((__f64x2)__a >= (__f64x2)__b);
908 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_not(v128_t __a) {
909 return ~__a;
912 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_and(v128_t __a,
914 return __a & __b;
917 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_or(v128_t __a,
919 return __a | __b;
922 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_xor(v128_t __a,
924 return __a ^ __b;
927 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_andnot(v128_t __a,
929 return __a & ~__b;
932 static __inline__ bool __DEFAULT_FN_ATTRS wasm_v128_any_true(v128_t __a) {
933 return __builtin_wasm_any_true_v128((__i8x16)__a);
936 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_bitselect(v128_t __a,
939 return (v128_t)__builtin_wasm_bitselect((__i32x4)__a, (__i32x4)__b,
943 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_abs(v128_t __a) {
944 return (v128_t)__builtin_wasm_abs_i8x16((__i8x16)__a);
947 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_neg(v128_t __a) {
948 return (v128_t)(-(__u8x16)__a);
951 static __inline__ bool __DEFAULT_FN_ATTRS wasm_i8x16_all_true(v128_t __a) {
952 return __builtin_wasm_all_true_i8x16((__i8x16)__a);
955 static __inline__ uint32_t __DEFAULT_FN_ATTRS wasm_i8x16_bitmask(v128_t __a) {
956 return __builtin_wasm_bitmask_i8x16((__i8x16)__a);
959 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_popcnt(v128_t __a) {
960 return (v128_t)__builtin_elementwise_popcount((__i8x16)__a);
963 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_shl(v128_t __a,
965 return (v128_t)((__i8x16)__a << (__b & 0x7));
968 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_shr(v128_t __a,
970 return (v128_t)((__i8x16)__a >> (__b & 0x7));
973 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_shr(v128_t __a,
975 return (v128_t)((__u8x16)__a >> (__b & 0x7));
978 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_add(v128_t __a,
980 return (v128_t)((__u8x16)__a + (__u8x16)__b);
983 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_add_sat(v128_t __a,
985 return (v128_t)__builtin_elementwise_add_sat((__i8x16)__a, (__i8x16)__b);
988 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_add_sat(v128_t __a,
990 return (v128_t)__builtin_elementwise_add_sat((__u8x16)__a, (__u8x16)__b);
993 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_sub(v128_t __a,
995 return (v128_t)((__u8x16)__a - (__u8x16)__b);
998 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_sub_sat(v128_t __a,
1000 return (v128_t)__builtin_elementwise_sub_sat((__i8x16)__a, (__i8x16)__b);
1003 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_sub_sat(v128_t __a,
1005 return (v128_t)__builtin_elementwise_sub_sat((__u8x16)__a, (__u8x16)__b);
1008 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_min(v128_t __a,
1010 return (v128_t)__builtin_elementwise_min((__i8x16)__a, (__i8x16)__b);
1013 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_min(v128_t __a,
1015 return (v128_t)__builtin_elementwise_min((__u8x16)__a, (__u8x16)__b);
1018 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_max(v128_t __a,
1020 return (v128_t)__builtin_elementwise_max((__i8x16)__a, (__i8x16)__b);
1023 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_max(v128_t __a,
1025 return (v128_t)__builtin_elementwise_max((__u8x16)__a, (__u8x16)__b);
1028 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_avgr(v128_t __a,
1030 return (v128_t)__builtin_wasm_avgr_u_i8x16((__u8x16)__a, (__u8x16)__b);
1033 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_abs(v128_t __a) {
1034 return (v128_t)__builtin_wasm_abs_i16x8((__i16x8)__a);
1037 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_neg(v128_t __a) {
1038 return (v128_t)(-(__u16x8)__a);
1041 static __inline__ bool __DEFAULT_FN_ATTRS wasm_i16x8_all_true(v128_t __a) {
1042 return __builtin_wasm_all_true_i16x8((__i16x8)__a);
1045 static __inline__ uint32_t __DEFAULT_FN_ATTRS wasm_i16x8_bitmask(v128_t __a) {
1046 return __builtin_wasm_bitmask_i16x8((__i16x8)__a);
1049 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_shl(v128_t __a,
1051 return (v128_t)((__i16x8)__a << (__b & 0xF));
1054 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_shr(v128_t __a,
1056 return (v128_t)((__i16x8)__a >> (__b & 0xF));
1059 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_shr(v128_t __a,
1061 return (v128_t)((__u16x8)__a >> (__b & 0xF));
1064 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_add(v128_t __a,
1066 return (v128_t)((__u16x8)__a + (__u16x8)__b);
1069 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_add_sat(v128_t __a,
1071 return (v128_t)__builtin_elementwise_add_sat((__i16x8)__a, (__i16x8)__b);
1074 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_add_sat(v128_t __a,
1076 return (v128_t)__builtin_elementwise_add_sat((__u16x8)__a, (__u16x8)__b);
1079 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_sub(v128_t __a,
1081 return (v128_t)((__i16x8)__a - (__i16x8)__b);
1084 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_sub_sat(v128_t __a,
1086 return (v128_t)__builtin_elementwise_sub_sat((__i16x8)__a, (__i16x8)__b);
1089 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_sub_sat(v128_t __a,
1091 return (v128_t)__builtin_elementwise_sub_sat((__u16x8)__a, (__u16x8)__b);
1094 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_mul(v128_t __a,
1096 return (v128_t)((__u16x8)__a * (__u16x8)__b);
1099 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_min(v128_t __a,
1101 return (v128_t)__builtin_elementwise_min((__i16x8)__a, (__i16x8)__b);
1104 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_min(v128_t __a,
1106 return (v128_t)__builtin_elementwise_min((__u16x8)__a, (__u16x8)__b);
1109 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_max(v128_t __a,
1111 return (v128_t)__builtin_elementwise_max((__i16x8)__a, (__i16x8)__b);
1114 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_max(v128_t __a,
1116 return (v128_t)__builtin_elementwise_max((__u16x8)__a, (__u16x8)__b);
1119 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_avgr(v128_t __a,
1121 return (v128_t)__builtin_wasm_avgr_u_i16x8((__u16x8)__a, (__u16x8)__b);
1124 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_abs(v128_t __a) {
1125 return (v128_t)__builtin_wasm_abs_i32x4((__i32x4)__a);
1128 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_neg(v128_t __a) {
1129 return (v128_t)(-(__u32x4)__a);
1132 static __inline__ bool __DEFAULT_FN_ATTRS wasm_i32x4_all_true(v128_t __a) {
1133 return __builtin_wasm_all_true_i32x4((__i32x4)__a);
1136 static __inline__ uint32_t __DEFAULT_FN_ATTRS wasm_i32x4_bitmask(v128_t __a) {
1137 return __builtin_wasm_bitmask_i32x4((__i32x4)__a);
1140 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_shl(v128_t __a,
1142 return (v128_t)((__i32x4)__a << (__b & 0x1F));
1145 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_shr(v128_t __a,
1147 return (v128_t)((__i32x4)__a >> (__b & 0x1F));
1150 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u32x4_shr(v128_t __a,
1152 return (v128_t)((__u32x4)__a >> (__b & 0x1F));
1155 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_add(v128_t __a,
1157 return (v128_t)((__u32x4)__a + (__u32x4)__b);
1160 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_sub(v128_t __a,
1162 return (v128_t)((__u32x4)__a - (__u32x4)__b);
1165 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_mul(v128_t __a,
1167 return (v128_t)((__u32x4)__a * (__u32x4)__b);
1170 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_min(v128_t __a,
1172 return (v128_t)__builtin_elementwise_min((__i32x4)__a, (__i32x4)__b);
1175 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u32x4_min(v128_t __a,
1177 return (v128_t)__builtin_elementwise_min((__u32x4)__a, (__u32x4)__b);
1180 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_max(v128_t __a,
1182 return (v128_t)__builtin_elementwise_max((__i32x4)__a, (__i32x4)__b);
1185 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u32x4_max(v128_t __a,
1187 return (v128_t)__builtin_elementwise_max((__u32x4)__a, (__u32x4)__b);
1190 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_dot_i16x8(v128_t __a,
1192 return (v128_t)__builtin_wasm_dot_s_i32x4_i16x8((__i16x8)__a, (__i16x8)__b);
1195 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_abs(v128_t __a) {
1196 return (v128_t)__builtin_wasm_abs_i64x2((__i64x2)__a);
1199 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_neg(v128_t __a) {
1200 return (v128_t)(-(__u64x2)__a);
1203 static __inline__ bool __DEFAULT_FN_ATTRS wasm_i64x2_all_true(v128_t __a) {
1204 return __builtin_wasm_all_true_i64x2((__i64x2)__a);
1207 static __inline__ uint32_t __DEFAULT_FN_ATTRS wasm_i64x2_bitmask(v128_t __a) {
1208 return __builtin_wasm_bitmask_i64x2((__i64x2)__a);
1211 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_shl(v128_t __a,
1213 return (v128_t)((__i64x2)__a << ((int64_t)__b & 0x3F));
1216 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_shr(v128_t __a,
1218 return (v128_t)((__i64x2)__a >> ((int64_t)__b & 0x3F));
1221 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u64x2_shr(v128_t __a,
1223 return (v128_t)((__u64x2)__a >> ((int64_t)__b & 0x3F));
1226 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_add(v128_t __a,
1228 return (v128_t)((__u64x2)__a + (__u64x2)__b);
1231 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_sub(v128_t __a,
1233 return (v128_t)((__u64x2)__a - (__u64x2)__b);
1236 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_mul(v128_t __a,
1238 return (v128_t)((__u64x2)__a * (__u64x2)__b);
1241 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_abs(v128_t __a) {
1242 return (v128_t)__builtin_wasm_abs_f32x4((__f32x4)__a);
1245 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_neg(v128_t __a) {
1246 return (v128_t)(-(__f32x4)__a);
1249 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_sqrt(v128_t __a) {
1250 return (v128_t)__builtin_wasm_sqrt_f32x4((__f32x4)__a);
1253 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_ceil(v128_t __a) {
1254 return (v128_t)__builtin_wasm_ceil_f32x4((__f32x4)__a);
1257 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_floor(v128_t __a) {
1258 return (v128_t)__builtin_wasm_floor_f32x4((__f32x4)__a);
1261 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_trunc(v128_t __a) {
1262 return (v128_t)__builtin_wasm_trunc_f32x4((__f32x4)__a);
1265 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_nearest(v128_t __a) {
1266 return (v128_t)__builtin_wasm_nearest_f32x4((__f32x4)__a);
1269 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_add(v128_t __a,
1271 return (v128_t)((__f32x4)__a + (__f32x4)__b);
1274 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_sub(v128_t __a,
1276 return (v128_t)((__f32x4)__a - (__f32x4)__b);
1279 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_mul(v128_t __a,
1281 return (v128_t)((__f32x4)__a * (__f32x4)__b);
1284 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_div(v128_t __a,
1286 return (v128_t)((__f32x4)__a / (__f32x4)__b);
1289 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_min(v128_t __a,
1291 return (v128_t)__builtin_wasm_min_f32x4((__f32x4)__a, (__f32x4)__b);
1294 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_max(v128_t __a,
1296 return (v128_t)__builtin_wasm_max_f32x4((__f32x4)__a, (__f32x4)__b);
1299 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_pmin(v128_t __a,
1301 return (v128_t)__builtin_wasm_pmin_f32x4((__f32x4)__a, (__f32x4)__b);
1304 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_pmax(v128_t __a,
1306 return (v128_t)__builtin_wasm_pmax_f32x4((__f32x4)__a, (__f32x4)__b);
1309 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_abs(v128_t __a) {
1310 return (v128_t)__builtin_wasm_abs_f64x2((__f64x2)__a);
1313 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_neg(v128_t __a) {
1314 return (v128_t)(-(__f64x2)__a);
1317 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_sqrt(v128_t __a) {
1318 return (v128_t)__builtin_wasm_sqrt_f64x2((__f64x2)__a);
1321 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_ceil(v128_t __a) {
1322 return (v128_t)__builtin_wasm_ceil_f64x2((__f64x2)__a);
1325 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_floor(v128_t __a) {
1326 return (v128_t)__builtin_wasm_floor_f64x2((__f64x2)__a);
1329 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_trunc(v128_t __a) {
1330 return (v128_t)__builtin_wasm_trunc_f64x2((__f64x2)__a);
1333 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_nearest(v128_t __a) {
1334 return (v128_t)__builtin_wasm_nearest_f64x2((__f64x2)__a);
1337 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_add(v128_t __a,
1339 return (v128_t)((__f64x2)__a + (__f64x2)__b);
1342 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_sub(v128_t __a,
1344 return (v128_t)((__f64x2)__a - (__f64x2)__b);
1347 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_mul(v128_t __a,
1349 return (v128_t)((__f64x2)__a * (__f64x2)__b);
1352 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_div(v128_t __a,
1354 return (v128_t)((__f64x2)__a / (__f64x2)__b);
1357 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_min(v128_t __a,
1359 return (v128_t)__builtin_wasm_min_f64x2((__f64x2)__a, (__f64x2)__b);
1362 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_max(v128_t __a,
1364 return (v128_t)__builtin_wasm_max_f64x2((__f64x2)__a, (__f64x2)__b);
1367 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_pmin(v128_t __a,
1369 return (v128_t)__builtin_wasm_pmin_f64x2((__f64x2)__a, (__f64x2)__b);
1372 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_pmax(v128_t __a,
1374 return (v128_t)__builtin_wasm_pmax_f64x2((__f64x2)__a, (__f64x2)__b);
1378 wasm_i32x4_trunc_sat_f32x4(v128_t __a) {
1379 return (v128_t)__builtin_wasm_trunc_saturate_s_i32x4_f32x4((__f32x4)__a);
1383 wasm_u32x4_trunc_sat_f32x4(v128_t __a) {
1384 return (v128_t)__builtin_wasm_trunc_saturate_u_i32x4_f32x4((__f32x4)__a);
1388 wasm_f32x4_convert_i32x4(v128_t __a) {
1389 return (v128_t) __builtin_convertvector((__i32x4)__a, __f32x4);
1393 wasm_f32x4_convert_u32x4(v128_t __a) {
1394 return (v128_t) __builtin_convertvector((__u32x4)__a, __f32x4);
1398 wasm_f64x2_convert_low_i32x4(v128_t __a) {
1399 return (v128_t) __builtin_convertvector((__i32x2){__a[0], __a[1]}, __f64x2);
1403 wasm_f64x2_convert_low_u32x4(v128_t __a) {
1404 return (v128_t) __builtin_convertvector((__u32x2){__a[0], __a[1]}, __f64x2);
1408 wasm_i32x4_trunc_sat_f64x2_zero(v128_t __a) {
1409 return (v128_t)__builtin_wasm_trunc_sat_s_zero_f64x2_i32x4((__f64x2)__a);
1413 wasm_u32x4_trunc_sat_f64x2_zero(v128_t __a) {
1414 return (v128_t)__builtin_wasm_trunc_sat_u_zero_f64x2_i32x4((__f64x2)__a);
1418 wasm_f32x4_demote_f64x2_zero(v128_t __a) {
1420 __builtin_shufflevector((__f64x2)__a, (__f64x2){0, 0}, 0, 1, 2, 3),
1425 wasm_f64x2_promote_low_f32x4(v128_t __a) {
1427 (__f32x2){((__f32x4)__a)[0], ((__f32x4)__a)[1]}, __f64x2);
1430 #define wasm_i8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \
1434 (__i8x16)(__a), (__i8x16)(__b), __c0, __c1, __c2, __c3, __c4, __c5, \
1437 #define wasm_i16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \
1440 (__i8x16)(__a), (__i8x16)(__b), (__c0)*2, (__c0)*2 + 1, (__c1)*2, \
1445 #define wasm_i32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3) \
1447 (__i8x16)(__a), (__i8x16)(__b), (__c0)*4, (__c0)*4 + 1, (__c0)*4 + 2, \
1452 #define wasm_i64x2_shuffle(__a, __b, __c0, __c1) \
1454 (__i8x16)(__a), (__i8x16)(__b), (__c0)*8, (__c0)*8 + 1, (__c0)*8 + 2, \
1459 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_swizzle(v128_t __a,
1461 return (v128_t)__builtin_wasm_swizzle_i8x16((__i8x16)__a, (__i8x16)__b);
1465 wasm_i8x16_narrow_i16x8(v128_t __a, v128_t __b) {
1466 return (v128_t)__builtin_wasm_narrow_s_i8x16_i16x8((__i16x8)__a,
1471 wasm_u8x16_narrow_i16x8(v128_t __a, v128_t __b) {
1472 return (v128_t)__builtin_wasm_narrow_u_i8x16_i16x8((__i16x8)__a,
1477 wasm_i16x8_narrow_i32x4(v128_t __a, v128_t __b) {
1478 return (v128_t)__builtin_wasm_narrow_s_i16x8_i32x4((__i32x4)__a,
1483 wasm_u16x8_narrow_i32x4(v128_t __a, v128_t __b) {
1484 return (v128_t)__builtin_wasm_narrow_u_i16x8_i32x4((__i32x4)__a,
1489 wasm_i16x8_extend_low_i8x16(v128_t __a) {
1491 (__i8x8){((__i8x16)__a)[0], ((__i8x16)__a)[1], ((__i8x16)__a)[2],
1492 ((__i8x16)__a)[3], ((__i8x16)__a)[4], ((__i8x16)__a)[5],
1493 ((__i8x16)__a)[6], ((__i8x16)__a)[7]},
1498 wasm_i16x8_extend_high_i8x16(v128_t __a) {
1500 (__i8x8){((__i8x16)__a)[8], ((__i8x16)__a)[9], ((__i8x16)__a)[10],
1501 ((__i8x16)__a)[11], ((__i8x16)__a)[12], ((__i8x16)__a)[13],
1502 ((__i8x16)__a)[14], ((__i8x16)__a)[15]},
1507 wasm_u16x8_extend_low_u8x16(v128_t __a) {
1509 (__u8x8){((__u8x16)__a)[0], ((__u8x16)__a)[1], ((__u8x16)__a)[2],
1510 ((__u8x16)__a)[3], ((__u8x16)__a)[4], ((__u8x16)__a)[5],
1511 ((__u8x16)__a)[6], ((__u8x16)__a)[7]},
1516 wasm_u16x8_extend_high_u8x16(v128_t __a) {
1518 (__u8x8){((__u8x16)__a)[8], ((__u8x16)__a)[9], ((__u8x16)__a)[10],
1519 ((__u8x16)__a)[11], ((__u8x16)__a)[12], ((__u8x16)__a)[13],
1520 ((__u8x16)__a)[14], ((__u8x16)__a)[15]},
1525 wasm_i32x4_extend_low_i16x8(v128_t __a) {
1527 (__i16x4){((__i16x8)__a)[0], ((__i16x8)__a)[1], ((__i16x8)__a)[2],
1528 ((__i16x8)__a)[3]},
1533 wasm_i32x4_extend_high_i16x8(v128_t __a) {
1535 (__i16x4){((__i16x8)__a)[4], ((__i16x8)__a)[5], ((__i16x8)__a)[6],
1536 ((__i16x8)__a)[7]},
1541 wasm_u32x4_extend_low_u16x8(v128_t __a) {
1543 (__u16x4){((__u16x8)__a)[0], ((__u16x8)__a)[1], ((__u16x8)__a)[2],
1544 ((__u16x8)__a)[3]},
1549 wasm_u32x4_extend_high_u16x8(v128_t __a) {
1551 (__u16x4){((__u16x8)__a)[4], ((__u16x8)__a)[5], ((__u16x8)__a)[6],
1552 ((__u16x8)__a)[7]},
1557 wasm_i64x2_extend_low_i32x4(v128_t __a) {
1559 (__i32x2){((__i32x4)__a)[0], ((__i32x4)__a)[1]}, __i64x2);
1563 wasm_i64x2_extend_high_i32x4(v128_t __a) {
1565 (__i32x2){((__i32x4)__a)[2], ((__i32x4)__a)[3]}, __i64x2);
1569 wasm_u64x2_extend_low_u32x4(v128_t __a) {
1571 (__u32x2){((__u32x4)__a)[0], ((__u32x4)__a)[1]}, __u64x2);
1575 wasm_u64x2_extend_high_u32x4(v128_t __a) {
1577 (__u32x2){((__u32x4)__a)[2], ((__u32x4)__a)[3]}, __u64x2);
1581 wasm_i16x8_extadd_pairwise_i8x16(v128_t __a) {
1582 return (v128_t)__builtin_wasm_extadd_pairwise_i8x16_s_i16x8((__i8x16)__a);
1586 wasm_u16x8_extadd_pairwise_u8x16(v128_t __a) {
1587 return (v128_t)__builtin_wasm_extadd_pairwise_i8x16_u_i16x8((__u8x16)__a);
1591 wasm_i32x4_extadd_pairwise_i16x8(v128_t __a) {
1592 return (v128_t)__builtin_wasm_extadd_pairwise_i16x8_s_i32x4((__i16x8)__a);
1596 wasm_u32x4_extadd_pairwise_u16x8(v128_t __a) {
1597 return (v128_t)__builtin_wasm_extadd_pairwise_i16x8_u_i32x4((__u16x8)__a);
1601 wasm_i16x8_extmul_low_i8x16(v128_t __a, v128_t __b) {
1602 return (v128_t)((__i16x8)wasm_i16x8_extend_low_i8x16(__a) *
1607 wasm_i16x8_extmul_high_i8x16(v128_t __a, v128_t __b) {
1608 return (v128_t)((__i16x8)wasm_i16x8_extend_high_i8x16(__a) *
1613 wasm_u16x8_extmul_low_u8x16(v128_t __a, v128_t __b) {
1614 return (v128_t)((__u16x8)wasm_u16x8_extend_low_u8x16(__a) *
1619 wasm_u16x8_extmul_high_u8x16(v128_t __a, v128_t __b) {
1620 return (v128_t)((__u16x8)wasm_u16x8_extend_high_u8x16(__a) *
1625 wasm_i32x4_extmul_low_i16x8(v128_t __a, v128_t __b) {
1626 return (v128_t)((__i32x4)wasm_i32x4_extend_low_i16x8(__a) *
1631 wasm_i32x4_extmul_high_i16x8(v128_t __a, v128_t __b) {
1632 return (v128_t)((__i32x4)wasm_i32x4_extend_high_i16x8(__a) *
1637 wasm_u32x4_extmul_low_u16x8(v128_t __a, v128_t __b) {
1638 return (v128_t)((__u32x4)wasm_u32x4_extend_low_u16x8(__a) *
1643 wasm_u32x4_extmul_high_u16x8(v128_t __a, v128_t __b) {
1644 return (v128_t)((__u32x4)wasm_u32x4_extend_high_u16x8(__a) *
1649 wasm_i64x2_extmul_low_i32x4(v128_t __a, v128_t __b) {
1650 return (v128_t)((__i64x2)wasm_i64x2_extend_low_i32x4(__a) *
1655 wasm_i64x2_extmul_high_i32x4(v128_t __a, v128_t __b) {
1656 return (v128_t)((__i64x2)wasm_i64x2_extend_high_i32x4(__a) *
1661 wasm_u64x2_extmul_low_u32x4(v128_t __a, v128_t __b) {
1662 return (v128_t)((__u64x2)wasm_u64x2_extend_low_u32x4(__a) *
1667 wasm_u64x2_extmul_high_u32x4(v128_t __a, v128_t __b) {
1668 return (v128_t)((__u64x2)wasm_u64x2_extend_high_u32x4(__a) *
1672 static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_q15mulr_sat(v128_t __a,
1674 return (v128_t)__builtin_wasm_q15mulr_sat_s_i16x8((__i16x8)__a, (__i16x8)__b);
1744 #define wasm_v8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \
1748 wasm_i8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7, \
1751 #define wasm_v16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \
1754 wasm_i16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7)
1756 #define wasm_v32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3) \
1758 wasm_i32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3)
1760 #define wasm_v64x2_shuffle(__a, __b, __c0, __c1) \
1762 wasm_i64x2_shuffle(__a, __b, __c0, __c1)
1771 wasm_f32x4_relaxed_madd(v128_t __a, v128_t __b, v128_t __c) {
1772 return (v128_t)__builtin_wasm_relaxed_madd_f32x4((__f32x4)__a, (__f32x4)__b,
1777 wasm_f32x4_relaxed_nmadd(v128_t __a, v128_t __b, v128_t __c) {
1778 return (v128_t)__builtin_wasm_relaxed_nmadd_f32x4((__f32x4)__a, (__f32x4)__b,
1783 wasm_f64x2_relaxed_madd(v128_t __a, v128_t __b, v128_t __c) {
1784 return (v128_t)__builtin_wasm_relaxed_madd_f64x2((__f64x2)__a, (__f64x2)__b,
1789 wasm_f64x2_relaxed_nmadd(v128_t __a, v128_t __b, v128_t __c) {
1790 return (v128_t)__builtin_wasm_relaxed_nmadd_f64x2((__f64x2)__a, (__f64x2)__b,
1795 wasm_i8x16_relaxed_laneselect(v128_t __a, v128_t __b, v128_t __m) {
1797 (__i8x16)__a, (__i8x16)__b, (__i8x16)__m);
1801 wasm_i16x8_relaxed_laneselect(v128_t __a, v128_t __b, v128_t __m) {
1803 (__i16x8)__a, (__i16x8)__b, (__i16x8)__m);
1807 wasm_i32x4_relaxed_laneselect(v128_t __a, v128_t __b, v128_t __m) {
1809 (__i32x4)__a, (__i32x4)__b, (__i32x4)__m);
1813 wasm_i64x2_relaxed_laneselect(v128_t __a, v128_t __b, v128_t __m) {
1815 (__i64x2)__a, (__i64x2)__b, (__i64x2)__m);
1819 wasm_i8x16_relaxed_swizzle(v128_t __a, v128_t __s) {
1820 return (v128_t)__builtin_wasm_relaxed_swizzle_i8x16((__i8x16)__a,
1824 static __inline__ v128_t __RELAXED_FN_ATTRS wasm_f32x4_relaxed_min(v128_t __a,
1826 return (v128_t)__builtin_wasm_relaxed_min_f32x4((__f32x4)__a, (__f32x4)__b);
1829 static __inline__ v128_t __RELAXED_FN_ATTRS wasm_f32x4_relaxed_max(v128_t __a,
1831 return (v128_t)__builtin_wasm_relaxed_max_f32x4((__f32x4)__a, (__f32x4)__b);
1834 static __inline__ v128_t __RELAXED_FN_ATTRS wasm_f64x2_relaxed_min(v128_t __a,
1836 return (v128_t)__builtin_wasm_relaxed_min_f64x2((__f64x2)__a, (__f64x2)__b);
1839 static __inline__ v128_t __RELAXED_FN_ATTRS wasm_f64x2_relaxed_max(v128_t __a,
1841 return (v128_t)__builtin_wasm_relaxed_max_f64x2((__f64x2)__a, (__f64x2)__b);
1845 wasm_i32x4_relaxed_trunc_f32x4(v128_t __a) {
1846 return (v128_t)__builtin_wasm_relaxed_trunc_s_i32x4_f32x4((__f32x4)__a);
1850 wasm_u32x4_relaxed_trunc_f32x4(v128_t __a) {
1851 return (v128_t)__builtin_wasm_relaxed_trunc_u_i32x4_f32x4((__f32x4)__a);
1855 wasm_i32x4_relaxed_trunc_f64x2_zero(v128_t __a) {
1856 return (v128_t)__builtin_wasm_relaxed_trunc_s_zero_i32x4_f64x2((__f64x2)__a);
1860 wasm_u32x4_relaxed_trunc_f64x2_zero(v128_t __a) {
1861 return (v128_t)__builtin_wasm_relaxed_trunc_u_zero_i32x4_f64x2((__f64x2)__a);
1865 wasm_i16x8_relaxed_q15mulr(v128_t __a, v128_t __b) {
1866 return (v128_t)__builtin_wasm_relaxed_q15mulr_s_i16x8((__i16x8)__a,
1871 wasm_i16x8_relaxed_dot_i8x16_i7x16(v128_t __a, v128_t __b) {
1872 return (v128_t)__builtin_wasm_relaxed_dot_i8x16_i7x16_s_i16x8((__i8x16)__a,
1877 wasm_i32x4_relaxed_dot_i8x16_i7x16_add(v128_t __a, v128_t __b, v128_t __c) {
1879 (__i8x16)__a, (__i8x16)__b, (__i32x4)__c);
1887 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_splat(float __a) {
1888 return (v128_t)__builtin_wasm_splat_f16x8(__a);
1895 #define wasm_f16x8_extract_lane(__a, __i) \
1896 (__builtin_wasm_extract_lane_f16x8((__f16x8)(__a), __i))
1898 #define wasm_f16x8_replace_lane(__a, __i, __b) \
1899 ((v128_t)__builtin_wasm_replace_lane_f16x8((__f16x8)(__a), __i, __b))
1903 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_abs(v128_t __a) {
1904 return (v128_t)__builtin_wasm_abs_f16x8((__f16x8)__a);
1907 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_neg(v128_t __a) {
1908 return (v128_t)(-(__f16x8)__a);
1911 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_sqrt(v128_t __a) {
1912 return (v128_t)__builtin_wasm_sqrt_f16x8((__f16x8)__a);
1915 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_ceil(v128_t __a) {
1916 return (v128_t)__builtin_wasm_ceil_f16x8((__f16x8)__a);
1919 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_floor(v128_t __a) {
1920 return (v128_t)__builtin_wasm_floor_f16x8((__f16x8)__a);
1923 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_trunc(v128_t __a) {
1924 return (v128_t)__builtin_wasm_trunc_f16x8((__f16x8)__a);
1927 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_nearest(v128_t __a) {
1928 return (v128_t)__builtin_wasm_nearest_f16x8((__f16x8)__a);
1931 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_eq(v128_t __a, v128_t __b) {
1932 return (v128_t)((__f16x8)__a == (__f16x8)__b);
1935 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_ne(v128_t __a, v128_t __b) {
1936 return (v128_t)((__f16x8)__a != (__f16x8)__b);
1939 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_lt(v128_t __a, v128_t __b) {
1940 return (v128_t)((__f16x8)__a < (__f16x8)__b);
1943 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_gt(v128_t __a, v128_t __b) {
1944 return (v128_t)((__f16x8)__a > (__f16x8)__b);
1947 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_le(v128_t __a, v128_t __b) {
1948 return (v128_t)((__f16x8)__a <= (__f16x8)__b);
1951 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_ge(v128_t __a, v128_t __b) {
1952 return (v128_t)((__f16x8)__a >= (__f16x8)__b);
1955 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_add(v128_t __a,
1957 return (v128_t)((__f16x8)__a + (__f16x8)__b);
1960 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_sub(v128_t __a,
1962 return (v128_t)((__f16x8)__a - (__f16x8)__b);
1965 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_mul(v128_t __a,
1967 return (v128_t)((__f16x8)__a * (__f16x8)__b);
1970 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_div(v128_t __a,
1972 return (v128_t)((__f16x8)__a / (__f16x8)__b);
1975 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_min(v128_t __a,
1977 return (v128_t)__builtin_wasm_min_f16x8((__f16x8)__a, (__f16x8)__b);
1980 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_max(v128_t __a,
1982 return (v128_t)__builtin_wasm_max_f16x8((__f16x8)__a, (__f16x8)__b);
1985 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_pmin(v128_t __a,
1987 return (v128_t)__builtin_wasm_pmin_f16x8((__f16x8)__a, (__f16x8)__b);
1990 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_pmax(v128_t __a,
1992 return (v128_t)__builtin_wasm_pmax_f16x8((__f16x8)__a, (__f16x8)__b);
1996 wasm_i16x8_trunc_sat_f16x8(v128_t __a) {
1997 return (v128_t)__builtin_wasm_trunc_saturate_s_i16x8_f16x8((__f16x8)__a);
2001 wasm_u16x8_trunc_sat_f16x8(v128_t __a) {
2002 return (v128_t)__builtin_wasm_trunc_saturate_u_i16x8_f16x8((__f16x8)__a);
2005 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_convert_i16x8(v128_t __a) {
2006 return (v128_t) __builtin_convertvector((__i16x8)__a, __f16x8);
2009 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_convert_u16x8(v128_t __a) {
2010 return (v128_t) __builtin_convertvector((__u16x8)__a, __f16x8);
2013 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_relaxed_madd(v128_t __a,
2016 return (v128_t)__builtin_wasm_relaxed_madd_f16x8((__f16x8)__a, (__f16x8)__b,
2020 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_relaxed_nmadd(v128_t __a,
2023 return (v128_t)__builtin_wasm_relaxed_nmadd_f16x8((__f16x8)__a, (__f16x8)__b,
2030 wasm_v8x16_swizzle(v128_t __a, v128_t __b) {
2031 return wasm_i8x16_swizzle(__a, __b);
2035 wasm_i8x16_any_true(v128_t __a) {
2036 return wasm_v128_any_true(__a);
2040 wasm_i16x8_any_true(v128_t __a) {
2041 return wasm_v128_any_true(__a);
2045 wasm_i32x4_any_true(v128_t __a) {
2046 return wasm_v128_any_true(__a);
2050 wasm_i8x16_add_saturate(v128_t __a, v128_t __b) {
2051 return wasm_i8x16_add_sat(__a, __b);
2055 wasm_u8x16_add_saturate(v128_t __a, v128_t __b) {
2056 return wasm_u8x16_add_sat(__a, __b);
2060 wasm_i8x16_sub_saturate(v128_t __a, v128_t __b) {
2061 return wasm_i8x16_sub_sat(__a, __b);
2065 wasm_u8x16_sub_saturate(v128_t __a, v128_t __b) {
2066 return wasm_u8x16_sub_sat(__a, __b);
2070 wasm_i16x8_add_saturate(v128_t __a, v128_t __b) {
2071 return wasm_i16x8_add_sat(__a, __b);
2075 wasm_u16x8_add_saturate(v128_t __a, v128_t __b) {
2076 return wasm_u16x8_add_sat(__a, __b);
2080 wasm_i16x8_sub_saturate(v128_t __a, v128_t __b) {
2081 return wasm_i16x8_sub_sat(__a, __b);
2085 wasm_u16x8_sub_saturate(v128_t __a, v128_t __b) {
2086 return wasm_u16x8_sub_sat(__a, __b);
2090 wasm_i16x8_widen_low_i8x16(v128_t __a) {
2091 return wasm_i16x8_extend_low_i8x16(__a);
2095 wasm_i16x8_widen_high_i8x16(v128_t __a) {
2096 return wasm_i16x8_extend_high_i8x16(__a);
2100 wasm_i16x8_widen_low_u8x16(v128_t __a) {
2101 return wasm_u16x8_extend_low_u8x16(__a);
2105 wasm_i16x8_widen_high_u8x16(v128_t __a) {
2106 return wasm_u16x8_extend_high_u8x16(__a);
2110 wasm_i32x4_widen_low_i16x8(v128_t __a) {
2111 return wasm_i32x4_extend_low_i16x8(__a);
2115 wasm_i32x4_widen_high_i16x8(v128_t __a) {
2116 return wasm_i32x4_extend_high_i16x8(__a);
2120 wasm_i32x4_widen_low_u16x8(v128_t __a) {
2121 return wasm_u32x4_extend_low_u16x8(__a);
2125 wasm_i32x4_widen_high_u16x8(v128_t __a) {
2126 return wasm_u32x4_extend_high_u16x8(__a);
2130 wasm_i32x4_trunc_saturate_f32x4(v128_t __a) {
2131 return wasm_i32x4_trunc_sat_f32x4(__a);
2135 wasm_u32x4_trunc_saturate_f32x4(v128_t __a) {
2136 return wasm_u32x4_trunc_sat_f32x4(__a);