Lines Matching defs:__b
509 int8_t __b)
512 __v[__i] = __b;
518 uint8_t __b)
521 __v[__i] = __b;
546 int16_t __b)
549 __v[__i] = __b;
554 v128_t __a, int __i, uint16_t __b) __REQUIRE_CONSTANT(__i) {
556 __v[__i] = __b;
581 int32_t __b)
584 __v[__i] = __b;
589 v128_t __a, int __i, uint32_t __b) __REQUIRE_CONSTANT(__i) {
591 __v[__i] = __b;
616 int64_t __b)
619 __v[__i] = __b;
624 v128_t __a, int __i, uint64_t __b) __REQUIRE_CONSTANT(__i) {
626 __v[__i] = __b;
642 float __b)
645 __v[__i] = __b;
661 double __b)
664 __v[__i] = __b;
669 v128_t __b) {
670 return (v128_t)((__i8x16)__a == (__i8x16)__b);
674 v128_t __b) {
675 return (v128_t)((__i8x16)__a != (__i8x16)__b);
679 v128_t __b) {
680 return (v128_t)((__i8x16)__a < (__i8x16)__b);
684 v128_t __b) {
685 return (v128_t)((__u8x16)__a < (__u8x16)__b);
689 v128_t __b) {
690 return (v128_t)((__i8x16)__a > (__i8x16)__b);
694 v128_t __b) {
695 return (v128_t)((__u8x16)__a > (__u8x16)__b);
699 v128_t __b) {
700 return (v128_t)((__i8x16)__a <= (__i8x16)__b);
704 v128_t __b) {
705 return (v128_t)((__u8x16)__a <= (__u8x16)__b);
709 v128_t __b) {
710 return (v128_t)((__i8x16)__a >= (__i8x16)__b);
714 v128_t __b) {
715 return (v128_t)((__u8x16)__a >= (__u8x16)__b);
719 v128_t __b) {
720 return (v128_t)((__i16x8)__a == (__i16x8)__b);
724 v128_t __b) {
725 return (v128_t)((__u16x8)__a != (__u16x8)__b);
729 v128_t __b) {
730 return (v128_t)((__i16x8)__a < (__i16x8)__b);
734 v128_t __b) {
735 return (v128_t)((__u16x8)__a < (__u16x8)__b);
739 v128_t __b) {
740 return (v128_t)((__i16x8)__a > (__i16x8)__b);
744 v128_t __b) {
745 return (v128_t)((__u16x8)__a > (__u16x8)__b);
749 v128_t __b) {
750 return (v128_t)((__i16x8)__a <= (__i16x8)__b);
754 v128_t __b) {
755 return (v128_t)((__u16x8)__a <= (__u16x8)__b);
759 v128_t __b) {
760 return (v128_t)((__i16x8)__a >= (__i16x8)__b);
764 v128_t __b) {
765 return (v128_t)((__u16x8)__a >= (__u16x8)__b);
769 v128_t __b) {
770 return (v128_t)((__i32x4)__a == (__i32x4)__b);
774 v128_t __b) {
775 return (v128_t)((__i32x4)__a != (__i32x4)__b);
779 v128_t __b) {
780 return (v128_t)((__i32x4)__a < (__i32x4)__b);
784 v128_t __b) {
785 return (v128_t)((__u32x4)__a < (__u32x4)__b);
789 v128_t __b) {
790 return (v128_t)((__i32x4)__a > (__i32x4)__b);
794 v128_t __b) {
795 return (v128_t)((__u32x4)__a > (__u32x4)__b);
799 v128_t __b) {
800 return (v128_t)((__i32x4)__a <= (__i32x4)__b);
804 v128_t __b) {
805 return (v128_t)((__u32x4)__a <= (__u32x4)__b);
809 v128_t __b) {
810 return (v128_t)((__i32x4)__a >= (__i32x4)__b);
814 v128_t __b) {
815 return (v128_t)((__u32x4)__a >= (__u32x4)__b);
819 v128_t __b) {
820 return (v128_t)((__i64x2)__a == (__i64x2)__b);
824 v128_t __b) {
825 return (v128_t)((__i64x2)__a != (__i64x2)__b);
829 v128_t __b) {
830 return (v128_t)((__i64x2)__a < (__i64x2)__b);
834 v128_t __b) {
835 return (v128_t)((__i64x2)__a > (__i64x2)__b);
839 v128_t __b) {
840 return (v128_t)((__i64x2)__a <= (__i64x2)__b);
844 v128_t __b) {
845 return (v128_t)((__i64x2)__a >= (__i64x2)__b);
849 v128_t __b) {
850 return (v128_t)((__f32x4)__a == (__f32x4)__b);
854 v128_t __b) {
855 return (v128_t)((__f32x4)__a != (__f32x4)__b);
859 v128_t __b) {
860 return (v128_t)((__f32x4)__a < (__f32x4)__b);
864 v128_t __b) {
865 return (v128_t)((__f32x4)__a > (__f32x4)__b);
869 v128_t __b) {
870 return (v128_t)((__f32x4)__a <= (__f32x4)__b);
874 v128_t __b) {
875 return (v128_t)((__f32x4)__a >= (__f32x4)__b);
879 v128_t __b) {
880 return (v128_t)((__f64x2)__a == (__f64x2)__b);
884 v128_t __b) {
885 return (v128_t)((__f64x2)__a != (__f64x2)__b);
889 v128_t __b) {
890 return (v128_t)((__f64x2)__a < (__f64x2)__b);
894 v128_t __b) {
895 return (v128_t)((__f64x2)__a > (__f64x2)__b);
899 v128_t __b) {
900 return (v128_t)((__f64x2)__a <= (__f64x2)__b);
904 v128_t __b) {
905 return (v128_t)((__f64x2)__a >= (__f64x2)__b);
913 v128_t __b) {
914 return __a & __b;
918 v128_t __b) {
919 return __a | __b;
923 v128_t __b) {
924 return __a ^ __b;
928 v128_t __b) {
929 return __a & ~__b;
937 v128_t __b,
939 return (v128_t)__builtin_wasm_bitselect((__i32x4)__a, (__i32x4)__b,
964 uint32_t __b) {
965 return (v128_t)((__i8x16)__a << (__b & 0x7));
969 uint32_t __b) {
970 return (v128_t)((__i8x16)__a >> (__b & 0x7));
974 uint32_t __b) {
975 return (v128_t)((__u8x16)__a >> (__b & 0x7));
979 v128_t __b) {
980 return (v128_t)((__u8x16)__a + (__u8x16)__b);
984 v128_t __b) {
985 return (v128_t)__builtin_elementwise_add_sat((__i8x16)__a, (__i8x16)__b);
989 v128_t __b) {
990 return (v128_t)__builtin_elementwise_add_sat((__u8x16)__a, (__u8x16)__b);
994 v128_t __b) {
995 return (v128_t)((__u8x16)__a - (__u8x16)__b);
999 v128_t __b) {
1000 return (v128_t)__builtin_elementwise_sub_sat((__i8x16)__a, (__i8x16)__b);
1004 v128_t __b) {
1005 return (v128_t)__builtin_elementwise_sub_sat((__u8x16)__a, (__u8x16)__b);
1009 v128_t __b) {
1010 return (v128_t)__builtin_elementwise_min((__i8x16)__a, (__i8x16)__b);
1014 v128_t __b) {
1015 return (v128_t)__builtin_elementwise_min((__u8x16)__a, (__u8x16)__b);
1019 v128_t __b) {
1020 return (v128_t)__builtin_elementwise_max((__i8x16)__a, (__i8x16)__b);
1024 v128_t __b) {
1025 return (v128_t)__builtin_elementwise_max((__u8x16)__a, (__u8x16)__b);
1029 v128_t __b) {
1030 return (v128_t)__builtin_wasm_avgr_u_i8x16((__u8x16)__a, (__u8x16)__b);
1050 uint32_t __b) {
1051 return (v128_t)((__i16x8)__a << (__b & 0xF));
1055 uint32_t __b) {
1056 return (v128_t)((__i16x8)__a >> (__b & 0xF));
1060 uint32_t __b) {
1061 return (v128_t)((__u16x8)__a >> (__b & 0xF));
1065 v128_t __b) {
1066 return (v128_t)((__u16x8)__a + (__u16x8)__b);
1070 v128_t __b) {
1071 return (v128_t)__builtin_elementwise_add_sat((__i16x8)__a, (__i16x8)__b);
1075 v128_t __b) {
1076 return (v128_t)__builtin_elementwise_add_sat((__u16x8)__a, (__u16x8)__b);
1080 v128_t __b) {
1081 return (v128_t)((__i16x8)__a - (__i16x8)__b);
1085 v128_t __b) {
1086 return (v128_t)__builtin_elementwise_sub_sat((__i16x8)__a, (__i16x8)__b);
1090 v128_t __b) {
1091 return (v128_t)__builtin_elementwise_sub_sat((__u16x8)__a, (__u16x8)__b);
1095 v128_t __b) {
1096 return (v128_t)((__u16x8)__a * (__u16x8)__b);
1100 v128_t __b) {
1101 return (v128_t)__builtin_elementwise_min((__i16x8)__a, (__i16x8)__b);
1105 v128_t __b) {
1106 return (v128_t)__builtin_elementwise_min((__u16x8)__a, (__u16x8)__b);
1110 v128_t __b) {
1111 return (v128_t)__builtin_elementwise_max((__i16x8)__a, (__i16x8)__b);
1115 v128_t __b) {
1116 return (v128_t)__builtin_elementwise_max((__u16x8)__a, (__u16x8)__b);
1120 v128_t __b) {
1121 return (v128_t)__builtin_wasm_avgr_u_i16x8((__u16x8)__a, (__u16x8)__b);
1141 uint32_t __b) {
1142 return (v128_t)((__i32x4)__a << (__b & 0x1F));
1146 uint32_t __b) {
1147 return (v128_t)((__i32x4)__a >> (__b & 0x1F));
1151 uint32_t __b) {
1152 return (v128_t)((__u32x4)__a >> (__b & 0x1F));
1156 v128_t __b) {
1157 return (v128_t)((__u32x4)__a + (__u32x4)__b);
1161 v128_t __b) {
1162 return (v128_t)((__u32x4)__a - (__u32x4)__b);
1166 v128_t __b) {
1167 return (v128_t)((__u32x4)__a * (__u32x4)__b);
1171 v128_t __b) {
1172 return (v128_t)__builtin_elementwise_min((__i32x4)__a, (__i32x4)__b);
1176 v128_t __b) {
1177 return (v128_t)__builtin_elementwise_min((__u32x4)__a, (__u32x4)__b);
1181 v128_t __b) {
1182 return (v128_t)__builtin_elementwise_max((__i32x4)__a, (__i32x4)__b);
1186 v128_t __b) {
1187 return (v128_t)__builtin_elementwise_max((__u32x4)__a, (__u32x4)__b);
1191 v128_t __b) {
1192 return (v128_t)__builtin_wasm_dot_s_i32x4_i16x8((__i16x8)__a, (__i16x8)__b);
1212 uint32_t __b) {
1213 return (v128_t)((__i64x2)__a << ((int64_t)__b & 0x3F));
1217 uint32_t __b) {
1218 return (v128_t)((__i64x2)__a >> ((int64_t)__b & 0x3F));
1222 uint32_t __b) {
1223 return (v128_t)((__u64x2)__a >> ((int64_t)__b & 0x3F));
1227 v128_t __b) {
1228 return (v128_t)((__u64x2)__a + (__u64x2)__b);
1232 v128_t __b) {
1233 return (v128_t)((__u64x2)__a - (__u64x2)__b);
1237 v128_t __b) {
1238 return (v128_t)((__u64x2)__a * (__u64x2)__b);
1270 v128_t __b) {
1271 return (v128_t)((__f32x4)__a + (__f32x4)__b);
1275 v128_t __b) {
1276 return (v128_t)((__f32x4)__a - (__f32x4)__b);
1280 v128_t __b) {
1281 return (v128_t)((__f32x4)__a * (__f32x4)__b);
1285 v128_t __b) {
1286 return (v128_t)((__f32x4)__a / (__f32x4)__b);
1290 v128_t __b) {
1291 return (v128_t)__builtin_wasm_min_f32x4((__f32x4)__a, (__f32x4)__b);
1295 v128_t __b) {
1296 return (v128_t)__builtin_wasm_max_f32x4((__f32x4)__a, (__f32x4)__b);
1300 v128_t __b) {
1301 return (v128_t)__builtin_wasm_pmin_f32x4((__f32x4)__a, (__f32x4)__b);
1305 v128_t __b) {
1306 return (v128_t)__builtin_wasm_pmax_f32x4((__f32x4)__a, (__f32x4)__b);
1338 v128_t __b) {
1339 return (v128_t)((__f64x2)__a + (__f64x2)__b);
1343 v128_t __b) {
1344 return (v128_t)((__f64x2)__a - (__f64x2)__b);
1348 v128_t __b) {
1349 return (v128_t)((__f64x2)__a * (__f64x2)__b);
1353 v128_t __b) {
1354 return (v128_t)((__f64x2)__a / (__f64x2)__b);
1358 v128_t __b) {
1359 return (v128_t)__builtin_wasm_min_f64x2((__f64x2)__a, (__f64x2)__b);
1363 v128_t __b) {
1364 return (v128_t)__builtin_wasm_max_f64x2((__f64x2)__a, (__f64x2)__b);
1368 v128_t __b) {
1369 return (v128_t)__builtin_wasm_pmin_f64x2((__f64x2)__a, (__f64x2)__b);
1373 v128_t __b) {
1374 return (v128_t)__builtin_wasm_pmax_f64x2((__f64x2)__a, (__f64x2)__b);
1430 #define wasm_i8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \
1434 (__i8x16)(__a), (__i8x16)(__b), __c0, __c1, __c2, __c3, __c4, __c5, \
1437 #define wasm_i16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \
1440 (__i8x16)(__a), (__i8x16)(__b), (__c0)*2, (__c0)*2 + 1, (__c1)*2, \
1445 #define wasm_i32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3) \
1447 (__i8x16)(__a), (__i8x16)(__b), (__c0)*4, (__c0)*4 + 1, (__c0)*4 + 2, \
1452 #define wasm_i64x2_shuffle(__a, __b, __c0, __c1) \
1454 (__i8x16)(__a), (__i8x16)(__b), (__c0)*8, (__c0)*8 + 1, (__c0)*8 + 2, \
1460 v128_t __b) {
1461 return (v128_t)__builtin_wasm_swizzle_i8x16((__i8x16)__a, (__i8x16)__b);
1465 wasm_i8x16_narrow_i16x8(v128_t __a, v128_t __b) {
1467 (__i16x8)__b);
1471 wasm_u8x16_narrow_i16x8(v128_t __a, v128_t __b) {
1473 (__i16x8)__b);
1477 wasm_i16x8_narrow_i32x4(v128_t __a, v128_t __b) {
1479 (__i32x4)__b);
1483 wasm_u16x8_narrow_i32x4(v128_t __a, v128_t __b) {
1485 (__i32x4)__b);
1601 wasm_i16x8_extmul_low_i8x16(v128_t __a, v128_t __b) {
1603 (__i16x8)wasm_i16x8_extend_low_i8x16(__b));
1607 wasm_i16x8_extmul_high_i8x16(v128_t __a, v128_t __b) {
1609 (__i16x8)wasm_i16x8_extend_high_i8x16(__b));
1613 wasm_u16x8_extmul_low_u8x16(v128_t __a, v128_t __b) {
1615 (__u16x8)wasm_u16x8_extend_low_u8x16(__b));
1619 wasm_u16x8_extmul_high_u8x16(v128_t __a, v128_t __b) {
1621 (__u16x8)wasm_u16x8_extend_high_u8x16(__b));
1625 wasm_i32x4_extmul_low_i16x8(v128_t __a, v128_t __b) {
1627 (__i32x4)wasm_i32x4_extend_low_i16x8(__b));
1631 wasm_i32x4_extmul_high_i16x8(v128_t __a, v128_t __b) {
1633 (__i32x4)wasm_i32x4_extend_high_i16x8(__b));
1637 wasm_u32x4_extmul_low_u16x8(v128_t __a, v128_t __b) {
1639 (__u32x4)wasm_u32x4_extend_low_u16x8(__b));
1643 wasm_u32x4_extmul_high_u16x8(v128_t __a, v128_t __b) {
1645 (__u32x4)wasm_u32x4_extend_high_u16x8(__b));
1649 wasm_i64x2_extmul_low_i32x4(v128_t __a, v128_t __b) {
1651 (__i64x2)wasm_i64x2_extend_low_i32x4(__b));
1655 wasm_i64x2_extmul_high_i32x4(v128_t __a, v128_t __b) {
1657 (__i64x2)wasm_i64x2_extend_high_i32x4(__b));
1661 wasm_u64x2_extmul_low_u32x4(v128_t __a, v128_t __b) {
1663 (__u64x2)wasm_u64x2_extend_low_u32x4(__b));
1667 wasm_u64x2_extmul_high_u32x4(v128_t __a, v128_t __b) {
1669 (__u64x2)wasm_u64x2_extend_high_u32x4(__b));
1673 v128_t __b) {
1674 return (v128_t)__builtin_wasm_q15mulr_sat_s_i16x8((__i16x8)__a, (__i16x8)__b);
1744 #define wasm_v8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \
1748 wasm_i8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7, \
1751 #define wasm_v16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \
1754 wasm_i16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7)
1756 #define wasm_v32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3) \
1758 wasm_i32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3)
1760 #define wasm_v64x2_shuffle(__a, __b, __c0, __c1) \
1762 wasm_i64x2_shuffle(__a, __b, __c0, __c1)
1771 wasm_f32x4_relaxed_madd(v128_t __a, v128_t __b, v128_t __c) {
1772 return (v128_t)__builtin_wasm_relaxed_madd_f32x4((__f32x4)__a, (__f32x4)__b,
1777 wasm_f32x4_relaxed_nmadd(v128_t __a, v128_t __b, v128_t __c) {
1778 return (v128_t)__builtin_wasm_relaxed_nmadd_f32x4((__f32x4)__a, (__f32x4)__b,
1783 wasm_f64x2_relaxed_madd(v128_t __a, v128_t __b, v128_t __c) {
1784 return (v128_t)__builtin_wasm_relaxed_madd_f64x2((__f64x2)__a, (__f64x2)__b,
1789 wasm_f64x2_relaxed_nmadd(v128_t __a, v128_t __b, v128_t __c) {
1790 return (v128_t)__builtin_wasm_relaxed_nmadd_f64x2((__f64x2)__a, (__f64x2)__b,
1795 wasm_i8x16_relaxed_laneselect(v128_t __a, v128_t __b, v128_t __m) {
1797 (__i8x16)__a, (__i8x16)__b, (__i8x16)__m);
1801 wasm_i16x8_relaxed_laneselect(v128_t __a, v128_t __b, v128_t __m) {
1803 (__i16x8)__a, (__i16x8)__b, (__i16x8)__m);
1807 wasm_i32x4_relaxed_laneselect(v128_t __a, v128_t __b, v128_t __m) {
1809 (__i32x4)__a, (__i32x4)__b, (__i32x4)__m);
1813 wasm_i64x2_relaxed_laneselect(v128_t __a, v128_t __b, v128_t __m) {
1815 (__i64x2)__a, (__i64x2)__b, (__i64x2)__m);
1825 v128_t __b) {
1826 return (v128_t)__builtin_wasm_relaxed_min_f32x4((__f32x4)__a, (__f32x4)__b);
1830 v128_t __b) {
1831 return (v128_t)__builtin_wasm_relaxed_max_f32x4((__f32x4)__a, (__f32x4)__b);
1835 v128_t __b) {
1836 return (v128_t)__builtin_wasm_relaxed_min_f64x2((__f64x2)__a, (__f64x2)__b);
1840 v128_t __b) {
1841 return (v128_t)__builtin_wasm_relaxed_max_f64x2((__f64x2)__a, (__f64x2)__b);
1865 wasm_i16x8_relaxed_q15mulr(v128_t __a, v128_t __b) {
1867 (__i16x8)__b);
1871 wasm_i16x8_relaxed_dot_i8x16_i7x16(v128_t __a, v128_t __b) {
1873 (__i8x16)__b);
1877 wasm_i32x4_relaxed_dot_i8x16_i7x16_add(v128_t __a, v128_t __b, v128_t __c) {
1879 (__i8x16)__a, (__i8x16)__b, (__i32x4)__c);
1898 #define wasm_f16x8_replace_lane(__a, __i, __b) \
1899 ((v128_t)__builtin_wasm_replace_lane_f16x8((__f16x8)(__a), __i, __b))
1931 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_eq(v128_t __a, v128_t __b) {
1932 return (v128_t)((__f16x8)__a == (__f16x8)__b);
1935 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_ne(v128_t __a, v128_t __b) {
1936 return (v128_t)((__f16x8)__a != (__f16x8)__b);
1939 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_lt(v128_t __a, v128_t __b) {
1940 return (v128_t)((__f16x8)__a < (__f16x8)__b);
1943 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_gt(v128_t __a, v128_t __b) {
1944 return (v128_t)((__f16x8)__a > (__f16x8)__b);
1947 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_le(v128_t __a, v128_t __b) {
1948 return (v128_t)((__f16x8)__a <= (__f16x8)__b);
1951 static __inline__ v128_t __FP16_FN_ATTRS wasm_f16x8_ge(v128_t __a, v128_t __b) {
1952 return (v128_t)((__f16x8)__a >= (__f16x8)__b);
1956 v128_t __b) {
1957 return (v128_t)((__f16x8)__a + (__f16x8)__b);
1961 v128_t __b) {
1962 return (v128_t)((__f16x8)__a - (__f16x8)__b);
1966 v128_t __b) {
1967 return (v128_t)((__f16x8)__a * (__f16x8)__b);
1971 v128_t __b) {
1972 return (v128_t)((__f16x8)__a / (__f16x8)__b);
1976 v128_t __b) {
1977 return (v128_t)__builtin_wasm_min_f16x8((__f16x8)__a, (__f16x8)__b);
1981 v128_t __b) {
1982 return (v128_t)__builtin_wasm_max_f16x8((__f16x8)__a, (__f16x8)__b);
1986 v128_t __b) {
1987 return (v128_t)__builtin_wasm_pmin_f16x8((__f16x8)__a, (__f16x8)__b);
1991 v128_t __b) {
1992 return (v128_t)__builtin_wasm_pmax_f16x8((__f16x8)__a, (__f16x8)__b);
2014 v128_t __b,
2016 return (v128_t)__builtin_wasm_relaxed_madd_f16x8((__f16x8)__a, (__f16x8)__b,
2021 v128_t __b,
2023 return (v128_t)__builtin_wasm_relaxed_nmadd_f16x8((__f16x8)__a, (__f16x8)__b,
2030 wasm_v8x16_swizzle(v128_t __a, v128_t __b) {
2031 return wasm_i8x16_swizzle(__a, __b);
2050 wasm_i8x16_add_saturate(v128_t __a, v128_t __b) {
2051 return wasm_i8x16_add_sat(__a, __b);
2055 wasm_u8x16_add_saturate(v128_t __a, v128_t __b) {
2056 return wasm_u8x16_add_sat(__a, __b);
2060 wasm_i8x16_sub_saturate(v128_t __a, v128_t __b) {
2061 return wasm_i8x16_sub_sat(__a, __b);
2065 wasm_u8x16_sub_saturate(v128_t __a, v128_t __b) {
2066 return wasm_u8x16_sub_sat(__a, __b);
2070 wasm_i16x8_add_saturate(v128_t __a, v128_t __b) {
2071 return wasm_i16x8_add_sat(__a, __b);
2075 wasm_u16x8_add_saturate(v128_t __a, v128_t __b) {
2076 return wasm_u16x8_add_sat(__a, __b);
2080 wasm_i16x8_sub_saturate(v128_t __a, v128_t __b) {
2081 return wasm_i16x8_sub_sat(__a, __b);
2085 wasm_u16x8_sub_saturate(v128_t __a, v128_t __b) {
2086 return wasm_u16x8_sub_sat(__a, __b);