Lines Matching defs:__b
508 int8_t __b) in wasm_i8x16_replace_lane()
517 uint8_t __b) in wasm_u8x16_replace_lane()
545 int16_t __b) in wasm_i16x8_replace_lane()
553 v128_t __a, int __i, uint16_t __b) __REQUIRE_CONSTANT(__i) { in wasm_u16x8_replace_lane()
580 int32_t __b) in wasm_i32x4_replace_lane()
588 v128_t __a, int __i, uint32_t __b) __REQUIRE_CONSTANT(__i) { in wasm_u32x4_replace_lane()
615 int64_t __b) in wasm_i64x2_replace_lane()
623 v128_t __a, int __i, uint64_t __b) __REQUIRE_CONSTANT(__i) { in wasm_u64x2_replace_lane()
641 float __b) in wasm_f32x4_replace_lane()
660 double __b) in wasm_f64x2_replace_lane()
668 v128_t __b) { in wasm_i8x16_eq()
673 v128_t __b) { in wasm_i8x16_ne()
678 v128_t __b) { in wasm_i8x16_lt()
683 v128_t __b) { in wasm_u8x16_lt()
688 v128_t __b) { in wasm_i8x16_gt()
693 v128_t __b) { in wasm_u8x16_gt()
698 v128_t __b) { in wasm_i8x16_le()
703 v128_t __b) { in wasm_u8x16_le()
708 v128_t __b) { in wasm_i8x16_ge()
713 v128_t __b) { in wasm_u8x16_ge()
718 v128_t __b) { in wasm_i16x8_eq()
723 v128_t __b) { in wasm_i16x8_ne()
728 v128_t __b) { in wasm_i16x8_lt()
733 v128_t __b) { in wasm_u16x8_lt()
738 v128_t __b) { in wasm_i16x8_gt()
743 v128_t __b) { in wasm_u16x8_gt()
748 v128_t __b) { in wasm_i16x8_le()
753 v128_t __b) { in wasm_u16x8_le()
758 v128_t __b) { in wasm_i16x8_ge()
763 v128_t __b) { in wasm_u16x8_ge()
768 v128_t __b) { in wasm_i32x4_eq()
773 v128_t __b) { in wasm_i32x4_ne()
778 v128_t __b) { in wasm_i32x4_lt()
783 v128_t __b) { in wasm_u32x4_lt()
788 v128_t __b) { in wasm_i32x4_gt()
793 v128_t __b) { in wasm_u32x4_gt()
798 v128_t __b) { in wasm_i32x4_le()
803 v128_t __b) { in wasm_u32x4_le()
808 v128_t __b) { in wasm_i32x4_ge()
813 v128_t __b) { in wasm_u32x4_ge()
818 v128_t __b) { in wasm_i64x2_eq()
823 v128_t __b) { in wasm_i64x2_ne()
828 v128_t __b) { in wasm_i64x2_lt()
833 v128_t __b) { in wasm_i64x2_gt()
838 v128_t __b) { in wasm_i64x2_le()
843 v128_t __b) { in wasm_i64x2_ge()
848 v128_t __b) { in wasm_f32x4_eq()
853 v128_t __b) { in wasm_f32x4_ne()
858 v128_t __b) { in wasm_f32x4_lt()
863 v128_t __b) { in wasm_f32x4_gt()
868 v128_t __b) { in wasm_f32x4_le()
873 v128_t __b) { in wasm_f32x4_ge()
878 v128_t __b) { in wasm_f64x2_eq()
883 v128_t __b) { in wasm_f64x2_ne()
888 v128_t __b) { in wasm_f64x2_lt()
893 v128_t __b) { in wasm_f64x2_gt()
898 v128_t __b) { in wasm_f64x2_le()
903 v128_t __b) { in wasm_f64x2_ge()
912 v128_t __b) { in wasm_v128_and()
917 v128_t __b) { in wasm_v128_or()
922 v128_t __b) { in wasm_v128_xor()
927 v128_t __b) { in wasm_v128_andnot()
936 v128_t __b, in wasm_v128_bitselect()
963 uint32_t __b) { in wasm_i8x16_shl()
968 uint32_t __b) { in wasm_i8x16_shr()
973 uint32_t __b) { in wasm_u8x16_shr()
978 v128_t __b) { in wasm_i8x16_add()
983 v128_t __b) { in wasm_i8x16_add_sat()
988 v128_t __b) { in wasm_u8x16_add_sat()
993 v128_t __b) { in wasm_i8x16_sub()
998 v128_t __b) { in wasm_i8x16_sub_sat()
1003 v128_t __b) { in wasm_u8x16_sub_sat()
1008 v128_t __b) { in wasm_i8x16_min()
1013 v128_t __b) { in wasm_u8x16_min()
1018 v128_t __b) { in wasm_i8x16_max()
1023 v128_t __b) { in wasm_u8x16_max()
1028 v128_t __b) { in wasm_u8x16_avgr()
1049 uint32_t __b) { in wasm_i16x8_shl()
1054 uint32_t __b) { in wasm_i16x8_shr()
1059 uint32_t __b) { in wasm_u16x8_shr()
1064 v128_t __b) { in wasm_i16x8_add()
1069 v128_t __b) { in wasm_i16x8_add_sat()
1074 v128_t __b) { in wasm_u16x8_add_sat()
1079 v128_t __b) { in wasm_i16x8_sub()
1084 v128_t __b) { in wasm_i16x8_sub_sat()
1089 v128_t __b) { in wasm_u16x8_sub_sat()
1094 v128_t __b) { in wasm_i16x8_mul()
1099 v128_t __b) { in wasm_i16x8_min()
1104 v128_t __b) { in wasm_u16x8_min()
1109 v128_t __b) { in wasm_i16x8_max()
1114 v128_t __b) { in wasm_u16x8_max()
1119 v128_t __b) { in wasm_u16x8_avgr()
1140 uint32_t __b) { in wasm_i32x4_shl()
1145 uint32_t __b) { in wasm_i32x4_shr()
1150 uint32_t __b) { in wasm_u32x4_shr()
1155 v128_t __b) { in wasm_i32x4_add()
1160 v128_t __b) { in wasm_i32x4_sub()
1165 v128_t __b) { in wasm_i32x4_mul()
1170 v128_t __b) { in wasm_i32x4_min()
1175 v128_t __b) { in wasm_u32x4_min()
1180 v128_t __b) { in wasm_i32x4_max()
1185 v128_t __b) { in wasm_u32x4_max()
1190 v128_t __b) { in wasm_i32x4_dot_i16x8()
1211 uint32_t __b) { in wasm_i64x2_shl()
1216 uint32_t __b) { in wasm_i64x2_shr()
1221 uint32_t __b) { in wasm_u64x2_shr()
1226 v128_t __b) { in wasm_i64x2_add()
1231 v128_t __b) { in wasm_i64x2_sub()
1236 v128_t __b) { in wasm_i64x2_mul()
1269 v128_t __b) { in wasm_f32x4_add()
1274 v128_t __b) { in wasm_f32x4_sub()
1279 v128_t __b) { in wasm_f32x4_mul()
1284 v128_t __b) { in wasm_f32x4_div()
1289 v128_t __b) { in wasm_f32x4_min()
1294 v128_t __b) { in wasm_f32x4_max()
1299 v128_t __b) { in wasm_f32x4_pmin()
1304 v128_t __b) { in wasm_f32x4_pmax()
1337 v128_t __b) { in wasm_f64x2_add()
1342 v128_t __b) { in wasm_f64x2_sub()
1347 v128_t __b) { in wasm_f64x2_mul()
1352 v128_t __b) { in wasm_f64x2_div()
1357 v128_t __b) { in wasm_f64x2_min()
1362 v128_t __b) { in wasm_f64x2_max()
1367 v128_t __b) { in wasm_f64x2_pmin()
1372 v128_t __b) { in wasm_f64x2_pmax()
1429 #define wasm_i8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ argument
1436 #define wasm_i16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ argument
1444 #define wasm_i32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3) \ argument
1451 #define wasm_i64x2_shuffle(__a, __b, __c0, __c1) \ argument
1459 v128_t __b) { in wasm_i8x16_swizzle()
1464 wasm_i8x16_narrow_i16x8(v128_t __a, v128_t __b) { in wasm_i8x16_narrow_i16x8()
1470 wasm_u8x16_narrow_i16x8(v128_t __a, v128_t __b) { in wasm_u8x16_narrow_i16x8()
1476 wasm_i16x8_narrow_i32x4(v128_t __a, v128_t __b) { in wasm_i16x8_narrow_i32x4()
1482 wasm_u16x8_narrow_i32x4(v128_t __a, v128_t __b) { in wasm_u16x8_narrow_i32x4()
1600 wasm_i16x8_extmul_low_i8x16(v128_t __a, v128_t __b) { in wasm_i16x8_extmul_low_i8x16()
1606 wasm_i16x8_extmul_high_i8x16(v128_t __a, v128_t __b) { in wasm_i16x8_extmul_high_i8x16()
1612 wasm_u16x8_extmul_low_u8x16(v128_t __a, v128_t __b) { in wasm_u16x8_extmul_low_u8x16()
1618 wasm_u16x8_extmul_high_u8x16(v128_t __a, v128_t __b) { in wasm_u16x8_extmul_high_u8x16()
1624 wasm_i32x4_extmul_low_i16x8(v128_t __a, v128_t __b) { in wasm_i32x4_extmul_low_i16x8()
1630 wasm_i32x4_extmul_high_i16x8(v128_t __a, v128_t __b) { in wasm_i32x4_extmul_high_i16x8()
1636 wasm_u32x4_extmul_low_u16x8(v128_t __a, v128_t __b) { in wasm_u32x4_extmul_low_u16x8()
1642 wasm_u32x4_extmul_high_u16x8(v128_t __a, v128_t __b) { in wasm_u32x4_extmul_high_u16x8()
1648 wasm_i64x2_extmul_low_i32x4(v128_t __a, v128_t __b) { in wasm_i64x2_extmul_low_i32x4()
1654 wasm_i64x2_extmul_high_i32x4(v128_t __a, v128_t __b) { in wasm_i64x2_extmul_high_i32x4()
1660 wasm_u64x2_extmul_low_u32x4(v128_t __a, v128_t __b) { in wasm_u64x2_extmul_low_u32x4()
1666 wasm_u64x2_extmul_high_u32x4(v128_t __a, v128_t __b) { in wasm_u64x2_extmul_high_u32x4()
1672 v128_t __b) { in wasm_i16x8_q15mulr_sat()
1743 #define wasm_v8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ argument
1750 #define wasm_v16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ argument
1755 #define wasm_v32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3) \ argument
1759 #define wasm_v64x2_shuffle(__a, __b, __c0, __c1) \ argument
1764 wasm_v8x16_swizzle(v128_t __a, v128_t __b) { in wasm_v8x16_swizzle()
1784 wasm_i8x16_add_saturate(v128_t __a, v128_t __b) { in wasm_i8x16_add_saturate()
1789 wasm_u8x16_add_saturate(v128_t __a, v128_t __b) { in wasm_u8x16_add_saturate()
1794 wasm_i8x16_sub_saturate(v128_t __a, v128_t __b) { in wasm_i8x16_sub_saturate()
1799 wasm_u8x16_sub_saturate(v128_t __a, v128_t __b) { in wasm_u8x16_sub_saturate()
1804 wasm_i16x8_add_saturate(v128_t __a, v128_t __b) { in wasm_i16x8_add_saturate()
1809 wasm_u16x8_add_saturate(v128_t __a, v128_t __b) { in wasm_u16x8_add_saturate()
1814 wasm_i16x8_sub_saturate(v128_t __a, v128_t __b) { in wasm_i16x8_sub_saturate()
1819 wasm_u16x8_sub_saturate(v128_t __a, v128_t __b) { in wasm_u16x8_sub_saturate()