/freebsd-src/contrib/llvm-project/clang/lib/Headers/ |
H A D | velintrin_approx.h | 12 static inline __vr _vel_approx_vfdivs_vvvl(__vr v0, __vr v1, int l) { in _vel_approx_vfdivs_vvvl() argument 19 v2 = _vel_vfmuls_vvvl(v0, v3, l); in _vel_approx_vfdivs_vvvl() 20 v4 = _vel_vfnmsbs_vvvvl(v0, v2, v1, l); in _vel_approx_vfdivs_vvvl() 22 v0 = _vel_vfnmsbs_vvvvl(v0, v2, v1, l); in _vel_approx_vfdivs_vvvl() 23 v0 = _vel_vfmads_vvvvl(v2, v3, v0, l); in _vel_approx_vfdivs_vvvl() 24 return v0; in _vel_approx_vfdivs_vvvl() 27 static inline __vr _vel_approx_pvfdiv_vvvl(__vr v0, __vr v1, int l) { in _vel_approx_pvfdiv_vvvl() argument 34 v2 = _vel_pvfmul_vvvl(v0, v3, l); in _vel_approx_pvfdiv_vvvl() 35 v4 = _vel_pvfnmsb_vvvvl(v0, v2, v1, l); in _vel_approx_pvfdiv_vvvl() 37 v0 = _vel_pvfnmsb_vvvvl(v0, v2, v1, l); in _vel_approx_pvfdiv_vvvl() [all …]
|
/freebsd-src/sys/crypto/openssl/aarch64/ |
H A D | aesv8-armx.S | 38 eor v0.16b,v0.16b,v0.16b 50 ext v5.16b,v0.16b,v3.16b,#12 52 aese v6.16b,v0.16b 56 ext v5.16b,v0.16b,v5.16b,#12 58 ext v5.16b,v0.16b,v5.16b,#12 68 ext v5.16b,v0.16b,v3.16b,#12 70 aese v6.16b,v0.16b 73 ext v5.16b,v0.16b,v5.16b,#12 75 ext v5.16b,v0.16b,v5.16b,#12 82 ext v5.16b,v0.16b,v3.16b,#12 [all …]
|
H A D | ghashv8-armx.S | 31 pmull v0.1q,v20.1d,v20.1d 36 ext v17.16b,v0.16b,v2.16b,#8 //Karatsuba post-processing 37 eor v18.16b,v0.16b,v2.16b 40 pmull v18.1q,v0.1d,v19.1d //1st phase 43 ins v1.d[1],v0.d[0] 44 eor v0.16b,v1.16b,v18.16b 46 ext v18.16b,v0.16b,v0.16b,#8 //2nd phase 47 pmull v0.1q,v0.1d,v19.1d 49 eor v22.16b,v0.16b,v18.16b 56 pmull v0.1q,v20.1d, v22.1d [all …]
|
H A D | vpaes-armv8.S | 137 ushr v0.16b, v7.16b, #4 // vpsrlb $4, %xmm0, %xmm0 140 tbl v2.16b, {v21.16b}, v0.16b // vpshufb %xmm0, %xmm3, %xmm2 141 eor v0.16b, v1.16b, v16.16b // vpxor %xmm5, %xmm1, %xmm0 142 eor v0.16b, v0.16b, v2.16b // vpxor %xmm2, %xmm0, %xmm0 151 tbl v0.16b, {v24.16b}, v3.16b // vpshufb %xmm3, %xmm12, %xmm0 # 0 = sb1t 154 eor v0.16b, v0.16b, v4.16b // vpxor %xmm4, %xmm0, %xmm0 # 0 = A 157 tbl v3.16b, {v0.16b}, v1.16b // vpshufb %xmm1, %xmm0, %xmm3 # 0 = B 159 tbl v0.16b, {v0.16b}, v4.16b // vpshufb %xmm4, %xmm0, %xmm0 # 3 = D 162 eor v0.16b, v0.16b, v3.16b // vpxor %xmm3, %xmm0, %xmm0 # 3 = 2A+B+D 164 eor v0.16b, v0.16b, v4.16b // vpxor %xmm4, %xmm0, %xmm0 # 0 = 2A+3B+C+D [all …]
|
H A D | aes-gcm-armv8_64.S | 60 …ld1 { v0.16b}, [x16] //special case vector load initial counter so we … 93 aese v0.16b, v18.16b 94 aesmc v0.16b, v0.16b //AES block 0 - round 0 105 aese v0.16b, v19.16b 106 aesmc v0.16b, v0.16b //AES block 0 - round 1 117 aese v0.16b, v20.16b 118 aesmc v0.16b, v0.16b //AES block 0 - round 2 135 aese v0.16b, v21.16b 136 aesmc v0.16b, v0.16b //AES block 0 - round 3 159 aese v0.16b, v22.16b [all …]
|
/freebsd-src/contrib/file/tests/ |
H A D | Makefile.am | 147 zstd-v0.2-FF.result \ 148 zstd-v0.2-FF.testfile \ 149 zstd-v0.3-FF.result \ 150 zstd-v0.3-FF.testfile \ 151 zstd-v0.4-FF.result \ 152 zstd-v0.4-FF.testfile \ 153 zstd-v0.5-FF.result \ 154 zstd-v0.5-FF.testfile \ 155 zstd-v0.6-FF.result \ 156 zstd-v0 [all...] |
/freebsd-src/crypto/openssl/crypto/aes/asm/ |
H A D | vpaes-ppc.pl | 246 vsrb v1, v0, v8 # vpsrlb \$4, %xmm0, %xmm0 247 vperm v0, $iptlo, $iptlo, v0 # vpshufb %xmm1, %xmm2, %xmm1 249 vxor v0, v0, v5 # vpxor %xmm5, %xmm1, %xmm0 250 vxor v0, v0, v1 # vpxor %xmm2, %xmm0, %xmm0 260 vperm v0, $sb1u, v7, v3 # vpshufb %xmm3, %xmm12, %xmm0 # 0 = sb1t 264 vxor v0, v0, v4 # vpxor %xmm4, %xmm0, %xmm0 # 0 = A 268 vperm v3, v0, v7, v1 # vpshufb %xmm1, %xmm0, %xmm3 # 0 = B 270 vperm v0, v0, v7, v4 # vpshufb %xmm4, %xmm0, %xmm0 # 3 = D 273 vxor v0, v0, v3 # vpxor %xmm3, %xmm0, %xmm0 # 3 = 2A+B+D 274 vxor v0, v0, v4 # vpxor %xmm4, %xmm0, %xmm0 # 0 = 2A+3B+C+D [all …]
|
H A D | vpaes-armv8.pl | 200 ushr v0.16b, v7.16b, #4 // vpsrlb \$4, %xmm0, %xmm0 203 tbl v2.16b, {$ipthi}, v0.16b // vpshufb %xmm0, %xmm3, %xmm2 204 eor v0.16b, v1.16b, v16.16b // vpxor %xmm5, %xmm1, %xmm0 205 eor v0.16b, v0.16b, v2.16b // vpxor %xmm2, %xmm0, %xmm0 214 tbl v0.16b, {$sb1u}, v3.16b // vpshufb %xmm3, %xmm12, %xmm0 # 0 = sb1t 217 eor v0.16b, v0.16b, v4.16b // vpxor %xmm4, %xmm0, %xmm0 # 0 = A 220 tbl v3.16b, {v0.16b}, v1.16b // vpshufb %xmm1, %xmm0, %xmm3 # 0 = B 222 tbl v0.16b, {v0.16b}, v4.16b // vpshufb %xmm4, %xmm0, %xmm0 # 3 = D 225 eor v0.16b, v0.16b, v3.16b // vpxor %xmm3, %xmm0, %xmm0 # 3 = 2A+B+D 227 eor v0.16b, v0.16b, v4.16b // vpxor %xmm4, %xmm0, %xmm0 # 0 = 2A+3B+C+D [all …]
|
/freebsd-src/sys/contrib/openzfs/module/icp/asm-aarch64/blake3/ |
H A D | b3_aarch64_sse2.S | 81 eor v0.16b, v2.16b, v0.16b 114 and v0.8b, v1.8b, v0.8b 116 mov v3.d[1], v0.d[0] 119 uzp1 v1.4s, v0.4s, v6.4s 120 uzp2 v0.4s, v0.4s, v6.4s 125 add v2.4s, v2.4s, v0.4s 167 uzp2 v17.4s, v17.4s, v0.4s 179 zip1 v18.2d, v16.2d, v0.2d 180 zip2 v0.4s, v0.4s, v16.4s 184 zip1 v16.4s, v0.4s, v6.4s [all …]
|
H A D | b3_aarch64_sse41.S | 81 eor v0.16b, v2.16b, v0.16b 147 and v0.8b, v1.8b, v0.8b 150 mov v1.d[1], v0.d[0] 154 add v0.4s, v2.4s, v3.4s 156 add v16.4s, v0.4s, v5.4s 161 tbl v1.16b, { v1.16b }, v0.16b 189 tbl v7.16b, { v7.16b }, v0.16b 215 tbl v7.16b, { v7.16b }, v0.16b 246 tbl v7.16b, { v7.16b }, v0.16b 275 tbl v7.16b, { v7.16b }, v0.16b [all …]
|
/freebsd-src/sys/sys/ |
H A D | ktr.h | 143 #define KTR_EVENT1(m, egroup, ident, etype, edat, a0, v0) \ argument 144 CTR3(m, KTR_EFMT(egroup, ident, etype) a0, ident, edat, (v0)) 145 #define KTR_EVENT2(m, egroup, ident, etype, edat, a0, v0, a1, v1) \ argument 147 ident, edat, (v0), (v1)) 148 #define KTR_EVENT3(m, egroup, ident, etype, edat, a0, v0, a1, v1, a2, v2)\ argument 150 ident, edat, (v0), (v1), (v2)) 152 a0, v0, a1, v1, a2, v2, a3, v3) \ argument 154 ident, edat, (v0), (v1), (v2), (v3)) 161 #define KTR_STATE1(m, egroup, ident, state, a0, v0) \ argument 162 KTR_EVENT1(m, egroup, ident, "state:\"%s\"", state, a0, (v0)) [all …]
|
/freebsd-src/contrib/bearssl/src/rsa/ |
H A D | rsa_i15_privexp.c | 65 uint32_t r, a, b, u0, v0, u1, v1, he, hr; in br_rsa_i15_compute_privexp() local 161 * u0, u1, v0 and v1. Initial values are: in br_rsa_i15_compute_privexp() 162 * a = e u0 = 1 v0 = 0 in br_rsa_i15_compute_privexp() 165 * a = u0*e - v0*r in br_rsa_i15_compute_privexp() 170 * 0 <= v0 <= e in br_rsa_i15_compute_privexp() 175 * adjust u0, u1, v0 and v1 to maintain the invariants: in br_rsa_i15_compute_privexp() 182 * key or public exponent is not valid). The (u0,v0) or (u1,v1) in br_rsa_i15_compute_privexp() 191 * - When a is divided by 2, u0 and v0 must be divided by 2. in br_rsa_i15_compute_privexp() 194 * u0 and v0, respectively. in br_rsa_i15_compute_privexp() 195 * - When a is subtracted from b, u0 and v0 are subtracted from in br_rsa_i15_compute_privexp() [all …]
|
H A D | rsa_i31_privexp.c | 65 uint32_t r, a, b, u0, v0, u1, v1, he, hr; in br_rsa_i31_compute_privexp() local 161 * u0, u1, v0 and v1. Initial values are: in br_rsa_i31_compute_privexp() 162 * a = e u0 = 1 v0 = 0 in br_rsa_i31_compute_privexp() 165 * a = u0*e - v0*r in br_rsa_i31_compute_privexp() 170 * 0 <= v0 <= e in br_rsa_i31_compute_privexp() 175 * adjust u0, u1, v0 and v1 to maintain the invariants: in br_rsa_i31_compute_privexp() 182 * key or public exponent is not valid). The (u0,v0) or (u1,v1) in br_rsa_i31_compute_privexp() 191 * - When a is divided by 2, u0 and v0 must be divided by 2. in br_rsa_i31_compute_privexp() 194 * u0 and v0, respectively. in br_rsa_i31_compute_privexp() 195 * - When a is subtracted from b, u0 and v0 are subtracted from in br_rsa_i31_compute_privexp() [all …]
|
/freebsd-src/crypto/openssl/crypto/des/ |
H A D | cfb64ede.c | 29 register DES_LONG v0, v1; in DES_ede3_cfb64_encrypt() local 39 c2l(iv, v0); in DES_ede3_cfb64_encrypt() 42 ti[0] = v0; in DES_ede3_cfb64_encrypt() 45 v0 = ti[0]; in DES_ede3_cfb64_encrypt() 49 l2c(v0, iv); in DES_ede3_cfb64_encrypt() 61 c2l(iv, v0); in DES_ede3_cfb64_encrypt() 64 ti[0] = v0; in DES_ede3_cfb64_encrypt() 67 v0 = ti[0]; in DES_ede3_cfb64_encrypt() 71 l2c(v0, iv); in DES_ede3_cfb64_encrypt() 82 v0 = v1 = ti[0] = ti[1] = c = cc = 0; in DES_ede3_cfb64_encrypt() [all …]
|
H A D | cfb64enc.c | 28 register DES_LONG v0, v1; in DES_cfb64_encrypt() local 38 c2l(iv, v0); in DES_cfb64_encrypt() 39 ti[0] = v0; in DES_cfb64_encrypt() 44 v0 = ti[0]; in DES_cfb64_encrypt() 45 l2c(v0, iv); in DES_cfb64_encrypt() 46 v0 = ti[1]; in DES_cfb64_encrypt() 47 l2c(v0, iv); in DES_cfb64_encrypt() 58 c2l(iv, v0); in DES_cfb64_encrypt() 59 ti[0] = v0; in DES_cfb64_encrypt() 64 v0 = ti[0]; in DES_cfb64_encrypt() [all …]
|
H A D | cfb_enc.c | 34 register DES_LONG d0, d1, v0, v1; in DES_cfb_encrypt() local 55 c2l(iv, v0); in DES_cfb_encrypt() 60 ti[0] = v0; in DES_cfb_encrypt() 74 v0 = v1; in DES_cfb_encrypt() 77 v0 = d0; in DES_cfb_encrypt() 82 l2c(v0, iv); in DES_cfb_encrypt() 87 sh[0] = v0, sh[1] = v1, sh[2] = d0, sh[3] = d1; in DES_cfb_encrypt() 96 v0 = sh[0], v1 = sh[1]; in DES_cfb_encrypt() 99 c2l(iv, v0); in DES_cfb_encrypt() 107 ti[0] = v0; in DES_cfb_encrypt() [all …]
|
H A D | ofb_enc.c | 28 register DES_LONG d0, d1, vv0, vv1, v0, v1, n = (numbits + 7) / 8; in DES_ofb_encrypt() local 52 c2l(iv, v0); in DES_ofb_encrypt() 54 ti[0] = v0; in DES_ofb_encrypt() 57 ti[0] = v0; in DES_ofb_encrypt() 70 v0 = v1; in DES_ofb_encrypt() 73 v0 = vv0; in DES_ofb_encrypt() 76 v0 = ((v1 >> (num - 32)) | (vv0 << (64 - num))) & 0xffffffffL; in DES_ofb_encrypt() 80 v0 = ((v0 >> num) | (v1 << (32 - num))) & 0xffffffffL; in DES_ofb_encrypt() 85 l2c(v0, iv); in DES_ofb_encrypt() 87 v0 = v1 = d0 = d1 = ti[0] = ti[1] = vv0 = vv1 = 0; in DES_ofb_encrypt()
|
/freebsd-src/contrib/xz/src/liblzma/check/ |
H A D | crc_x86_clmul.h | 75 crc_simd_body(const uint8_t *buf, const size_t size, __m128i *v0, __m128i *v1, in crc_simd_body() argument 137 // and store the result in v1 (high bytes) and v0 (low bytes). in crc_simd_body() 143 // [ v1 ][ v0 ] in crc_simd_body() 147 // [ initial_crc ] size == 16 (all in v0) in crc_simd_body() 150 MASK_LH(initial_crc, mask_low, *v0, *v1); in crc_simd_body() 158 // (part of) initial_crc is in v0. in crc_simd_body() 167 // low bits of initial_crc in v0. Note that the in crc_simd_body() 172 *v0 = _mm_xor_si128(*v0, v2); in crc_simd_body() 175 *v0 in crc_simd_body() 267 __m128i v0, v1, v2; crc32_arch_optimized() local 355 __m128i v0, v1, v2; crc64_arch_optimized() local [all...] |
/freebsd-src/sys/libkern/arm/ |
H A D | muldi3.c | 50 * v = 2^n v1 * v0 54 * uv = 2^2n u1 v1 + 2^n u1 v0 + 2^n v1 u0 + u0 v0 55 * = 2^2n u1 v1 + 2^n (u1 v0 + v1 u0) + u0 v0 58 * and add 2^n u0 v0 to the last term and subtract it from the middle. 62 * (2^n) (u1 v0 - u1 v1 + u0 v1 - u0 v0) + 63 * (2^n + 1) (u0 v0) 68 * (2^n) (u1 - u0) (v0 - v1) + [(u1-u0)... = mid] 69 * (2^n + 1) (u0 v0) [u0v0 = low] 71 * The terms (u1 v1), (u1 - u0) (v0 - v1), and (u0 v0) can all be done 73 * of (u1 - u0) or (v0 - v1) may be negative.) [all …]
|
/freebsd-src/lib/libc/quad/ |
H A D | muldi3.c | 48 * v = 2^n v1 * v0 52 * uv = 2^2n u1 v1 + 2^n u1 v0 + 2^n v1 u0 + u0 v0 53 * = 2^2n u1 v1 + 2^n (u1 v0 + v1 u0) + u0 v0 56 * and add 2^n u0 v0 to the last term and subtract it from the middle. 60 * (2^n) (u1 v0 - u1 v1 + u0 v1 - u0 v0) + 61 * (2^n + 1) (u0 v0) 66 * (2^n) (u1 - u0) (v0 - v1) + [(u1-u0)... = mid] 67 * (2^n + 1) (u0 v0) [u0v0 = low] 69 * The terms (u1 v1), (u1 - u0) (v0 - v1), and (u0 v0) can all be done 71 * of (u1 - u0) or (v0 - v1) may be negative.) [all …]
|
/freebsd-src/contrib/libucl/klib/ |
H A D | kvec.h | 87 #define kv_copy_safe(type, v1, v0, el) do { \ argument 88 if ((v1).m < (v0).n) kv_resize_safe(type, v1, (v0).n, el); \ 89 (v1).n = (v0).n; \ 90 memcpy((v1).a, (v0).a, sizeof(type) * (v0).n); \ 109 #define kv_concat_safe(type, v1, v0, el) do { \ argument 110 if ((v1).m < (v0).n + (v1).n) \ 111 kv_resize_safe(type, v1, (v0).n + (v1).n, el); \ 112 memcpy((v1).a + (v1).n, (v0).a, sizeof(type) * (v0).n); \ 113 (v1).n = (v0).n + (v1).n; \ 133 #define kv_copy(type, v1, v0) do { \ argument [all …]
|
/freebsd-src/contrib/bearssl/src/symcipher/ |
H A D | poly1305_ctmulq.c | 140 uint64_t v0, v1, v2; in poly1305_inner_big() local 143 v0 = br_dec64le(buf + 0); in poly1305_inner_big() 146 v1 = ((v0 >> 44) | (v1 << 20)) & MASK44; in poly1305_inner_big() 147 v0 &= MASK44; in poly1305_inner_big() 148 a0 += v0; in poly1305_inner_big() 158 v0 = br_dec64le(buf + 16); in poly1305_inner_big() 161 v1 = ((v0 >> 44) | (v1 << 20)) & MASK44; in poly1305_inner_big() 162 v0 &= MASK44; in poly1305_inner_big() 163 a0 += v0; in poly1305_inner_big() 173 v0 = br_dec64le(buf + 32); in poly1305_inner_big() [all …]
|
/freebsd-src/contrib/llvm-project/llvm/lib/Target/RISCV/ |
H A D | RISCVInstrInfoVVLPatterns.td | 642 (mask_type V0), 651 (mask_type V0), GPR:$vl, log2sew, TAIL_AGNOSTIC)>; 669 (mask_type V0), 676 (mask_type V0), 698 (mask_type V0), 707 (mask_type V0), 761 (mask_type V0), 766 (mask_type V0), GPR:$vl, sew, TU_MU)>; 830 (mask_type V0), 839 (mask_type V0), GP [all...] |
/freebsd-src/contrib/unbound/util/ |
H A D | siphash.c | 58 v0 += v1; \ 60 v1 ^= v0; \ 61 v0 = ROTL(v0, 32); \ 65 v0 += v3; \ 67 v3 ^= v0; \ 77 printf("(%3d) v0 %08x %08x\n", (int)inlen, (uint32_t)(v0 >> 32), \ 78 (uint32_t)v0); \ 93 uint64_t v0 in siphash() local [all...] |
/freebsd-src/crypto/openssl/crypto/siphash/ |
H A D | siphash.c | 52 v0 += v1; \ 54 v1 ^= v0; \ 55 v0 = ROTL(v0, 32); \ 59 v0 += v3; \ 61 v3 ^= v0; \ 128 ctx->v0 = 0x736f6d6570736575ULL ^ k0; in SipHash_Init() 145 uint64_t v0 = ctx->v0; in SipHash_Update() local 173 v0 ^= m; in SipHash_Update() 183 v0 ^= m; in SipHash_Update() 191 ctx->v0 = v0; in SipHash_Update() [all …]
|