/dragonfly/crypto/libressl/crypto/modes/ |
H A D | xts128.c | 80 memcpy(scratch.c,inp,16); in CRYPTO_xts128_encrypt() 81 scratch.u[0] ^= tweak.u[0]; in CRYPTO_xts128_encrypt() 87 (*ctx->block1)(scratch.c,scratch.c,ctx->key1); in CRYPTO_xts128_encrypt() 91 memcpy(out,scratch.c,16); in CRYPTO_xts128_encrypt() 124 out[i] = scratch.c[i]; in CRYPTO_xts128_encrypt() 125 scratch.c[i] = c; in CRYPTO_xts128_encrypt() 129 (*ctx->block1)(scratch.c,scratch.c,ctx->key1); in CRYPTO_xts128_encrypt() 156 memcpy(scratch.c,inp,16); in CRYPTO_xts128_encrypt() 163 (*ctx->block1)(scratch.c,scratch.c,ctx->key1); in CRYPTO_xts128_encrypt() 170 scratch.c[i] = c; in CRYPTO_xts128_encrypt() [all …]
|
H A D | ccm128.c | 214 temp.u[0] ^= scratch.u[0]; in CRYPTO_ccm128_encrypt() 215 temp.u[1] ^= scratch.u[1]; in CRYPTO_ccm128_encrypt() 237 ctx->cmac.u[0] ^= scratch.u[0]; in CRYPTO_ccm128_encrypt() 238 ctx->cmac.u[1] ^= scratch.u[1]; in CRYPTO_ccm128_encrypt() 280 memcpy (out,scratch.c,16); in CRYPTO_ccm128_decrypt() 303 ctx->cmac.u[0] ^= scratch.u[0]; in CRYPTO_ccm128_decrypt() 304 ctx->cmac.u[1] ^= scratch.u[1]; in CRYPTO_ccm128_decrypt() 373 ctx->cmac.u[0] ^= scratch.u[0]; in CRYPTO_ccm128_encrypt_ccm64() 374 ctx->cmac.u[1] ^= scratch.u[1]; in CRYPTO_ccm128_encrypt_ccm64() 426 ctx->cmac.u[0] ^= scratch.u[0]; in CRYPTO_ccm128_decrypt_ccm64() [all …]
|
/dragonfly/contrib/gmp/mpn/generic/ |
H A D | mul.c | 165 mp_ptr scratch; in mpn_mul() local 168 scratch = TMP_SALLOC_LIMBS (ITCH); in mpn_mul() 201 mpn_toom22_mul (ws, up, un, vp, vn, scratch); in mpn_mul() 203 mpn_toom32_mul (ws, up, un, vp, vn, scratch); in mpn_mul() 205 mpn_toom42_mul (ws, up, un, vp, vn, scratch); in mpn_mul() 232 mp_ptr scratch; in mpn_mul() local 235 scratch = TMP_SALLOC_LIMBS (ITCH); in mpn_mul() 281 mpn_toom33_mul (prodp, up, un, vp, vn, scratch); in mpn_mul() 294 mpn_toom32_mul (prodp, up, un, vp, vn, scratch); in mpn_mul() 296 mpn_toom53_mul (prodp, up, un, vp, vn, scratch); in mpn_mul() [all …]
|
H A D | invert.c | 35 mpn_invert (mp_ptr ip, mp_srcptr dp, mp_size_t n, mp_ptr scratch) in mpn_invert() argument 40 ASSERT (! MPN_OVERLAP_P (ip, n, scratch, mpn_invertappr_itch(n))); in mpn_invert() 41 ASSERT (! MPN_OVERLAP_P (dp, n, scratch, mpn_invertappr_itch(n))); in mpn_invert() 49 if (scratch == NULL) in mpn_invert() 50 scratch = TMP_ALLOC_LIMBS (mpn_invert_itch (n)); in mpn_invert() 58 xp = scratch; /* 2 * n limbs */ in mpn_invert() 75 e = mpn_ni_invertappr (ip, dp, n, scratch); in mpn_invert() 79 mpn_mul_n (scratch, ip, dp, n); in mpn_invert() 80 ASSERT_NOCARRY (mpn_add_n (scratch + n, scratch + n, dp, n)); in mpn_invert() 81 if (! mpn_add (scratch, scratch, 2*n, dp, n)) in mpn_invert()
|
H A D | mu_div_qr.c | 92 mp_ptr scratch) in mpn_mu_div_qr() argument 113 scratch); in mpn_mu_div_qr() 122 cy = mpn_add_n (scratch + qn, scratch + qn, dp, dn - (qn + 1)); in mpn_mu_div_qr() 125 scratch[dn - 1] = cy; in mpn_mu_div_qr() 153 mp_ptr scratch) in mpn_mu_div_qr2() argument 171 ip = scratch; in mpn_mu_div_qr2() 172 tp = scratch + in + 1; in mpn_mu_div_qr2() 196 ip = scratch; in mpn_mu_div_qr2() 197 tp = scratch + in; in mpn_mu_div_qr2() 231 mp_ptr scratch) in mpn_preinv_mu_div_qr() argument [all …]
|
H A D | invertappr.c | 165 #define rp scratch in mpn_ni_invertappr() 170 ASSERT (! MPN_OVERLAP_P (ip, n, scratch, mpn_invertappr_itch(n))); in mpn_ni_invertappr() 171 ASSERT (! MPN_OVERLAP_P (dp, n, scratch, mpn_invertappr_itch(n))); in mpn_ni_invertappr() 188 mpn_bc_invertappr (ip - rn, dp - rn, rn, scratch); in mpn_ni_invertappr() 201 xp = scratch + n + 3; /* n + rn limbs */ in mpn_ni_invertappr() 295 if (scratch == NULL) in mpn_invertappr() 296 scratch = TMP_ALLOC_LIMBS (mpn_invertappr_itch (n)); in mpn_invertappr() 301 ASSERT (! MPN_OVERLAP_P (ip, n, scratch, mpn_invertappr_itch(n))); in mpn_invertappr() 302 ASSERT (! MPN_OVERLAP_P (dp, n, scratch, mpn_invertappr_itch(n))); in mpn_invertappr() 305 res = mpn_bc_invertappr (ip, dp, n, scratch); in mpn_invertappr() [all …]
|
H A D | mu_div_q.c | 64 mp_ptr scratch) in mpn_mu_div_q() argument 98 MPN_COPY (scratch + 1, dp, in); in mpn_mu_div_q() 99 scratch[0] = 1; in mpn_mu_div_q() 100 mpn_invertappr (ip, scratch, in + 1, NULL); in mpn_mu_div_q() 105 cy = mpn_add_1 (scratch, dp + dn - (in + 1), in + 1, 1); in mpn_mu_div_q() 110 mpn_invertappr (ip, scratch, in + 1, NULL); in mpn_mu_div_q() 122 this_ip, this_in, scratch); in mpn_mu_div_q() 129 this_ip, this_in, scratch); in mpn_mu_div_q() 175 dp + dn - (qn + 1), qn + 1, scratch); in mpn_mu_div_q()
|
H A D | div_q.c | 92 mp_srcptr dp, mp_size_t dn, mp_ptr scratch) in mpn_div_q() argument 107 ASSERT (MPN_SAME_OR_SEPARATE_P (np, scratch, nn)); in mpn_div_q() 123 new_np = scratch; in mpn_div_q() 158 mp_ptr scratch = TMP_ALLOC_LIMBS (itch); in mpn_div_q() local 159 qh = mpn_mu_div_q (qp, new_np, new_nn, new_dp, dn, scratch); in mpn_div_q() 200 mp_ptr scratch = TMP_ALLOC_LIMBS (itch); in mpn_div_q() local 201 qh = mpn_mu_div_q (qp, np, nn, dp, dn, scratch); in mpn_div_q() 212 new_np = scratch; in mpn_div_q() 251 mp_ptr scratch = TMP_ALLOC_LIMBS (itch); in mpn_div_q() local 252 qh = mpn_mu_divappr_q (tp, new_np, new_nn, new_dp, qn + 1, scratch); in mpn_div_q() [all …]
|
H A D | toom43_mul.c | 53 mp_srcptr bp, mp_size_t bn, mp_ptr scratch) in mpn_toom43_mul() argument 81 #define vm1 (scratch) /* 2n+1 */ in mpn_toom43_mul() 83 #define vm2 (scratch + 2 * n + 1) /* 2n+1 */ in mpn_toom43_mul() 84 #define v2 (scratch + 4 * n + 2) /* 2n+1 */ in mpn_toom43_mul() 87 #define bsm1 (scratch + 2 * n + 2) /* n+1 */ in mpn_toom43_mul() 88 #define asm1 (scratch + 3 * n + 3) /* n+1 */ in mpn_toom43_mul() 89 #define asm2 (scratch + 4 * n + 4) /* n+1 */ in mpn_toom43_mul() 98 #define a0a2 scratch in mpn_toom43_mul() 99 #define b0b2 scratch in mpn_toom43_mul()
|
H A D | redc_n.c | 40 mp_ptr xp, yp, scratch; in mpn_redc_n() local 48 scratch = TMP_ALLOC_LIMBS (n + rn + mpn_mulmod_bnm1_itch (rn, n, n)); in mpn_redc_n() 50 xp = scratch; in mpn_redc_n() 53 yp = scratch + n; in mpn_redc_n() 54 mpn_mulmod_bnm1 (yp, rn, xp, n, mp, n, scratch + n + rn); in mpn_redc_n()
|
H A D | mu_divappr_q.c | 70 mp_ptr scratch) in mpn_mu_divappr_q() argument 97 ip = scratch; in mpn_mu_divappr_q() 98 tp = scratch + in + 1; in mpn_mu_divappr_q() 122 ip = scratch; in mpn_mu_divappr_q() 123 tp = scratch + in; in mpn_mu_divappr_q() 143 qh = mpn_preinv_mu_divappr_q (qp, np, nn, dp, dn, ip, in, scratch + in); in mpn_mu_divappr_q() 156 mp_ptr scratch) in mpn_preinv_mu_divappr_q() argument 163 #define rp scratch in mpn_preinv_mu_divappr_q() 164 #define tp (scratch + dn) in mpn_preinv_mu_divappr_q() 165 #define scratch_out (scratch + dn + tn) in mpn_preinv_mu_divappr_q()
|
H A D | gcd.c | 58 mp_size_t scratch; in mpn_gcd() local 70 scratch = usize - n + 1; in mpn_gcd() 71 if (scratch > talloc) in mpn_gcd() 72 talloc = scratch; in mpn_gcd() 83 mp_size_t scratch; in mpn_gcd() local 95 scratch = matrix_scratch + MAX(hgcd_scratch, update_scratch); in mpn_gcd() 96 if (scratch > talloc) in mpn_gcd() 97 talloc = scratch; in mpn_gcd()
|
H A D | mu_bdiv_q.c | 62 mp_ptr scratch) in mpn_mu_bdiv_q() argument 81 #define ip scratch /* in */ in mpn_mu_bdiv_q() 82 #define rp (scratch + in) /* dn or rest >= binvert_itch(in) */ in mpn_mu_bdiv_q() 83 #define tp (scratch + in + dn) /* dn+in or next_size(dn) */ in mpn_mu_bdiv_q() 84 #define scratch_out (scratch + in + dn + tn) /* mulmod_bnm1_itch(next_size(dn)) */ in mpn_mu_bdiv_q() 183 #define ip scratch /* in */ in mpn_mu_bdiv_q() 184 #define tp (scratch + in) /* qn+in or next_size(qn) or rest >= binvert_itch(in) */ in mpn_mu_bdiv_q() 185 #define scratch_out (scratch + in + tn)/* mulmod_bnm1_itch(next_size(qn)) */ in mpn_mu_bdiv_q()
|
H A D | toom52_mul.c | 57 mp_srcptr bp, mp_size_t bn, mp_ptr scratch) in mpn_toom52_mul() argument 84 #define vm1 (scratch) /* 2n+1 */ in mpn_toom52_mul() 86 #define vm2 (scratch + 2 * n + 1) /* 2n+1 */ in mpn_toom52_mul() 87 #define v2 (scratch + 4 * n + 2) /* 2n+1 */ in mpn_toom52_mul() 90 #define bsm1 (scratch + 2 * n + 2) /* n */ in mpn_toom52_mul() 91 #define asm1 (scratch + 3 * n + 3) /* n+1 */ in mpn_toom52_mul() 92 #define asm2 (scratch + 4 * n + 4) /* n+1 */ in mpn_toom52_mul() 101 #define a0a2 scratch in mpn_toom52_mul()
|
H A D | toom4_sqr.c | 76 mp_ptr scratch) in mpn_toom4_sqr() argument 98 #define v2 scratch /* 2n+1 */ in mpn_toom4_sqr() 99 #define vm2 (scratch + 2 * n + 1) /* 2n+1 */ in mpn_toom4_sqr() 100 #define vh (scratch + 4 * n + 2) /* 2n+1 */ in mpn_toom4_sqr() 101 #define vm1 (scratch + 6 * n + 3) /* 2n+1 */ in mpn_toom4_sqr() 102 #define tp (scratch + 8*n + 5) in mpn_toom4_sqr()
|
H A D | divexact.c | 126 mp_ptr scratch) in mpn_divexact() argument 148 tp = scratch; in mpn_divexact() 165 tp = scratch; in mpn_divexact() 173 tp = scratch; in mpn_divexact() 180 mpn_mu_bdiv_q (qp, np, qn, dp, dn, scratch); in mpn_divexact() 258 mpn_mu_divappr_q (qp + qn0 - 1, tp, nn1, xdp, qn1, scratch); in mpn_divexact() 277 mpn_mu_bdiv_q (qp, np, qn0, dp, qn0, scratch); in mpn_divexact()
|
H A D | mu_bdiv_qr.c | 62 mp_ptr scratch) in mpn_mu_bdiv_qr() argument 83 #define ip scratch /* in */ in mpn_mu_bdiv_qr() 84 #define tp (scratch + in) /* dn+in or next_size(dn) or rest >= binvert_itch(in) */ in mpn_mu_bdiv_qr() 85 #define scratch_out (scratch + in + tn)/* mulmod_bnm1_itch(next_size(dn)) */ in mpn_mu_bdiv_qr() 178 #define ip scratch /* in */ in mpn_mu_bdiv_qr() 179 #define tp (scratch + in) /* dn+in or next_size(dn) or rest >= binvert_itch(in) */ in mpn_mu_bdiv_qr() 180 #define scratch_out (scratch + in + tn)/* mulmod_bnm1_itch(next_size(dn)) */ in mpn_mu_bdiv_qr()
|
H A D | toom3_sqr.c | 69 mp_ptr scratch) in mpn_toom3_sqr() argument 86 as1 = scratch + 4 * n + 4; in mpn_toom3_sqr() 87 asm1 = scratch + 2 * n + 2; in mpn_toom3_sqr() 90 gp = scratch; in mpn_toom3_sqr() 152 #define vm1 scratch /* 2n+1 */ in mpn_toom3_sqr() 153 #define v2 (scratch + 2 * n + 1) /* 2n+2 */ in mpn_toom3_sqr() 154 #define scratch_out (scratch + 5 * n + 5) in mpn_toom3_sqr()
|
H A D | gcdext.c | 193 mp_size_t scratch; in mpn_gcdext() local 215 scratch = an - n + 1; in mpn_gcdext() 216 if (scratch > talloc) in mpn_gcdext() 217 talloc = scratch; in mpn_gcdext() 232 scratch = matrix_scratch + MAX(hgcd_scratch, update_scratch); in mpn_gcdext() 233 if (scratch > talloc) in mpn_gcdext() 234 talloc = scratch; in mpn_gcdext() 238 scratch = MPN_GCDEXT_LEHMER_N_ITCH (GCDEXT_DC_THRESHOLD) in mpn_gcdext() 241 if (scratch > talloc) in mpn_gcdext() 242 talloc = scratch; in mpn_gcdext()
|
H A D | toom33_mul.c | 80 mp_ptr scratch) in mpn_toom33_mul() argument 106 as1 = scratch + 4 * n + 4; in mpn_toom33_mul() 107 asm1 = scratch + 2 * n + 2; in mpn_toom33_mul() 111 bsm1 = scratch + 3 * n + 3; /* we need 4n+4 <= 4n+s+t */ in mpn_toom33_mul() 114 gp = scratch; in mpn_toom33_mul() 240 #define vm1 scratch /* 2n+1 */ in mpn_toom33_mul() 241 #define v2 (scratch + 2 * n + 1) /* 2n+2 */ in mpn_toom33_mul() 242 #define scratch_out (scratch + 5 * n + 5) in mpn_toom33_mul()
|
H A D | toom8_sqr.c | 107 mpn_toom8_sqr (mp_ptr pp, mp_srcptr ap, mp_size_t an, mp_ptr scratch) in mpn_toom8_sqr() argument 126 #define r7 (scratch) /* 3n+1 */ in mpn_toom8_sqr() 127 #define r5 (scratch + 3 * n + 1) /* 3n+1 */ in mpn_toom8_sqr() 128 #define r3 (scratch + 6 * n + 2) /* 3n+1 */ in mpn_toom8_sqr() 129 #define r1 (scratch + 9 * n + 3) /* 3n+1 */ in mpn_toom8_sqr() 132 #define wse (scratch +12 * n + 4) /* 3n+1 */ in mpn_toom8_sqr()
|
/dragonfly/contrib/lvm2/dist/libdm/regex/ |
H A D | matcher.c | 39 struct dm_pool *scratch, *mem; member 197 struct ttree *tt = ttree_create(m->scratch, iwidth); in _calc_states() 245 _create_state_queue(m->scratch, in _calc_states() 278 if (!scratch) in dm_regex_create() 282 dm_pool_destroy(scratch); in dm_regex_create() 292 ptr = all = dm_pool_alloc(scratch, len + 1); in dm_regex_create() 304 if (!(rx = rx_parse_tok(scratch, all, ptr))) { in dm_regex_create() 310 m->scratch = scratch; in dm_regex_create() 321 dm_pool_destroy(scratch); in dm_regex_create() 322 m->scratch = NULL; in dm_regex_create() [all …]
|
/dragonfly/sys/vfs/hammer2/ |
H A D | hammer2.h | 822 char *scratch; /* MAXPHYS */ member 1777 void *scratch, int clindex); 1779 void *scratch, int clindex); 1781 void hammer2_xop_inode_chain_sync(hammer2_xop_t *xop, void *scratch, 1783 void hammer2_xop_inode_unlinkall(hammer2_xop_t *xop, void *scratch, 1798 void hammer2_dmsg_inode_mkdirent(hammer2_xop_t *xop, void *scratch, 1804 void hammer2_dmsg_inode_unlinkall(hammer2_xop_t *xop, void *scratch, 1809 void hammer2_dmsg_strategy_write(hammer2_xop_t *xop, void *scratch, 1820 void hammer2_rmsg_inode_mkdirent(hammer2_xop_t *xop, void *scratch, 1826 void hammer2_rmsg_inode_unlinkall(hammer2_xop_t *xop, void *scratch, [all …]
|
/dragonfly/usr.sbin/makefs/hammer2/ |
H A D | hammer2.h | 863 char *scratch; /* MAXPHYS */ member 1838 void *scratch, int clindex); 1840 void *scratch, int clindex); 1842 void hammer2_xop_inode_chain_sync(hammer2_xop_t *xop, void *scratch, 1844 void hammer2_xop_inode_unlinkall(hammer2_xop_t *xop, void *scratch, 1859 void hammer2_dmsg_inode_mkdirent(hammer2_xop_t *xop, void *scratch, 1865 void hammer2_dmsg_inode_unlinkall(hammer2_xop_t *xop, void *scratch, 1870 void hammer2_dmsg_strategy_write(hammer2_xop_t *xop, void *scratch, 1881 void hammer2_rmsg_inode_mkdirent(hammer2_xop_t *xop, void *scratch, 1887 void hammer2_rmsg_inode_unlinkall(hammer2_xop_t *xop, void *scratch, [all …]
|
/dragonfly/contrib/lvm2/dist/lib/filters/ |
H A D | filter-regex.c | 94 struct dm_pool *scratch; in _build_matcher() local 100 if (!(scratch = dm_pool_create("filter dm_regex", 1024))) in _build_matcher() 118 if (!(regex = dm_pool_alloc(scratch, sizeof(*regex) * count))) in _build_matcher() 132 if (!_extract_pattern(scratch, v->v.str, regex, rf->accept, i)) { in _build_matcher() 146 dm_pool_destroy(scratch); in _build_matcher()
|