Home
last modified time | relevance | path

Searched refs:SHA1_BLOCK_SIZE (Results 1 – 25 of 45) sorted by relevance

12

/linux/include/crypto/
H A Dsha1_base.h41 unsigned int partial = sctx->count % SHA1_BLOCK_SIZE; in sha1_base_do_update()
45 if (unlikely((partial + len) >= SHA1_BLOCK_SIZE)) { in sha1_base_do_update()
49 int p = SHA1_BLOCK_SIZE - partial; in sha1_base_do_update()
58 blocks = len / SHA1_BLOCK_SIZE; in sha1_base_do_update()
59 len %= SHA1_BLOCK_SIZE; in sha1_base_do_update()
63 data += blocks * SHA1_BLOCK_SIZE; in sha1_base_do_update()
76 const int bit_offset = SHA1_BLOCK_SIZE - sizeof(__be64); in sha1_base_do_finalize()
79 unsigned int partial = sctx->count % SHA1_BLOCK_SIZE; in sha1_base_do_finalize()
83 memset(sctx->buffer + partial, 0x0, SHA1_BLOCK_SIZE - partial); in sha1_base_do_finalize()
H A Dsha1.h12 #define SHA1_BLOCK_SIZE 64 macro
25 u8 buffer[SHA1_BLOCK_SIZE];
/linux/arch/sparc/crypto/
H A Dsha1_glue.c37 done = SHA1_BLOCK_SIZE - partial; in __sha1_sparc64_update()
41 if (len - done >= SHA1_BLOCK_SIZE) { in __sha1_sparc64_update()
42 const unsigned int rounds = (len - done) / SHA1_BLOCK_SIZE; in __sha1_sparc64_update()
45 done += rounds * SHA1_BLOCK_SIZE; in __sha1_sparc64_update()
55 unsigned int partial = sctx->count % SHA1_BLOCK_SIZE; in sha1_sparc64_update()
58 if (partial + len < SHA1_BLOCK_SIZE) { in sha1_sparc64_update()
74 static const u8 padding[SHA1_BLOCK_SIZE] = { 0x80, }; in sha1_sparc64_final()
79 index = sctx->count % SHA1_BLOCK_SIZE; in sha1_sparc64_final()
80 padlen = (index < 56) ? (56 - index) : ((SHA1_BLOCK_SIZE+56) - index); in sha1_sparc64_final()
132 .cra_blocksize = SHA1_BLOCK_SIZE,
/linux/arch/mips/cavium-octeon/crypto/
H A Docteon-sha1.c82 partial = sctx->count % SHA1_BLOCK_SIZE; in __octeon_sha1_update()
87 if ((partial + len) >= SHA1_BLOCK_SIZE) { in __octeon_sha1_update()
91 done + SHA1_BLOCK_SIZE); in __octeon_sha1_update()
97 done += SHA1_BLOCK_SIZE; in __octeon_sha1_update()
99 } while (done + SHA1_BLOCK_SIZE <= len); in __octeon_sha1_update()
118 if ((sctx->count % SHA1_BLOCK_SIZE) + len < SHA1_BLOCK_SIZE) in octeon_sha1_update()
201 .cra_blocksize = SHA1_BLOCK_SIZE,
/linux/drivers/crypto/
H A Dpadlock-sha.c94 leftover = ((state.count - 1) & (SHA1_BLOCK_SIZE - 1)) + 1; in padlock_sha1_finup()
95 space = SHA1_BLOCK_SIZE - leftover; in padlock_sha1_finup()
108 state.count &= ~(SHA1_BLOCK_SIZE - 1); in padlock_sha1_finup()
169 state.count &= ~(SHA1_BLOCK_SIZE - 1); in padlock_sha256_finup()
238 .cra_blocksize = SHA1_BLOCK_SIZE,
297 if ((partial + len) >= SHA1_BLOCK_SIZE) { in padlock_sha1_update_nano()
303 done + SHA1_BLOCK_SIZE); in padlock_sha1_update_nano()
308 done += SHA1_BLOCK_SIZE; in padlock_sha1_update_nano()
313 if (len - done >= SHA1_BLOCK_SIZE) { in padlock_sha1_update_nano()
317 "c"((unsigned long)((len - done) / SHA1_BLOCK_SIZE))); in padlock_sha1_update_nano()
[all …]
H A Domap-sham.c410 d = SHA1_BLOCK_SIZE; in get_block_size()
960 bs = SHA1_BLOCK_SIZE; in omap_sham_init()
964 bs = SHA1_BLOCK_SIZE; in omap_sham_init()
1436 .cra_blocksize = SHA1_BLOCK_SIZE,
1458 .cra_blocksize = SHA1_BLOCK_SIZE,
1481 .cra_blocksize = SHA1_BLOCK_SIZE,
1505 .cra_blocksize = SHA1_BLOCK_SIZE,
H A Dsa2ul.c392 for (; i < SHA1_BLOCK_SIZE; i++) in prepare_kipad()
404 for (; i < SHA1_BLOCK_SIZE; i++) in prepare_kopad()
444 u8 k_pad[SHA1_BLOCK_SIZE]; in sa_prepare_iopads()
2073 .cra_blocksize = SHA1_BLOCK_SIZE,
H A Datmel-sha.c443 ctx->block_size = SHA1_BLOCK_SIZE; in atmel_sha_init()
1277 .halg.base.cra_blocksize = SHA1_BLOCK_SIZE,
1696 ctx->block_size = SHA1_BLOCK_SIZE; in atmel_sha_hmac_setup()
2065 .halg.base.cra_blocksize = SHA1_BLOCK_SIZE,
/linux/arch/arm/crypto/
H A Dsha1-ce-glue.c35 (sctx->count % SHA1_BLOCK_SIZE) + len < SHA1_BLOCK_SIZE) in sha1_ce_update()
76 .cra_blocksize = SHA1_BLOCK_SIZE,
H A Dsha1_neon_glue.c38 (sctx->count % SHA1_BLOCK_SIZE) + len < SHA1_BLOCK_SIZE) in sha1_neon_update()
79 .cra_blocksize = SHA1_BLOCK_SIZE,
H A Dsha1_glue.c62 .cra_blocksize = SHA1_BLOCK_SIZE,
/linux/arch/x86/crypto/
H A Dsha1_ssse3_glue.c47 (sctx->count % SHA1_BLOCK_SIZE) + len < SHA1_BLOCK_SIZE) in sha1_update()
110 .cra_blocksize = SHA1_BLOCK_SIZE,
159 .cra_blocksize = SHA1_BLOCK_SIZE,
241 .cra_blocksize = SHA1_BLOCK_SIZE,
291 .cra_blocksize = SHA1_BLOCK_SIZE,
/linux/arch/arm64/crypto/
H A Dsha1-ce-glue.c46 src += (blocks - rem) * SHA1_BLOCK_SIZE; in sha1_ce_transform()
72 bool finalize = !sctx->sst.count && !(len % SHA1_BLOCK_SIZE) && len; in sha1_ce_finup()
132 .cra_blocksize = SHA1_BLOCK_SIZE,
/linux/crypto/
H A Dsha1_generic.c37 src += SHA1_BLOCK_SIZE; in sha1_generic_block_fn()
74 .cra_blocksize = SHA1_BLOCK_SIZE,
/linux/arch/s390/crypto/
H A Dsha1_s390.c81 .cra_blocksize = SHA1_BLOCK_SIZE,
/linux/arch/powerpc/crypto/
H A Dsha1.c117 .cra_blocksize = SHA1_BLOCK_SIZE,
H A Dsha1-spe-glue.c169 .cra_blocksize = SHA1_BLOCK_SIZE,
/linux/drivers/crypto/rockchip/
H A Drk3288_crypto_ahash.c394 .cra_blocksize = SHA1_BLOCK_SIZE,
458 .cra_blocksize = SHA1_BLOCK_SIZE,
/linux/drivers/crypto/qce/
H A Dsha.c419 .blocksize = SHA1_BLOCK_SIZE,
437 .blocksize = SHA1_BLOCK_SIZE,
/linux/drivers/crypto/aspeed/
H A Daspeed-hace-hash.c681 rctx->block_size = SHA1_BLOCK_SIZE; in aspeed_sham_init()
871 .cra_blocksize = SHA1_BLOCK_SIZE,
969 .cra_blocksize = SHA1_BLOCK_SIZE,
/linux/drivers/crypto/inside-secure/
H A Dsafexcel_hash.c946 req->block_sz = SHA1_BLOCK_SIZE; in safexcel_sha1_init()
1002 .cra_blocksize = SHA1_BLOCK_SIZE,
1022 req->len = SHA1_BLOCK_SIZE; in safexcel_hmac_sha1_init()
1023 req->processed = SHA1_BLOCK_SIZE; in safexcel_hmac_sha1_init()
1029 req->block_sz = SHA1_BLOCK_SIZE; in safexcel_hmac_sha1_init()
1228 .cra_blocksize = SHA1_BLOCK_SIZE,
/linux/drivers/crypto/allwinner/sun8i-ss/
H A Dsun8i-ss-core.c360 .cra_blocksize = SHA1_BLOCK_SIZE,
457 .cra_blocksize = SHA1_BLOCK_SIZE,
/linux/drivers/crypto/allwinner/sun4i-ss/
H A Dsun4i-ss-core.c78 .cra_blocksize = SHA1_BLOCK_SIZE,
/linux/drivers/crypto/ccp/
H A Dccp-crypto-sha.c375 .block_size = SHA1_BLOCK_SIZE,
/linux/drivers/crypto/marvell/cesa/
H A Dhash.c1025 .cra_blocksize = SHA1_BLOCK_SIZE,
1380 .cra_blocksize = SHA1_BLOCK_SIZE,

12