/linux/drivers/md/ |
H A D | dm-switch.c | 66 if (!sctx) in alloc_switch_ctx() 69 sctx->ti = ti; in alloc_switch_ctx() 74 return sctx; in alloc_switch_ctx() 83 if (!(sctx->region_size & (sctx->region_size - 1))) in alloc_region_table() 84 sctx->region_size_bits = __ffs(sctx->region_size); in alloc_region_table() 212 &sctx->path_list[sctx->nr_paths].dmdev); in parse_path() 220 dm_put_device(ti, sctx->path_list[sctx->nr_paths].dmdev); in parse_path() 224 sctx->path_list[sctx->nr_paths].start = start; in parse_path() 242 kfree(sctx); in switch_dtr() 288 if (!sctx) { in switch_ctr() [all …]
|
/linux/include/crypto/ |
H A D | sha512_base.h | 26 sctx->state[0] = SHA384_H0; in sha384_base_init() 27 sctx->state[1] = SHA384_H1; in sha384_base_init() 34 sctx->count[0] = sctx->count[1] = 0; in sha384_base_init() 51 sctx->count[0] = sctx->count[1] = 0; in sha512_base_init() 64 sctx->count[0] += len; in sha512_base_do_update() 66 sctx->count[1]++; in sha512_base_do_update() 78 block_fn(sctx, sctx->buf, 1); in sha512_base_do_update() 109 block_fn(sctx, sctx->buf, 1); in sha512_base_do_finalize() 113 bits[0] = cpu_to_be64(sctx->count[1] << 3 | sctx->count[0] >> 61); in sha512_base_do_finalize() 115 block_fn(sctx, sctx->buf, 1); in sha512_base_do_finalize() [all …]
|
H A D | sm3_base.h | 25 sctx->state[0] = SM3_IVA; in sm3_base_init() 26 sctx->state[1] = SM3_IVB; in sm3_base_init() 27 sctx->state[2] = SM3_IVC; in sm3_base_init() 28 sctx->state[3] = SM3_IVD; in sm3_base_init() 29 sctx->state[4] = SM3_IVE; in sm3_base_init() 33 sctx->count = 0; in sm3_base_init() 46 sctx->count += len; in sm3_base_do_update() 58 block_fn(sctx, sctx->buffer, 1); in sm3_base_do_update() 89 block_fn(sctx, sctx->buffer, 1); in sm3_base_do_finalize() 94 block_fn(sctx, sctx->buffer, 1); in sm3_base_do_finalize() [all …]
|
H A D | sha256_base.h | 25 sha224_init(sctx); in sha224_base_init() 33 sha256_init(sctx); in sha256_base_init() 44 sctx->count += len; in lib_sha256_base_do_update() 52 memcpy(sctx->buf + partial, data, p); in lib_sha256_base_do_update() 56 block_fn(sctx, sctx->buf, 1); in lib_sha256_base_do_update() 63 block_fn(sctx, data, blocks); in lib_sha256_base_do_update() 91 sctx->buf[partial++] = 0x80; in lib_sha256_base_do_finalize() 96 block_fn(sctx, sctx->buf, 1); in lib_sha256_base_do_finalize() 100 *bits = cpu_to_be64(sctx->count << 3); in lib_sha256_base_do_finalize() 101 block_fn(sctx, sctx->buf, 1); in lib_sha256_base_do_finalize() [all …]
|
H A D | sha1_base.h | 25 sctx->state[0] = SHA1_H0; in sha1_base_init() 26 sctx->state[1] = SHA1_H1; in sha1_base_init() 27 sctx->state[2] = SHA1_H2; in sha1_base_init() 28 sctx->state[3] = SHA1_H3; in sha1_base_init() 29 sctx->state[4] = SHA1_H4; in sha1_base_init() 30 sctx->count = 0; in sha1_base_init() 43 sctx->count += len; in sha1_base_do_update() 55 block_fn(sctx, sctx->buffer, 1); in sha1_base_do_update() 86 block_fn(sctx, sctx->buffer, 1); in sha1_base_do_finalize() 91 block_fn(sctx, sctx->buffer, 1); in sha1_base_do_finalize() [all …]
|
H A D | sha2.h | 105 sctx->state[0] = SHA256_H0; in sha256_init() 106 sctx->state[1] = SHA256_H1; in sha256_init() 107 sctx->state[2] = SHA256_H2; in sha256_init() 108 sctx->state[3] = SHA256_H3; in sha256_init() 109 sctx->state[4] = SHA256_H4; in sha256_init() 110 sctx->state[5] = SHA256_H5; in sha256_init() 111 sctx->state[6] = SHA256_H6; in sha256_init() 112 sctx->state[7] = SHA256_H7; in sha256_init() 113 sctx->count = 0; in sha256_init() 121 sctx->state[0] = SHA224_H0; in sha224_init() [all …]
|
/linux/fs/btrfs/ |
H A D | send.c | 805 ret = write_buf(sctx->send_filp, sctx->send_buf, sctx->send_size, in send_cmd() 1633 backref_ctx.sctx = sctx; in find_extent_clone() 4235 ret = get_cur_path(sctx, sctx->cur_ino, sctx->cur_inode_gen, in process_recorded_refs() 4505 ret = can_rmdir(sctx, sctx->cur_ino, sctx->cur_inode_gen); in process_recorded_refs() 4955 ret = get_cur_path(sctx, sctx->cur_ino, sctx->cur_inode_gen, p); in __process_new_xattr() 4978 ret = get_cur_path(sctx, sctx->cur_ino, sctx->cur_inode_gen, p); in __process_deleted_xattr() 5225 ret = get_cur_path(sctx, sctx->cur_ino, sctx->cur_inode_gen, p); in process_verity() 6748 ret = send_chown(sctx, sctx->cur_ino, sctx->cur_inode_gen, in finish_inode_if_needed() 6754 ret = send_chmod(sctx, sctx->cur_ino, sctx->cur_inode_gen, in finish_inode_if_needed() 6760 ret = send_fileattr(sctx, sctx->cur_ino, sctx->cur_inode_gen, in finish_inode_if_needed() [all …]
|
/linux/arch/powerpc/crypto/ |
H A D | md5-glue.c | 23 u32 *ptr = (u32 *)sctx; in ppc_md5_clear_context() 34 sctx->hash[0] = MD5_H0; in ppc_md5_init() 35 sctx->hash[1] = MD5_H1; in ppc_md5_init() 36 sctx->hash[2] = MD5_H2; in ppc_md5_init() 37 sctx->hash[3] = MD5_H3; in ppc_md5_init() 38 sctx->byte_count = 0; in ppc_md5_init() 51 sctx->byte_count += len; in ppc_md5_update() 60 ppc_md5_transform(sctx->hash, (const u8 *)sctx->block, 1); in ppc_md5_update() 90 p = (char *)sctx->block; in ppc_md5_final() 111 memcpy(out, sctx, sizeof(*sctx)); in ppc_md5_export() [all …]
|
H A D | sha1-spe-glue.c | 52 u32 *ptr = (u32 *)sctx; in ppc_sha1_clear_context() 69 sctx->count += len; in ppc_spe_sha1_update() 74 sctx->count += len; in ppc_spe_sha1_update() 80 ppc_spe_sha1_transform(sctx->state, (const u8 *)sctx->buffer, 1); in ppc_spe_sha1_update() 119 ppc_spe_sha1_transform(sctx->state, sctx->buffer, 1); in ppc_spe_sha1_final() 120 p = (char *)sctx->buffer; in ppc_spe_sha1_final() 126 ppc_spe_sha1_transform(sctx->state, sctx->buffer, 1); in ppc_spe_sha1_final() 130 dst[0] = cpu_to_be32(sctx->state[0]); in ppc_spe_sha1_final() 136 ppc_sha1_clear_context(sctx); in ppc_spe_sha1_final() 144 memcpy(out, sctx, sizeof(*sctx)); in ppc_spe_sha1_export() [all …]
|
H A D | sha256-spe-glue.c | 53 u32 *ptr = (u32 *)sctx; in ppc_sha256_clear_context() 70 sctx->count += len; in ppc_spe_sha256_update() 75 sctx->count += len; in ppc_spe_sha256_update() 81 ppc_spe_sha256_transform(sctx->state, (const u8 *)sctx->buf, 1); in ppc_spe_sha256_update() 101 memcpy((char *)sctx->buf, src, len); in ppc_spe_sha256_update() 121 ppc_spe_sha256_transform(sctx->state, sctx->buf, 1); in ppc_spe_sha256_final() 122 p = (char *)sctx->buf; in ppc_spe_sha256_final() 128 ppc_spe_sha256_transform(sctx->state, sctx->buf, 1); in ppc_spe_sha256_final() 141 ppc_sha256_clear_context(sctx); in ppc_spe_sha256_final() 170 memcpy(out, sctx, sizeof(*sctx)); in ppc_spe_sha256_export() [all …]
|
H A D | sha1.c | 29 struct sha1_state *sctx = shash_desc_ctx(desc); in powerpc_sha1_update() local 33 partial = sctx->count & 0x3f; in powerpc_sha1_update() 34 sctx->count += len; in powerpc_sha1_update() 43 src = sctx->buffer; in powerpc_sha1_update() 47 powerpc_sha_transform(sctx->state, src); in powerpc_sha1_update() 69 bits = cpu_to_be64(sctx->count << 3); in powerpc_sha1_final() 72 index = sctx->count & 0x3f; in powerpc_sha1_final() 81 dst[i] = cpu_to_be32(sctx->state[i]); in powerpc_sha1_final() 84 memset(sctx, 0, sizeof *sctx); in powerpc_sha1_final() 93 memcpy(out, sctx, sizeof(*sctx)); in powerpc_sha1_export() [all …]
|
/linux/arch/mips/cavium-octeon/crypto/ |
H A D | octeon-sha512.c | 89 if ((sctx->count[0] += len) < len) in __octeon_sha512_update() 90 sctx->count[1]++; in __octeon_sha512_update() 97 octeon_sha512_transform(sctx->buf); in __octeon_sha512_update() 128 octeon_sha512_store_hash(sctx); in octeon_sha512_update() 130 __octeon_sha512_update(sctx, data, len); in octeon_sha512_update() 132 octeon_sha512_read_hash(sctx); in octeon_sha512_update() 152 bits[0] = cpu_to_be64(sctx->count[1] << 3 | sctx->count[0] >> 61); in octeon_sha512_final() 155 index = sctx->count[0] & 0x7f; in octeon_sha512_final() 159 octeon_sha512_store_hash(sctx); in octeon_sha512_final() 166 octeon_sha512_read_hash(sctx); in octeon_sha512_final() [all …]
|
H A D | octeon-sha1.c | 34 u64 *hash = (u64 *)sctx->state; in octeon_sha1_store_hash() 48 u64 *hash = (u64 *)sctx->state; in octeon_sha1_read_hash() 83 sctx->count += len; in __octeon_sha1_update() 92 src = sctx->buffer; in __octeon_sha1_update() 122 octeon_sha1_store_hash(sctx); in octeon_sha1_update() 126 octeon_sha1_read_hash(sctx); in octeon_sha1_update() 148 index = sctx->count & 0x3f; in octeon_sha1_final() 159 octeon_sha1_read_hash(sctx); in octeon_sha1_final() 167 memset(sctx, 0, sizeof(*sctx)); in octeon_sha1_final() 176 memcpy(out, sctx, sizeof(*sctx)); in octeon_sha1_export() [all …]
|
H A D | octeon-sha256.c | 35 u64 *hash = (u64 *)sctx->state; in octeon_sha256_store_hash() 45 u64 *hash = (u64 *)sctx->state; in octeon_sha256_read_hash() 75 sctx->count += len; in __octeon_sha256_update() 84 src = sctx->buf; in __octeon_sha256_update() 114 octeon_sha256_store_hash(sctx); in octeon_sha256_update() 118 octeon_sha256_read_hash(sctx); in octeon_sha256_update() 140 index = sctx->count & 0x3f; in octeon_sha256_final() 151 octeon_sha256_read_hash(sctx); in octeon_sha256_final() 159 memset(sctx, 0, sizeof(*sctx)); in octeon_sha256_final() 180 memcpy(out, sctx, sizeof(*sctx)); in octeon_sha256_export() [all …]
|
/linux/arch/s390/crypto/ |
H A D | sha256_s390.c | 24 sctx->state[0] = SHA256_H0; in s390_sha256_init() 25 sctx->state[1] = SHA256_H1; in s390_sha256_init() 26 sctx->state[2] = SHA256_H2; in s390_sha256_init() 27 sctx->state[3] = SHA256_H3; in s390_sha256_init() 28 sctx->state[4] = SHA256_H4; in s390_sha256_init() 29 sctx->state[5] = SHA256_H5; in s390_sha256_init() 30 sctx->state[6] = SHA256_H6; in s390_sha256_init() 31 sctx->state[7] = SHA256_H7; in s390_sha256_init() 32 sctx->count = 0; in s390_sha256_init() 43 octx->count = sctx->count; in sha256_export() [all …]
|
H A D | sha3_512_s390.c | 23 memset(sctx->state, 0, sizeof(sctx->state)); in sha3_512_init() 24 sctx->count = 0; in sha3_512_init() 25 sctx->func = CPACF_KIMD_SHA3_512; in sha3_512_init() 35 octx->rsiz = sctx->count; in sha3_512_export() 36 octx->rsizw = sctx->count >> 32; in sha3_512_export() 51 sctx->count = ictx->rsiz; in sha3_512_import() 55 sctx->func = CPACF_KIMD_SHA3_512; in sha3_512_import() 67 sctx->count = ictx->rsiz; in sha3_384_import() 71 sctx->func = CPACF_KIMD_SHA3_384; in sha3_384_import() 100 memset(sctx->state, 0, sizeof(sctx->state)); in sha3_384_init() [all …]
|
H A D | sha3_256_s390.c | 24 memset(sctx->state, 0, sizeof(sctx->state)); in sha3_256_init() 25 sctx->count = 0; in sha3_256_init() 26 sctx->func = CPACF_KIMD_SHA3_256; in sha3_256_init() 36 octx->rsiz = sctx->count; in sha3_256_export() 48 sctx->count = ictx->rsiz; in sha3_256_import() 51 sctx->func = CPACF_KIMD_SHA3_256; in sha3_256_import() 61 sctx->count = ictx->rsiz; in sha3_224_import() 64 sctx->func = CPACF_KIMD_SHA3_224; in sha3_224_import() 91 memset(sctx->state, 0, sizeof(sctx->state)); in sha3_224_init() 92 sctx->count = 0; in sha3_224_init() [all …]
|
/linux/arch/arm64/crypto/ |
H A D | sha3-ce-glue.c | 43 if ((sctx->partial + len) >= sctx->rsiz) { in sha3_update() 47 int p = sctx->rsiz - sctx->partial; in sha3_update() 49 memcpy(sctx->buf + sctx->partial, data, p); in sha3_update() 51 sha3_ce_transform(sctx->st, sctx->buf, 1, digest_size); in sha3_update() 60 len %= sctx->rsiz; in sha3_update() 75 memcpy(sctx->buf + sctx->partial, data, len); in sha3_update() 91 sctx->buf[sctx->partial++] = 0x06; in sha3_final() 92 memset(sctx->buf + sctx->partial, 0, sctx->rsiz - sctx->partial); in sha3_final() 93 sctx->buf[sctx->rsiz - 1] |= 0x80; in sha3_final() 96 sha3_ce_transform(sctx->st, sctx->buf, 1, digest_size); in sha3_final() [all …]
|
/linux/arch/sparc/crypto/ |
H A D | sha1_glue.c | 35 sctx->count += len; in __sha1_sparc64_update() 38 memcpy(sctx->buffer + partial, data, done); in __sha1_sparc64_update() 39 sha1_sparc64_transform(sctx->state, sctx->buffer, 1); in __sha1_sparc64_update() 59 sctx->count += len; in sha1_sparc64_update() 76 bits = cpu_to_be64(sctx->count << 3); in sha1_sparc64_final() 79 index = sctx->count % SHA1_BLOCK_SIZE; in sha1_sparc64_final() 84 sctx->count += padlen; in sha1_sparc64_final() 93 dst[i] = cpu_to_be32(sctx->state[i]); in sha1_sparc64_final() 96 memset(sctx, 0, sizeof(*sctx)); in sha1_sparc64_final() 105 memcpy(out, sctx, sizeof(*sctx)); in sha1_sparc64_export() [all …]
|
H A D | md5_glue.c | 51 sctx->byte_count += len; in __md5_sparc64_update() 55 md5_sparc64_transform(sctx->hash, (u8 *)sctx->block, 1); in __md5_sparc64_update() 64 memcpy(sctx->block, data + done, len - done); in __md5_sparc64_update() 70 struct md5_state *sctx = shash_desc_ctx(desc); in md5_sparc64_update() local 75 sctx->byte_count += len; in md5_sparc64_update() 92 bits = cpu_to_le64(sctx->byte_count << 3); in md5_sparc64_final() 100 sctx->byte_count += padlen; in md5_sparc64_final() 109 dst[i] = sctx->hash[i]; in md5_sparc64_final() 112 memset(sctx, 0, sizeof(*sctx)); in md5_sparc64_final() 121 memcpy(out, sctx, sizeof(*sctx)); in md5_sparc64_export() [all …]
|
H A D | sha256_glue.c | 35 sctx->count += len; in __sha256_sparc64_update() 38 memcpy(sctx->buf + partial, data, done); in __sha256_sparc64_update() 39 sha256_sparc64_transform(sctx->state, sctx->buf, 1); in __sha256_sparc64_update() 59 sctx->count += len; in sha256_sparc64_update() 60 memcpy(sctx->buf + partial, data, len); in sha256_sparc64_update() 75 bits = cpu_to_be64(sctx->count << 3); in sha256_sparc64_final() 83 sctx->count += padlen; in sha256_sparc64_final() 92 dst[i] = cpu_to_be32(sctx->state[i]); in sha256_sparc64_final() 95 memset(sctx, 0, sizeof(*sctx)); in sha256_sparc64_final() 116 memcpy(out, sctx, sizeof(*sctx)); in sha256_sparc64_export() [all …]
|
H A D | sha512_glue.c | 34 if ((sctx->count[0] += len) < len) in __sha512_sparc64_update() 35 sctx->count[1]++; in __sha512_sparc64_update() 38 memcpy(sctx->buf + partial, data, done); in __sha512_sparc64_update() 39 sha512_sparc64_transform(sctx->state, sctx->buf, 1); in __sha512_sparc64_update() 59 if ((sctx->count[0] += len) < len) in sha512_sparc64_update() 60 sctx->count[1]++; in sha512_sparc64_update() 61 memcpy(sctx->buf + partial, data, len); in sha512_sparc64_update() 78 bits[0] = cpu_to_be64(sctx->count[1] << 3 | sctx->count[0] >> 61); in sha512_sparc64_final() 87 sctx->count[1]++; in sha512_sparc64_final() 96 dst[i] = cpu_to_be64(sctx->state[i]); in sha512_sparc64_final() [all …]
|
/linux/crypto/ |
H A D | sha3_generic.c | 167 sctx->rsizw = sctx->rsiz / 8; in crypto_sha3_init() 168 sctx->partial = 0; in crypto_sha3_init() 170 memset(sctx->st, 0, sizeof(sctx->st)); in crypto_sha3_init() 185 if ((sctx->partial + len) > (sctx->rsiz - 1)) { in crypto_sha3_update() 188 memcpy(sctx->buf + sctx->partial, data, in crypto_sha3_update() 190 src = sctx->buf; in crypto_sha3_update() 206 memcpy(sctx->buf + sctx->partial, src, len - done); in crypto_sha3_update() 221 memset(sctx->buf + inlen, 0, sctx->rsiz - inlen); in crypto_sha3_final() 222 sctx->buf[sctx->rsiz - 1] |= 0x80; in crypto_sha3_final() 225 sctx->st[i] ^= get_unaligned_le64(sctx->buf + 8 * i); in crypto_sha3_final() [all …]
|
/linux/drivers/crypto/nx/ |
H A D | nx-sha256.c | 46 memset(sctx, 0, sizeof *sctx); in nx_sha256_init() 48 sctx->state[0] = __cpu_to_be32(SHA256_H0); in nx_sha256_init() 49 sctx->state[1] = __cpu_to_be32(SHA256_H1); in nx_sha256_init() 56 sctx->count = 0; in nx_sha256_init() 83 memcpy(sctx->buf + buf_len, data, len); in nx_sha256_update() 84 sctx->count += len; in nx_sha256_update() 114 (u8 *) sctx->buf, in nx_sha256_update() 172 memcpy(sctx->buf, data, leftover); in nx_sha256_update() 174 sctx->count += len; in nx_sha256_update() 255 memcpy(out, sctx, sizeof(*sctx)); in nx_sha256_export() [all …]
|
H A D | nx-sha512.c | 46 memset(sctx, 0, sizeof *sctx); in nx_sha512_init() 48 sctx->state[0] = __cpu_to_be64(SHA512_H0); in nx_sha512_init() 56 sctx->count[0] = 0; in nx_sha512_init() 83 memcpy(sctx->buf + buf_len, data, len); in nx_sha512_update() 84 sctx->count[0] += len; in nx_sha512_update() 114 (u8 *) sctx->buf, in nx_sha512_update() 176 memcpy(sctx->buf, data, leftover); in nx_sha512_update() 177 sctx->count[0] += len; in nx_sha512_update() 219 count0 = sctx->count[0] * 8; in nx_sha512_final() 261 memcpy(out, sctx, sizeof(*sctx)); in nx_sha512_export() [all …]
|