Home
last modified time | relevance | path

Searched refs:sctx (Results 1 – 25 of 71) sorted by relevance

123

/linux/drivers/md/
H A Ddm-switch.c66 if (!sctx) in alloc_switch_ctx()
69 sctx->ti = ti; in alloc_switch_ctx()
74 return sctx; in alloc_switch_ctx()
83 if (!(sctx->region_size & (sctx->region_size - 1))) in alloc_region_table()
84 sctx->region_size_bits = __ffs(sctx->region_size); in alloc_region_table()
212 &sctx->path_list[sctx->nr_paths].dmdev); in parse_path()
220 dm_put_device(ti, sctx->path_list[sctx->nr_paths].dmdev); in parse_path()
224 sctx->path_list[sctx->nr_paths].start = start; in parse_path()
242 kfree(sctx); in switch_dtr()
288 if (!sctx) { in switch_ctr()
[all …]
/linux/include/crypto/
H A Dsha512_base.h26 sctx->state[0] = SHA384_H0; in sha384_base_init()
27 sctx->state[1] = SHA384_H1; in sha384_base_init()
34 sctx->count[0] = sctx->count[1] = 0; in sha384_base_init()
51 sctx->count[0] = sctx->count[1] = 0; in sha512_base_init()
64 sctx->count[0] += len; in sha512_base_do_update()
66 sctx->count[1]++; in sha512_base_do_update()
78 block_fn(sctx, sctx->buf, 1); in sha512_base_do_update()
109 block_fn(sctx, sctx->buf, 1); in sha512_base_do_finalize()
113 bits[0] = cpu_to_be64(sctx->count[1] << 3 | sctx->count[0] >> 61); in sha512_base_do_finalize()
115 block_fn(sctx, sctx->buf, 1); in sha512_base_do_finalize()
[all …]
H A Dsm3_base.h25 sctx->state[0] = SM3_IVA; in sm3_base_init()
26 sctx->state[1] = SM3_IVB; in sm3_base_init()
27 sctx->state[2] = SM3_IVC; in sm3_base_init()
28 sctx->state[3] = SM3_IVD; in sm3_base_init()
29 sctx->state[4] = SM3_IVE; in sm3_base_init()
33 sctx->count = 0; in sm3_base_init()
46 sctx->count += len; in sm3_base_do_update()
58 block_fn(sctx, sctx->buffer, 1); in sm3_base_do_update()
89 block_fn(sctx, sctx->buffer, 1); in sm3_base_do_finalize()
94 block_fn(sctx, sctx->buffer, 1); in sm3_base_do_finalize()
[all …]
H A Dsha256_base.h25 sha224_init(sctx); in sha224_base_init()
33 sha256_init(sctx); in sha256_base_init()
44 sctx->count += len; in lib_sha256_base_do_update()
52 memcpy(sctx->buf + partial, data, p); in lib_sha256_base_do_update()
56 block_fn(sctx, sctx->buf, 1); in lib_sha256_base_do_update()
63 block_fn(sctx, data, blocks); in lib_sha256_base_do_update()
91 sctx->buf[partial++] = 0x80; in lib_sha256_base_do_finalize()
96 block_fn(sctx, sctx->buf, 1); in lib_sha256_base_do_finalize()
100 *bits = cpu_to_be64(sctx->count << 3); in lib_sha256_base_do_finalize()
101 block_fn(sctx, sctx->buf, 1); in lib_sha256_base_do_finalize()
[all …]
H A Dsha1_base.h25 sctx->state[0] = SHA1_H0; in sha1_base_init()
26 sctx->state[1] = SHA1_H1; in sha1_base_init()
27 sctx->state[2] = SHA1_H2; in sha1_base_init()
28 sctx->state[3] = SHA1_H3; in sha1_base_init()
29 sctx->state[4] = SHA1_H4; in sha1_base_init()
30 sctx->count = 0; in sha1_base_init()
43 sctx->count += len; in sha1_base_do_update()
55 block_fn(sctx, sctx->buffer, 1); in sha1_base_do_update()
86 block_fn(sctx, sctx->buffer, 1); in sha1_base_do_finalize()
91 block_fn(sctx, sctx->buffer, 1); in sha1_base_do_finalize()
[all …]
H A Dsha2.h105 sctx->state[0] = SHA256_H0; in sha256_init()
106 sctx->state[1] = SHA256_H1; in sha256_init()
107 sctx->state[2] = SHA256_H2; in sha256_init()
108 sctx->state[3] = SHA256_H3; in sha256_init()
109 sctx->state[4] = SHA256_H4; in sha256_init()
110 sctx->state[5] = SHA256_H5; in sha256_init()
111 sctx->state[6] = SHA256_H6; in sha256_init()
112 sctx->state[7] = SHA256_H7; in sha256_init()
113 sctx->count = 0; in sha256_init()
121 sctx->state[0] = SHA224_H0; in sha224_init()
[all …]
/linux/fs/btrfs/
H A Dsend.c805 ret = write_buf(sctx->send_filp, sctx->send_buf, sctx->send_size, in send_cmd()
1633 backref_ctx.sctx = sctx; in find_extent_clone()
4235 ret = get_cur_path(sctx, sctx->cur_ino, sctx->cur_inode_gen, in process_recorded_refs()
4505 ret = can_rmdir(sctx, sctx->cur_ino, sctx->cur_inode_gen); in process_recorded_refs()
4955 ret = get_cur_path(sctx, sctx->cur_ino, sctx->cur_inode_gen, p); in __process_new_xattr()
4978 ret = get_cur_path(sctx, sctx->cur_ino, sctx->cur_inode_gen, p); in __process_deleted_xattr()
5225 ret = get_cur_path(sctx, sctx->cur_ino, sctx->cur_inode_gen, p); in process_verity()
6748 ret = send_chown(sctx, sctx->cur_ino, sctx->cur_inode_gen, in finish_inode_if_needed()
6754 ret = send_chmod(sctx, sctx->cur_ino, sctx->cur_inode_gen, in finish_inode_if_needed()
6760 ret = send_fileattr(sctx, sctx->cur_ino, sctx->cur_inode_gen, in finish_inode_if_needed()
[all …]
/linux/arch/powerpc/crypto/
H A Dmd5-glue.c23 u32 *ptr = (u32 *)sctx; in ppc_md5_clear_context()
34 sctx->hash[0] = MD5_H0; in ppc_md5_init()
35 sctx->hash[1] = MD5_H1; in ppc_md5_init()
36 sctx->hash[2] = MD5_H2; in ppc_md5_init()
37 sctx->hash[3] = MD5_H3; in ppc_md5_init()
38 sctx->byte_count = 0; in ppc_md5_init()
51 sctx->byte_count += len; in ppc_md5_update()
60 ppc_md5_transform(sctx->hash, (const u8 *)sctx->block, 1); in ppc_md5_update()
90 p = (char *)sctx->block; in ppc_md5_final()
111 memcpy(out, sctx, sizeof(*sctx)); in ppc_md5_export()
[all …]
H A Dsha1-spe-glue.c52 u32 *ptr = (u32 *)sctx; in ppc_sha1_clear_context()
69 sctx->count += len; in ppc_spe_sha1_update()
74 sctx->count += len; in ppc_spe_sha1_update()
80 ppc_spe_sha1_transform(sctx->state, (const u8 *)sctx->buffer, 1); in ppc_spe_sha1_update()
119 ppc_spe_sha1_transform(sctx->state, sctx->buffer, 1); in ppc_spe_sha1_final()
120 p = (char *)sctx->buffer; in ppc_spe_sha1_final()
126 ppc_spe_sha1_transform(sctx->state, sctx->buffer, 1); in ppc_spe_sha1_final()
130 dst[0] = cpu_to_be32(sctx->state[0]); in ppc_spe_sha1_final()
136 ppc_sha1_clear_context(sctx); in ppc_spe_sha1_final()
144 memcpy(out, sctx, sizeof(*sctx)); in ppc_spe_sha1_export()
[all …]
H A Dsha256-spe-glue.c53 u32 *ptr = (u32 *)sctx; in ppc_sha256_clear_context()
70 sctx->count += len; in ppc_spe_sha256_update()
75 sctx->count += len; in ppc_spe_sha256_update()
81 ppc_spe_sha256_transform(sctx->state, (const u8 *)sctx->buf, 1); in ppc_spe_sha256_update()
101 memcpy((char *)sctx->buf, src, len); in ppc_spe_sha256_update()
121 ppc_spe_sha256_transform(sctx->state, sctx->buf, 1); in ppc_spe_sha256_final()
122 p = (char *)sctx->buf; in ppc_spe_sha256_final()
128 ppc_spe_sha256_transform(sctx->state, sctx->buf, 1); in ppc_spe_sha256_final()
141 ppc_sha256_clear_context(sctx); in ppc_spe_sha256_final()
170 memcpy(out, sctx, sizeof(*sctx)); in ppc_spe_sha256_export()
[all …]
H A Dsha1.c29 struct sha1_state *sctx = shash_desc_ctx(desc); in powerpc_sha1_update() local
33 partial = sctx->count & 0x3f; in powerpc_sha1_update()
34 sctx->count += len; in powerpc_sha1_update()
43 src = sctx->buffer; in powerpc_sha1_update()
47 powerpc_sha_transform(sctx->state, src); in powerpc_sha1_update()
69 bits = cpu_to_be64(sctx->count << 3); in powerpc_sha1_final()
72 index = sctx->count & 0x3f; in powerpc_sha1_final()
81 dst[i] = cpu_to_be32(sctx->state[i]); in powerpc_sha1_final()
84 memset(sctx, 0, sizeof *sctx); in powerpc_sha1_final()
93 memcpy(out, sctx, sizeof(*sctx)); in powerpc_sha1_export()
[all …]
/linux/arch/mips/cavium-octeon/crypto/
H A Docteon-sha512.c89 if ((sctx->count[0] += len) < len) in __octeon_sha512_update()
90 sctx->count[1]++; in __octeon_sha512_update()
97 octeon_sha512_transform(sctx->buf); in __octeon_sha512_update()
128 octeon_sha512_store_hash(sctx); in octeon_sha512_update()
130 __octeon_sha512_update(sctx, data, len); in octeon_sha512_update()
132 octeon_sha512_read_hash(sctx); in octeon_sha512_update()
152 bits[0] = cpu_to_be64(sctx->count[1] << 3 | sctx->count[0] >> 61); in octeon_sha512_final()
155 index = sctx->count[0] & 0x7f; in octeon_sha512_final()
159 octeon_sha512_store_hash(sctx); in octeon_sha512_final()
166 octeon_sha512_read_hash(sctx); in octeon_sha512_final()
[all …]
H A Docteon-sha1.c34 u64 *hash = (u64 *)sctx->state; in octeon_sha1_store_hash()
48 u64 *hash = (u64 *)sctx->state; in octeon_sha1_read_hash()
83 sctx->count += len; in __octeon_sha1_update()
92 src = sctx->buffer; in __octeon_sha1_update()
122 octeon_sha1_store_hash(sctx); in octeon_sha1_update()
126 octeon_sha1_read_hash(sctx); in octeon_sha1_update()
148 index = sctx->count & 0x3f; in octeon_sha1_final()
159 octeon_sha1_read_hash(sctx); in octeon_sha1_final()
167 memset(sctx, 0, sizeof(*sctx)); in octeon_sha1_final()
176 memcpy(out, sctx, sizeof(*sctx)); in octeon_sha1_export()
[all …]
H A Docteon-sha256.c35 u64 *hash = (u64 *)sctx->state; in octeon_sha256_store_hash()
45 u64 *hash = (u64 *)sctx->state; in octeon_sha256_read_hash()
75 sctx->count += len; in __octeon_sha256_update()
84 src = sctx->buf; in __octeon_sha256_update()
114 octeon_sha256_store_hash(sctx); in octeon_sha256_update()
118 octeon_sha256_read_hash(sctx); in octeon_sha256_update()
140 index = sctx->count & 0x3f; in octeon_sha256_final()
151 octeon_sha256_read_hash(sctx); in octeon_sha256_final()
159 memset(sctx, 0, sizeof(*sctx)); in octeon_sha256_final()
180 memcpy(out, sctx, sizeof(*sctx)); in octeon_sha256_export()
[all …]
/linux/arch/s390/crypto/
H A Dsha256_s390.c24 sctx->state[0] = SHA256_H0; in s390_sha256_init()
25 sctx->state[1] = SHA256_H1; in s390_sha256_init()
26 sctx->state[2] = SHA256_H2; in s390_sha256_init()
27 sctx->state[3] = SHA256_H3; in s390_sha256_init()
28 sctx->state[4] = SHA256_H4; in s390_sha256_init()
29 sctx->state[5] = SHA256_H5; in s390_sha256_init()
30 sctx->state[6] = SHA256_H6; in s390_sha256_init()
31 sctx->state[7] = SHA256_H7; in s390_sha256_init()
32 sctx->count = 0; in s390_sha256_init()
43 octx->count = sctx->count; in sha256_export()
[all …]
H A Dsha3_512_s390.c23 memset(sctx->state, 0, sizeof(sctx->state)); in sha3_512_init()
24 sctx->count = 0; in sha3_512_init()
25 sctx->func = CPACF_KIMD_SHA3_512; in sha3_512_init()
35 octx->rsiz = sctx->count; in sha3_512_export()
36 octx->rsizw = sctx->count >> 32; in sha3_512_export()
51 sctx->count = ictx->rsiz; in sha3_512_import()
55 sctx->func = CPACF_KIMD_SHA3_512; in sha3_512_import()
67 sctx->count = ictx->rsiz; in sha3_384_import()
71 sctx->func = CPACF_KIMD_SHA3_384; in sha3_384_import()
100 memset(sctx->state, 0, sizeof(sctx->state)); in sha3_384_init()
[all …]
H A Dsha3_256_s390.c24 memset(sctx->state, 0, sizeof(sctx->state)); in sha3_256_init()
25 sctx->count = 0; in sha3_256_init()
26 sctx->func = CPACF_KIMD_SHA3_256; in sha3_256_init()
36 octx->rsiz = sctx->count; in sha3_256_export()
48 sctx->count = ictx->rsiz; in sha3_256_import()
51 sctx->func = CPACF_KIMD_SHA3_256; in sha3_256_import()
61 sctx->count = ictx->rsiz; in sha3_224_import()
64 sctx->func = CPACF_KIMD_SHA3_224; in sha3_224_import()
91 memset(sctx->state, 0, sizeof(sctx->state)); in sha3_224_init()
92 sctx->count = 0; in sha3_224_init()
[all …]
/linux/arch/arm64/crypto/
H A Dsha3-ce-glue.c43 if ((sctx->partial + len) >= sctx->rsiz) { in sha3_update()
47 int p = sctx->rsiz - sctx->partial; in sha3_update()
49 memcpy(sctx->buf + sctx->partial, data, p); in sha3_update()
51 sha3_ce_transform(sctx->st, sctx->buf, 1, digest_size); in sha3_update()
60 len %= sctx->rsiz; in sha3_update()
75 memcpy(sctx->buf + sctx->partial, data, len); in sha3_update()
91 sctx->buf[sctx->partial++] = 0x06; in sha3_final()
92 memset(sctx->buf + sctx->partial, 0, sctx->rsiz - sctx->partial); in sha3_final()
93 sctx->buf[sctx->rsiz - 1] |= 0x80; in sha3_final()
96 sha3_ce_transform(sctx->st, sctx->buf, 1, digest_size); in sha3_final()
[all …]
/linux/arch/sparc/crypto/
H A Dsha1_glue.c35 sctx->count += len; in __sha1_sparc64_update()
38 memcpy(sctx->buffer + partial, data, done); in __sha1_sparc64_update()
39 sha1_sparc64_transform(sctx->state, sctx->buffer, 1); in __sha1_sparc64_update()
59 sctx->count += len; in sha1_sparc64_update()
76 bits = cpu_to_be64(sctx->count << 3); in sha1_sparc64_final()
79 index = sctx->count % SHA1_BLOCK_SIZE; in sha1_sparc64_final()
84 sctx->count += padlen; in sha1_sparc64_final()
93 dst[i] = cpu_to_be32(sctx->state[i]); in sha1_sparc64_final()
96 memset(sctx, 0, sizeof(*sctx)); in sha1_sparc64_final()
105 memcpy(out, sctx, sizeof(*sctx)); in sha1_sparc64_export()
[all …]
H A Dmd5_glue.c51 sctx->byte_count += len; in __md5_sparc64_update()
55 md5_sparc64_transform(sctx->hash, (u8 *)sctx->block, 1); in __md5_sparc64_update()
64 memcpy(sctx->block, data + done, len - done); in __md5_sparc64_update()
70 struct md5_state *sctx = shash_desc_ctx(desc); in md5_sparc64_update() local
75 sctx->byte_count += len; in md5_sparc64_update()
92 bits = cpu_to_le64(sctx->byte_count << 3); in md5_sparc64_final()
100 sctx->byte_count += padlen; in md5_sparc64_final()
109 dst[i] = sctx->hash[i]; in md5_sparc64_final()
112 memset(sctx, 0, sizeof(*sctx)); in md5_sparc64_final()
121 memcpy(out, sctx, sizeof(*sctx)); in md5_sparc64_export()
[all …]
H A Dsha256_glue.c35 sctx->count += len; in __sha256_sparc64_update()
38 memcpy(sctx->buf + partial, data, done); in __sha256_sparc64_update()
39 sha256_sparc64_transform(sctx->state, sctx->buf, 1); in __sha256_sparc64_update()
59 sctx->count += len; in sha256_sparc64_update()
60 memcpy(sctx->buf + partial, data, len); in sha256_sparc64_update()
75 bits = cpu_to_be64(sctx->count << 3); in sha256_sparc64_final()
83 sctx->count += padlen; in sha256_sparc64_final()
92 dst[i] = cpu_to_be32(sctx->state[i]); in sha256_sparc64_final()
95 memset(sctx, 0, sizeof(*sctx)); in sha256_sparc64_final()
116 memcpy(out, sctx, sizeof(*sctx)); in sha256_sparc64_export()
[all …]
H A Dsha512_glue.c34 if ((sctx->count[0] += len) < len) in __sha512_sparc64_update()
35 sctx->count[1]++; in __sha512_sparc64_update()
38 memcpy(sctx->buf + partial, data, done); in __sha512_sparc64_update()
39 sha512_sparc64_transform(sctx->state, sctx->buf, 1); in __sha512_sparc64_update()
59 if ((sctx->count[0] += len) < len) in sha512_sparc64_update()
60 sctx->count[1]++; in sha512_sparc64_update()
61 memcpy(sctx->buf + partial, data, len); in sha512_sparc64_update()
78 bits[0] = cpu_to_be64(sctx->count[1] << 3 | sctx->count[0] >> 61); in sha512_sparc64_final()
87 sctx->count[1]++; in sha512_sparc64_final()
96 dst[i] = cpu_to_be64(sctx->state[i]); in sha512_sparc64_final()
[all …]
/linux/crypto/
H A Dsha3_generic.c167 sctx->rsizw = sctx->rsiz / 8; in crypto_sha3_init()
168 sctx->partial = 0; in crypto_sha3_init()
170 memset(sctx->st, 0, sizeof(sctx->st)); in crypto_sha3_init()
185 if ((sctx->partial + len) > (sctx->rsiz - 1)) { in crypto_sha3_update()
188 memcpy(sctx->buf + sctx->partial, data, in crypto_sha3_update()
190 src = sctx->buf; in crypto_sha3_update()
206 memcpy(sctx->buf + sctx->partial, src, len - done); in crypto_sha3_update()
221 memset(sctx->buf + inlen, 0, sctx->rsiz - inlen); in crypto_sha3_final()
222 sctx->buf[sctx->rsiz - 1] |= 0x80; in crypto_sha3_final()
225 sctx->st[i] ^= get_unaligned_le64(sctx->buf + 8 * i); in crypto_sha3_final()
[all …]
/linux/drivers/crypto/nx/
H A Dnx-sha256.c46 memset(sctx, 0, sizeof *sctx); in nx_sha256_init()
48 sctx->state[0] = __cpu_to_be32(SHA256_H0); in nx_sha256_init()
49 sctx->state[1] = __cpu_to_be32(SHA256_H1); in nx_sha256_init()
56 sctx->count = 0; in nx_sha256_init()
83 memcpy(sctx->buf + buf_len, data, len); in nx_sha256_update()
84 sctx->count += len; in nx_sha256_update()
114 (u8 *) sctx->buf, in nx_sha256_update()
172 memcpy(sctx->buf, data, leftover); in nx_sha256_update()
174 sctx->count += len; in nx_sha256_update()
255 memcpy(out, sctx, sizeof(*sctx)); in nx_sha256_export()
[all …]
H A Dnx-sha512.c46 memset(sctx, 0, sizeof *sctx); in nx_sha512_init()
48 sctx->state[0] = __cpu_to_be64(SHA512_H0); in nx_sha512_init()
56 sctx->count[0] = 0; in nx_sha512_init()
83 memcpy(sctx->buf + buf_len, data, len); in nx_sha512_update()
84 sctx->count[0] += len; in nx_sha512_update()
114 (u8 *) sctx->buf, in nx_sha512_update()
176 memcpy(sctx->buf, data, leftover); in nx_sha512_update()
177 sctx->count[0] += len; in nx_sha512_update()
219 count0 = sctx->count[0] * 8; in nx_sha512_final()
261 memcpy(out, sctx, sizeof(*sctx)); in nx_sha512_export()
[all …]

123