/linux/fs/ext4/ |
H A D | indirect.c | 557 if (!partial) { in ext4_ind_map_blocks() 636 offsets + (partial - chain), partial); in ext4_ind_map_blocks() 675 partial--; in ext4_ind_map_blocks() 817 if (!partial) in ext4_find_shared() 823 if (!partial->key && *partial->p) in ext4_find_shared() 847 partial--; in ext4_find_shared() 1183 partial->p+1, (chain+n-1) - partial); in ext4_ind_truncate() 1193 partial--; in ext4_ind_truncate() 1297 partial->p+1, (chain+n-1) - partial); in ext4_ind_remove_space() 1310 partial--; in ext4_ind_remove_space() [all …]
|
/linux/fs/minix/ |
H A D | itree_common.c | 158 Indirect *partial; in get_block() local 169 if (!partial) { in get_block() 182 partial--; in get_block() 197 err = alloc_branch(inode, left, offsets+(partial-chain), partial); in get_block() 210 partial--; in get_block() 238 if (!partial) in find_shared() 240 if (!partial->key && *partial->p) { in find_shared() 257 partial--; in find_shared() 260 return partial; in find_shared() 336 free_branches(inode, partial->p + 1, block_end(partial->bh), in truncate() [all …]
|
/linux/include/crypto/ |
H A D | sha1_base.h | 45 if (unlikely((partial + len) >= SHA1_BLOCK_SIZE)) { in sha1_base_do_update() 48 if (partial) { in sha1_base_do_update() 49 int p = SHA1_BLOCK_SIZE - partial; in sha1_base_do_update() 51 memcpy(sctx->buffer + partial, data, p); in sha1_base_do_update() 65 partial = 0; in sha1_base_do_update() 68 memcpy(sctx->buffer + partial, data, len); in sha1_base_do_update() 81 sctx->buffer[partial++] = 0x80; in sha1_base_do_finalize() 82 if (partial > bit_offset) { in sha1_base_do_finalize() 83 memset(sctx->buffer + partial, 0x0, SHA1_BLOCK_SIZE - partial); in sha1_base_do_finalize() 84 partial = 0; in sha1_base_do_finalize() [all …]
|
H A D | sm3_base.h | 48 if (unlikely((partial + len) >= SM3_BLOCK_SIZE)) { in sm3_base_do_update() 51 if (partial) { in sm3_base_do_update() 52 int p = SM3_BLOCK_SIZE - partial; in sm3_base_do_update() 54 memcpy(sctx->buffer + partial, data, p); in sm3_base_do_update() 68 partial = 0; in sm3_base_do_update() 71 memcpy(sctx->buffer + partial, data, len); in sm3_base_do_update() 84 sctx->buffer[partial++] = 0x80; in sm3_base_do_finalize() 85 if (partial > bit_offset) { in sm3_base_do_finalize() 86 memset(sctx->buffer + partial, 0x0, SM3_BLOCK_SIZE - partial); in sm3_base_do_finalize() 87 partial = 0; in sm3_base_do_finalize() [all …]
|
H A D | sha256_base.h | 46 if (unlikely((partial + len) >= SHA256_BLOCK_SIZE)) { in lib_sha256_base_do_update() 49 if (partial) { in lib_sha256_base_do_update() 50 int p = SHA256_BLOCK_SIZE - partial; in lib_sha256_base_do_update() 52 memcpy(sctx->buf + partial, data, p); in lib_sha256_base_do_update() 66 partial = 0; in lib_sha256_base_do_update() 69 memcpy(sctx->buf + partial, data, len); in lib_sha256_base_do_update() 91 sctx->buf[partial++] = 0x80; in lib_sha256_base_do_finalize() 92 if (partial > bit_offset) { in lib_sha256_base_do_finalize() 93 memset(sctx->buf + partial, 0x0, SHA256_BLOCK_SIZE - partial); in lib_sha256_base_do_finalize() 94 partial = 0; in lib_sha256_base_do_finalize() [all …]
|
H A D | sha512_base.h | 68 if (unlikely((partial + len) >= SHA512_BLOCK_SIZE)) { in sha512_base_do_update() 71 if (partial) { in sha512_base_do_update() 72 int p = SHA512_BLOCK_SIZE - partial; in sha512_base_do_update() 74 memcpy(sctx->buf + partial, data, p); in sha512_base_do_update() 88 partial = 0; in sha512_base_do_update() 91 memcpy(sctx->buf + partial, data, len); in sha512_base_do_update() 104 sctx->buf[partial++] = 0x80; in sha512_base_do_finalize() 105 if (partial > bit_offset) { in sha512_base_do_finalize() 106 memset(sctx->buf + partial, 0x0, SHA512_BLOCK_SIZE - partial); in sha512_base_do_finalize() 107 partial = 0; in sha512_base_do_finalize() [all …]
|
/linux/fs/sysv/ |
H A D | itree.c | 214 Indirect *partial; in get_block() local 225 if (!partial) { in get_block() 239 partial--; in get_block() 254 err = alloc_branch(inode, left, offsets+(partial-chain), partial); in get_block() 267 partial--; in get_block() 295 if (!partial) in find_shared() 301 if (!partial->key && *partial->p) { in find_shared() 323 partial--; in find_shared() 326 return partial; in find_shared() 408 free_branches(inode, partial->p + 1, block_end(partial->bh), in sysv_truncate() [all …]
|
/linux/drivers/crypto/ |
H A D | padlock-sha.c | 284 unsigned int partial, done; in padlock_sha1_update_nano() local 291 partial = sctx->count & 0x3f; in padlock_sha1_update_nano() 300 if (partial) { in padlock_sha1_update_nano() 301 done = -partial; in padlock_sha1_update_nano() 321 partial = 0; in padlock_sha1_update_nano() 340 padlen = (partial < 56) ? (56 - partial) : ((64+56) - partial); in padlock_sha1_final_nano() 368 unsigned int partial, done; in padlock_sha256_update_nano() local 384 if (partial) { in padlock_sha256_update_nano() 385 done = -partial; in padlock_sha256_update_nano() 405 partial = 0; in padlock_sha256_update_nano() [all …]
|
/linux/arch/arm64/crypto/ |
H A D | sha3-ce-glue.c | 43 if ((sctx->partial + len) >= sctx->rsiz) { in sha3_update() 46 if (sctx->partial) { in sha3_update() 47 int p = sctx->rsiz - sctx->partial; in sha3_update() 49 memcpy(sctx->buf + sctx->partial, data, p); in sha3_update() 56 sctx->partial = 0; in sha3_update() 75 memcpy(sctx->buf + sctx->partial, data, len); in sha3_update() 76 sctx->partial += len; in sha3_update() 91 sctx->buf[sctx->partial++] = 0x06; in sha3_final() 92 memset(sctx->buf + sctx->partial, 0, sctx->rsiz - sctx->partial); in sha3_final()
|
/linux/arch/powerpc/crypto/ |
H A D | sha1.c | 30 unsigned int partial, done; in powerpc_sha1_update() local 33 partial = sctx->count & 0x3f; in powerpc_sha1_update() 38 if ((partial + len) > 63) { in powerpc_sha1_update() 40 if (partial) { in powerpc_sha1_update() 41 done = -partial; in powerpc_sha1_update() 42 memcpy(sctx->buffer + partial, data, done + 64); in powerpc_sha1_update() 52 partial = 0; in powerpc_sha1_update() 54 memcpy(sctx->buffer + partial, src, len - done); in powerpc_sha1_update()
|
H A D | aes-gcm-p10.S | 637 # check partial block 927 # check partial block 929 ld 15, 56(7) # partial? 1029 # store partial block 1045 # write partial 1057 # Complete loop partial 1076 std 15, 56(7) # partial 1082 # Write partial block 1178 # check partial block 1478 # check partial block [all …]
|
/linux/fs/ext2/ |
H A D | inode.c | 632 Indirect *partial; in ext2_get_blocks() local 650 if (!partial) { in ext2_get_blocks() 699 partial--; in ext2_get_blocks() 702 if (!partial) { in ext2_get_blocks() 735 offsets + (partial - chain), partial); in ext2_get_blocks() 776 partial--; in ext2_get_blocks() 1039 if (!partial) in ext2_find_shared() 1046 if (!partial->key && *partial->p) { in ext2_find_shared() 1069 partial--; in ext2_find_shared() 1072 return partial; in ext2_find_shared() [all …]
|
/linux/arch/sparc/crypto/ |
H A D | sha1_glue.c | 31 unsigned int len, unsigned int partial) in __sha1_sparc64_update() argument 36 if (partial) { in __sha1_sparc64_update() 37 done = SHA1_BLOCK_SIZE - partial; in __sha1_sparc64_update() 38 memcpy(sctx->buffer + partial, data, done); in __sha1_sparc64_update() 55 unsigned int partial = sctx->count % SHA1_BLOCK_SIZE; in sha1_sparc64_update() local 58 if (partial + len < SHA1_BLOCK_SIZE) { in sha1_sparc64_update() 60 memcpy(sctx->buffer + partial, data, len); in sha1_sparc64_update() 62 __sha1_sparc64_update(sctx, data, len, partial); in sha1_sparc64_update()
|
H A D | md5_glue.c | 47 unsigned int len, unsigned int partial) in __md5_sparc64_update() argument 52 if (partial) { in __md5_sparc64_update() 53 done = MD5_HMAC_BLOCK_SIZE - partial; in __md5_sparc64_update() 54 memcpy((u8 *)sctx->block + partial, data, done); in __md5_sparc64_update() 71 unsigned int partial = sctx->byte_count % MD5_HMAC_BLOCK_SIZE; in md5_sparc64_update() local 74 if (partial + len < MD5_HMAC_BLOCK_SIZE) { in md5_sparc64_update() 76 memcpy((u8 *)sctx->block + partial, data, len); in md5_sparc64_update() 78 __md5_sparc64_update(sctx, data, len, partial); in md5_sparc64_update()
|
H A D | sha512_glue.c | 30 unsigned int len, unsigned int partial) in __sha512_sparc64_update() argument 36 if (partial) { in __sha512_sparc64_update() 37 done = SHA512_BLOCK_SIZE - partial; in __sha512_sparc64_update() 38 memcpy(sctx->buf + partial, data, done); in __sha512_sparc64_update() 55 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_sparc64_update() local 58 if (partial + len < SHA512_BLOCK_SIZE) { in sha512_sparc64_update() 61 memcpy(sctx->buf + partial, data, len); in sha512_sparc64_update() 63 __sha512_sparc64_update(sctx, data, len, partial); in sha512_sparc64_update()
|
H A D | sha256_glue.c | 31 unsigned int len, unsigned int partial) in __sha256_sparc64_update() argument 36 if (partial) { in __sha256_sparc64_update() 37 done = SHA256_BLOCK_SIZE - partial; in __sha256_sparc64_update() 38 memcpy(sctx->buf + partial, data, done); in __sha256_sparc64_update() 55 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in sha256_sparc64_update() local 58 if (partial + len < SHA256_BLOCK_SIZE) { in sha256_sparc64_update() 60 memcpy(sctx->buf + partial, data, len); in sha256_sparc64_update() 62 __sha256_sparc64_update(sctx, data, len, partial); in sha256_sparc64_update()
|
/linux/arch/mips/cavium-octeon/crypto/ |
H A D | octeon-sha1.c | 78 unsigned int partial; in __octeon_sha1_update() local 82 partial = sctx->count % SHA1_BLOCK_SIZE; in __octeon_sha1_update() 87 if ((partial + len) >= SHA1_BLOCK_SIZE) { in __octeon_sha1_update() 88 if (partial) { in __octeon_sha1_update() 89 done = -partial; in __octeon_sha1_update() 90 memcpy(sctx->buffer + partial, data, in __octeon_sha1_update() 101 partial = 0; in __octeon_sha1_update() 103 memcpy(sctx->buffer + partial, src, len - done); in __octeon_sha1_update()
|
H A D | octeon-sha256.c | 70 unsigned int partial; in __octeon_sha256_update() local 74 partial = sctx->count % SHA256_BLOCK_SIZE; in __octeon_sha256_update() 79 if ((partial + len) >= SHA256_BLOCK_SIZE) { in __octeon_sha256_update() 80 if (partial) { in __octeon_sha256_update() 81 done = -partial; in __octeon_sha256_update() 82 memcpy(sctx->buf + partial, data, in __octeon_sha256_update() 93 partial = 0; in __octeon_sha256_update() 95 memcpy(sctx->buf + partial, src, len - done); in __octeon_sha256_update()
|
/linux/drivers/usb/storage/ |
H A D | freecom.c | 227 unsigned int partial; in freecom_transport() local 266 FCM_STATUS_PACKET_LENGTH, &partial); in freecom_transport() 271 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport() 308 FCM_STATUS_PACKET_LENGTH, &partial); in freecom_transport() 314 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport() 317 if (partial != 4) in freecom_transport() 376 FCM_PACKET_LENGTH, &partial); in freecom_transport() 377 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport() 379 if (partial != 4 || result > USB_STOR_XFER_SHORT) in freecom_transport() 407 FCM_PACKET_LENGTH, &partial); in freecom_transport() [all …]
|
/linux/net/smc/ |
H A D | smc_rx.c | 154 struct partial_page *partial; in smc_rx_splice() local 167 partial = kcalloc(nr_pages, sizeof(*partial), GFP_KERNEL); in smc_rx_splice() 168 if (!partial) in smc_rx_splice() 185 partial[0].len = len; in smc_rx_splice() 186 partial[0].private = (unsigned long)priv[0]; in smc_rx_splice() 197 partial[i].offset = offset; in smc_rx_splice() 198 partial[i].len = size; in smc_rx_splice() 199 partial[i].private = (unsigned long)priv[i]; in smc_rx_splice() 208 spd.partial = partial; in smc_rx_splice() 224 kfree(partial); in smc_rx_splice() [all …]
|
/linux/arch/x86/include/asm/ |
H A D | unwind.h | 72 bool *partial) in unwind_get_entry_regs() argument 77 if (partial) { in unwind_get_entry_regs() 79 *partial = !state->full_regs; in unwind_get_entry_regs() 81 *partial = false; in unwind_get_entry_regs() 89 bool *partial) in unwind_get_entry_regs() argument
|
/linux/crypto/ |
H A D | sha3_generic.c | 168 sctx->partial = 0; in crypto_sha3_init() 185 if ((sctx->partial + len) > (sctx->rsiz - 1)) { in crypto_sha3_update() 186 if (sctx->partial) { in crypto_sha3_update() 187 done = -sctx->partial; in crypto_sha3_update() 188 memcpy(sctx->buf + sctx->partial, data, in crypto_sha3_update() 204 sctx->partial = 0; in crypto_sha3_update() 206 memcpy(sctx->buf + sctx->partial, src, len - done); in crypto_sha3_update() 207 sctx->partial += (len - done); in crypto_sha3_update() 216 unsigned int i, inlen = sctx->partial; in crypto_sha3_final()
|
H A D | sm3.c | 185 if ((partial + len) >= SM3_BLOCK_SIZE) { in sm3_update() 188 if (partial) { in sm3_update() 189 int p = SM3_BLOCK_SIZE - partial; in sm3_update() 191 memcpy(sctx->buffer + partial, data, p); in sm3_update() 208 partial = 0; in sm3_update() 211 memcpy(sctx->buffer + partial, data, len); in sm3_update() 224 sctx->buffer[partial++] = 0x80; in sm3_final() 225 if (partial > bit_offset) { in sm3_final() 226 memset(sctx->buffer + partial, 0, SM3_BLOCK_SIZE - partial); in sm3_final() 227 partial = 0; in sm3_final() [all …]
|
/linux/arch/x86/kernel/ |
H A D | dumpstack.c | 155 bool partial, const char *log_lvl) in show_regs_if_on_stack() argument 166 if (!partial && on_stack(info, regs, sizeof(*regs))) { in show_regs_if_on_stack() 169 } else if (partial && on_stack(info, (void *)regs + IRET_FRAME_OFFSET, in show_regs_if_on_stack() 193 bool partial = false; in show_trace_log_lvl() local 198 regs = unwind_get_entry_regs(&state, &partial); in show_trace_log_lvl() 240 show_regs_if_on_stack(&stack_info, regs, partial, log_lvl); in show_trace_log_lvl() 299 regs = unwind_get_entry_regs(&state, &partial); in show_trace_log_lvl() 301 show_regs_if_on_stack(&stack_info, regs, partial, log_lvl); in show_trace_log_lvl()
|
/linux/drivers/crypto/stm32/ |
H A D | stm32-crc32.c | 69 u32 partial; /* crc32c: partial in first 4 bytes of that struct */ member 137 ctx->partial = readl_relaxed(crc->regs + CRC_DR); in stm32_crc_init() 163 ctx->partial = crc32_le(ctx->partial, d8, length); in burst_update() 165 ctx->partial = __crc32c_le(ctx->partial, d8, length); in burst_update() 176 writel_relaxed(bitrev32(ctx->partial), crc->regs + CRC_INIT); in burst_update() 206 ctx->partial = readl_relaxed(crc->regs + CRC_DR); in burst_update() 249 ~ctx->partial : ctx->partial, out); in stm32_crc_final()
|