/linux/drivers/crypto/allwinner/sun4i-ss/ |
H A D | sun4i-ss-cipher.c | 44 if (!areq->src || !areq->dst) { in sun4i_ss_opti_poll() 65 if (areq->iv) { in sun4i_ss_opti_poll() 80 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_ss_opti_poll() 108 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), in sun4i_ss_opti_poll() 137 scatterwalk_map_and_copy(areq->iv, areq->dst, areq->cryptlen - ivsize, in sun4i_ss_opti_poll() 164 areq->base.complete, areq->base.data); in sun4i_ss_cipher_poll_fallback() 166 areq->cryptlen, areq->iv); in sun4i_ss_cipher_poll_fallback() 212 if (!areq->src || !areq->dst) { in sun4i_ss_cipher_poll() 272 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_ss_cipher_poll() 326 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), in sun4i_ss_cipher_poll() [all …]
|
H A D | sun4i-ss-hash.c | 91 sun4i_hash_init(areq); in sun4i_hash_import_md5() 134 sun4i_hash_init(areq); in sun4i_hash_import_sha1() 220 copied = sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in sun4i_hash() 247 if (end > areq->nbytes || areq->nbytes - end > 63) { in sun4i_hash() 249 end, areq->nbytes); in sun4i_hash() 255 if (areq->nbytes < 4) in sun4i_hash() 272 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_hash() 513 return sun4i_hash(areq); in sun4i_hash_final() 521 return sun4i_hash(areq); in sun4i_hash_update() 530 return sun4i_hash(areq); in sun4i_hash_finup() [all …]
|
H A D | sun4i-ss.h | 201 int sun4i_hash_init(struct ahash_request *areq); 202 int sun4i_hash_update(struct ahash_request *areq); 203 int sun4i_hash_final(struct ahash_request *areq); 204 int sun4i_hash_finup(struct ahash_request *areq); 205 int sun4i_hash_digest(struct ahash_request *areq); 211 int sun4i_ss_cbc_aes_encrypt(struct skcipher_request *areq); 212 int sun4i_ss_cbc_aes_decrypt(struct skcipher_request *areq); 213 int sun4i_ss_ecb_aes_encrypt(struct skcipher_request *areq); 214 int sun4i_ss_ecb_aes_decrypt(struct skcipher_request *areq); 216 int sun4i_ss_cbc_des_encrypt(struct skcipher_request *areq); [all …]
|
/linux/drivers/crypto/allwinner/sun8i-ce/ |
H A D | sun8i-ce-cipher.c | 34 if (sg_nents_for_len(areq->src, areq->cryptlen) > MAX_SG || in sun8i_ce_cipher_need_fallback() 35 sg_nents_for_len(areq->dst, areq->cryptlen) > MAX_SG) { in sun8i_ce_cipher_need_fallback() 56 sg = areq->src; in sun8i_ce_cipher_need_fallback() 110 areq->base.complete, areq->base.data); in sun8i_ce_cipher_fallback() 112 areq->cryptlen, areq->iv); in sun8i_ce_cipher_fallback() 138 int ns = sg_nents_for_len(areq->src, areq->cryptlen); in sun8i_ce_cipher_prepare() 139 int nd = sg_nents_for_len(areq->dst, areq->cryptlen); in sun8i_ce_cipher_prepare() 214 if (areq->src == areq->dst) { in sun8i_ce_cipher_prepare() 273 if (areq->src == areq->dst) { in sun8i_ce_cipher_prepare() 290 scatterwalk_map_and_copy(areq->iv, areq->dst, offset, in sun8i_ce_cipher_prepare() [all …]
|
H A D | sun8i-ce-hash.c | 145 rctx->fallback_req.src = areq->src; in sun8i_ce_hash_update() 161 rctx->fallback_req.src = areq->src; in sun8i_ce_hash_finup() 215 if (areq->nbytes == 0) { in sun8i_ce_hash_need_fallback() 220 if (sg_nents_for_len(areq->src, areq->nbytes) > MAX_SG - 1) { in sun8i_ce_hash_need_fallback() 224 sg = areq->src; in sun8i_ce_hash_need_fallback() 253 nr_sgs = sg_nents_for_len(areq->src, areq->nbytes); in sun8i_ce_hash_digest() 352 int ns = sg_nents_for_len(areq->src, areq->nbytes); in sun8i_ce_hash_run() 384 dev_dbg(ce->dev, "%s %s len=%d\n", __func__, crypto_tfm_alg_name(areq->base.tfm), areq->nbytes); in sun8i_ce_hash_run() 404 len = areq->nbytes; in sun8i_ce_hash_run() 425 byte_count = areq->nbytes; in sun8i_ce_hash_run() [all …]
|
H A D | sun8i-ce.h | 352 int sun8i_ce_cipher_do_one(struct crypto_engine *engine, void *areq); 353 int sun8i_ce_skdecrypt(struct skcipher_request *areq); 354 int sun8i_ce_skencrypt(struct skcipher_request *areq); 362 int sun8i_ce_hash_init(struct ahash_request *areq); 363 int sun8i_ce_hash_export(struct ahash_request *areq, void *out); 364 int sun8i_ce_hash_import(struct ahash_request *areq, const void *in); 365 int sun8i_ce_hash_final(struct ahash_request *areq); 366 int sun8i_ce_hash_update(struct ahash_request *areq); 367 int sun8i_ce_hash_finup(struct ahash_request *areq); 368 int sun8i_ce_hash_digest(struct ahash_request *areq);
|
/linux/drivers/crypto/amlogic/ |
H A D | amlogic-gxl-cipher.c | 73 areq->base.complete, areq->base.data); in meson_cipher_do_fallback() 74 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in meson_cipher_do_fallback() 75 areq->cryptlen, areq->iv); in meson_cipher_do_fallback() 108 areq->cryptlen, in meson_cipher() 177 if (areq->src == areq->dst) { in meson_cipher() 178 nr_sgs = dma_map_sg(mc->dev, areq->src, sg_nents(areq->src), in meson_cipher() 187 nr_sgs = dma_map_sg(mc->dev, areq->src, sg_nents(areq->src), in meson_cipher() 194 nr_sgd = dma_map_sg(mc->dev, areq->dst, sg_nents(areq->dst), in meson_cipher() 203 src_sg = areq->src; in meson_cipher() 238 if (areq->src == areq->dst) { in meson_cipher() [all …]
|
H A D | amlogic-gxl.h | 153 int meson_enqueue(struct crypto_async_request *areq, u32 type); 159 int meson_skdecrypt(struct skcipher_request *areq); 160 int meson_skencrypt(struct skcipher_request *areq); 161 int meson_handle_cipher_request(struct crypto_engine *engine, void *areq);
|
/linux/drivers/crypto/gemini/ |
H A D | sl3516-ce-cipher.c | 34 if (areq->cryptlen == 0 || areq->cryptlen % 16) { in sl3516_ce_need_fallback() 53 sg = areq->src; in sl3516_ce_need_fallback() 69 sg = areq->dst; in sl3516_ce_need_fallback() 117 areq->base.complete, areq->base.data); in sl3516_ce_cipher_fallback() 119 areq->cryptlen, areq->iv); in sl3516_ce_cipher_fallback() 153 if (areq->src == areq->dst) { in sl3516_ce_cipher() 154 nr_sgs = dma_map_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher() 250 if (areq->src == areq->dst) { in sl3516_ce_cipher() 251 dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher() 254 dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher() [all …]
|
/linux/drivers/crypto/allwinner/sun8i-ss/ |
H A D | sun8i-ss-cipher.c | 33 if (areq->cryptlen == 0 || areq->cryptlen % 16) { in sun8i_ss_need_fallback() 38 if (sg_nents_for_len(areq->src, areq->cryptlen) > 8 || in sun8i_ss_need_fallback() 39 sg_nents_for_len(areq->dst, areq->cryptlen) > 8) { in sun8i_ss_need_fallback() 45 sg = areq->src; in sun8i_ss_need_fallback() 60 sg = areq->dst; in sun8i_ss_need_fallback() 110 areq->base.complete, areq->base.data); in sun8i_ss_cipher_fallback() 112 areq->cryptlen, areq->iv); in sun8i_ss_cipher_fallback() 197 int nsgs = sg_nents_for_len(areq->src, areq->cryptlen); in sun8i_ss_cipher() 198 int nsgd = sg_nents_for_len(areq->dst, areq->cryptlen); in sun8i_ss_cipher() 230 if (areq->src == areq->dst) { in sun8i_ss_cipher() [all …]
|
H A D | sun8i-ss-hash.c | 218 rctx->fallback_req.src = areq->src; in sun8i_ss_hash_update() 234 rctx->fallback_req.src = areq->src; in sun8i_ss_hash_finup() 263 rctx->fallback_req.src = areq->src; in sun8i_ss_hash_digest_fb() 349 if (areq->nbytes == 0) { in sun8i_ss_hash_need_fallback() 365 sg = areq->src; in sun8i_ss_hash_need_fallback() 520 nr_sgs = dma_map_sg(ss->dev, areq->src, sg_nents(areq->src), DMA_TO_DEVICE); in sun8i_ss_hash_run() 535 len = areq->nbytes; in sun8i_ss_hash_run() 536 sg = areq->src; in sun8i_ss_hash_run() 569 byte_count = areq->nbytes; in sun8i_ss_hash_run() 590 byte_count = areq->nbytes + bs; in sun8i_ss_hash_run() [all …]
|
H A D | sun8i-ss.h | 294 int sun8i_ss_handle_cipher_request(struct crypto_engine *engine, void *areq); 295 int sun8i_ss_skdecrypt(struct skcipher_request *areq); 296 int sun8i_ss_skencrypt(struct skcipher_request *areq); 309 int sun8i_ss_hash_init(struct ahash_request *areq); 310 int sun8i_ss_hash_export(struct ahash_request *areq, void *out); 311 int sun8i_ss_hash_import(struct ahash_request *areq, const void *in); 312 int sun8i_ss_hash_final(struct ahash_request *areq); 313 int sun8i_ss_hash_update(struct ahash_request *areq); 314 int sun8i_ss_hash_finup(struct ahash_request *areq); 315 int sun8i_ss_hash_digest(struct ahash_request *areq);
|
/linux/drivers/crypto/cavium/nitrox/ |
H A D | nitrox_aead.c | 229 rctx->srclen = areq->assoclen + areq->cryptlen; in nitrox_aes_gcm_enc() 235 rctx->src = areq->src; in nitrox_aes_gcm_enc() 244 areq); in nitrox_aes_gcm_enc() 263 rctx->srclen = areq->cryptlen + areq->assoclen; in nitrox_aes_gcm_dec() 278 areq); in nitrox_aes_gcm_dec() 400 if (areq->assoclen != 16 && areq->assoclen != 20) in nitrox_rfc4106_set_aead_rctx_sglist() 406 sg = scatterwalk_ffwd(rctx->src + 1, areq->src, areq->assoclen); in nitrox_rfc4106_set_aead_rctx_sglist() 410 if (areq->src != areq->dst) { in nitrox_rfc4106_set_aead_rctx_sglist() 413 sg = scatterwalk_ffwd(rctx->dst + 1, areq->dst, areq->assoclen); in nitrox_rfc4106_set_aead_rctx_sglist() 419 aead_rctx->dst = (areq->src == areq->dst) ? rctx->src : rctx->dst; in nitrox_rfc4106_set_aead_rctx_sglist() [all …]
|
/linux/drivers/crypto/inside-secure/ |
H A D | safexcel_hash.c | 285 areq->nbytes = 0; in safexcel_handle_req_result() 343 sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in safexcel_ahash_send_req() 425 areq->nbytes), in safexcel_ahash_send_req() 679 sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in safexcel_ahash_cache() 681 areq->nbytes, 0); in safexcel_ahash_cache() 748 if (!areq->nbytes) in safexcel_ahash_update() 799 memcpy(areq->result, in safexcel_ahash_final() 827 aes_encrypt(ctx->aes, areq->result, areq->result); in safexcel_ahash_final() 1131 if (!areq) { in __safexcel_hmac_setkey() 1930 return safexcel_crc32_init(areq) ?: safexcel_ahash_finup(areq); in safexcel_crc32_digest() [all …]
|
/linux/crypto/ |
H A D | algif_skcipher.c | 85 struct sock *sk = areq->sk; in algif_skcipher_done() 120 if (IS_ERR(areq)) in _skcipher_recvmsg() 121 return PTR_ERR(areq); in _skcipher_recvmsg() 142 if (!areq->tsgl_entries) in _skcipher_recvmsg() 143 areq->tsgl_entries = 1; in _skcipher_recvmsg() 144 areq->tsgl = sock_kmalloc(sk, array_size(sizeof(*areq->tsgl), in _skcipher_recvmsg() 145 areq->tsgl_entries), in _skcipher_recvmsg() 147 if (!areq->tsgl) { in _skcipher_recvmsg() 151 sg_init_table(areq->tsgl, areq->tsgl_entries); in _skcipher_recvmsg() 156 skcipher_request_set_crypt(&areq->cra_u.skcipher_req, areq->tsgl, in _skcipher_recvmsg() [all …]
|
H A D | algif_aead.c | 155 if (IS_ERR(areq)) in _aead_recvmsg() 156 return PTR_ERR(areq); in _aead_recvmsg() 254 if (!areq->tsgl_entries) in _aead_recvmsg() 255 areq->tsgl_entries = 1; in _aead_recvmsg() 256 areq->tsgl = sock_kmalloc(sk, array_size(sizeof(*areq->tsgl), in _aead_recvmsg() 257 areq->tsgl_entries), in _aead_recvmsg() 259 if (!areq->tsgl) { in _aead_recvmsg() 263 sg_init_table(areq->tsgl, areq->tsgl_entries); in _aead_recvmsg() 278 rsgl_src = areq->tsgl; in _aead_recvmsg() 290 areq->iocb = msg->msg_iocb; in _aead_recvmsg() [all …]
|
H A D | af_alg.c | 788 tsgl = areq->tsgl; in af_alg_free_areq_sgls() 1123 sock_kfree_s(sk, areq, areq->areqlen); in af_alg_free_resources() 1204 if (unlikely(!areq)) in af_alg_alloc_areq() 1209 areq->areqlen = areqlen; in af_alg_alloc_areq() 1210 areq->sk = sk; in af_alg_alloc_areq() 1211 areq->first_rsgl.sgl.sgt.sgl = areq->first_rsgl.sgl.sgl; in af_alg_alloc_areq() 1212 areq->last_rsgl = NULL; in af_alg_alloc_areq() 1214 areq->tsgl = NULL; in af_alg_alloc_areq() 1215 areq->tsgl_entries = 0; in af_alg_alloc_areq() 1217 return areq; in af_alg_alloc_areq() [all …]
|
/linux/drivers/crypto/intel/qat/qat_common/ |
H A D | qat_comp_algs.c | 78 areq->dlen = dlen; in qat_comp_resubmit() 130 areq->slen, areq->dlen, consumed, produced, cmp_err, xlt_err); in qat_comp_generic_callback() 132 areq->dlen = 0; in qat_comp_generic_callback() 171 areq->dlen = produced; in qat_comp_generic_callback() 245 if (!areq->src || !slen) in qat_comp_alg_compress_decompress() 248 if (areq->dst && !dlen) in qat_comp_alg_compress_decompress() 258 if (!areq->dst) { in qat_comp_alg_compress_decompress() 263 if (!areq->dst) in qat_comp_alg_compress_decompress() 267 areq->dlen = dlen; in qat_comp_alg_compress_decompress() 277 ret = qat_bl_sgl_to_bufl(ctx->inst->accel_dev, areq->src, areq->dst, in qat_comp_alg_compress_decompress() [all …]
|
/linux/drivers/crypto/ |
H A D | talitos.c | 1004 talitos_sg_unmap(dev, edesc, areq->src, areq->dst, in ipsec_esp_unmap() 1226 (areq->src == areq->dst) ? in ipsec_esp() 1264 if (areq->src != areq->dst) { in ipsec_esp() 1426 return talitos_edesc_alloc(ctx->dev, areq->src, areq->dst, in aead_edesc_alloc() 1538 talitos_sg_unmap(dev, edesc, areq->src, areq->dst, areq->cryptlen, 0); in common_nonsnoop_unmap() 1599 (areq->src == areq->dst) ? in common_nonsnoop() 1610 if (areq->src != areq->dst) { in common_nonsnoop() 1646 return talitos_edesc_alloc(ctx->dev, areq->src, areq->dst, in skcipher_edesc_alloc() 1647 areq->iv, 0, areq->cryptlen, 0, ivsize, 0, in skcipher_edesc_alloc() 2099 return ahash_process_req(areq, areq->nbytes); in ahash_update() [all …]
|
/linux/drivers/crypto/rockchip/ |
H A D | rk3288_crypto_ahash.c | 44 static int rk_ahash_digest_fb(struct ahash_request *areq) in rk_ahash_digest_fb() argument 46 struct rk_ahash_rctx *rctx = ahash_request_ctx(areq); in rk_ahash_digest_fb() 47 struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); in rk_ahash_digest_fb() 55 rctx->fallback_req.base.flags = areq->base.flags & in rk_ahash_digest_fb() 58 rctx->fallback_req.nbytes = areq->nbytes; in rk_ahash_digest_fb() 59 rctx->fallback_req.src = areq->src; in rk_ahash_digest_fb() 60 rctx->fallback_req.result = areq->result; in rk_ahash_digest_fb() 240 ret = dma_map_sg(rkc->dev, areq->src, sg_nents(areq->src), DMA_TO_DEVICE); in rk_hash_prepare() 265 struct scatterlist *sg = areq->src; in rk_hash_run() 299 rk_ahash_reg_init(areq, rkc); in rk_hash_run() [all …]
|
H A D | rk3288_crypto_skcipher.c | 81 areq->base.complete, areq->base.data); in rk_cipher_fallback() 82 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in rk_cipher_fallback() 83 areq->cryptlen, areq->iv); in rk_cipher_fallback() 310 u8 *ivtouse = areq->iv; in rk_cipher_run() 311 unsigned int len = areq->cryptlen; in rk_cipher_run() 333 sgs = areq->src; in rk_cipher_run() 334 sgd = areq->dst; in rk_cipher_run() 366 rk_cipher_hw_init(rkc, areq); in rk_cipher_run() 404 if (areq->iv && ivsize > 0) { in rk_cipher_run() 405 offset = areq->cryptlen - ivsize; in rk_cipher_run() [all …]
|
/linux/drivers/net/ethernet/fungible/funcore/ |
H A D | fun_dev.c | 225 const struct fun_dev_params *areq) in fun_enable_admin_queue() argument 228 .cqe_size_log2 = areq->cqe_size_log2, in fun_enable_admin_queue() 229 .sqe_size_log2 = areq->sqe_size_log2, in fun_enable_admin_queue() 230 .cq_depth = areq->cq_depth, in fun_enable_admin_queue() 231 .sq_depth = areq->sq_depth, in fun_enable_admin_queue() 232 .rq_depth = areq->rq_depth, in fun_enable_admin_queue() 234 unsigned int ntags = areq->sq_depth - 1; in fun_enable_admin_queue() 267 fdev->adminq_cb = areq->event_cb; in fun_enable_admin_queue() 276 rc = fun_enable_ctrl(fdev, areq->cqe_size_log2, areq->sqe_size_log2); in fun_enable_admin_queue() 280 if (areq->rq_depth) { in fun_enable_admin_queue() [all …]
|
/linux/drivers/crypto/marvell/octeontx2/ |
H A D | otx2_cptvf_reqmgr.c | 79 pentry->areq = NULL; in free_pentry() 99 gfp = (req->areq->flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? GFP_KERNEL : in process_request() 145 pentry->areq = req->areq; in process_request() 294 struct crypto_async_request *areq; in process_pending_queue() local 355 areq = resume_pentry->areq; in process_pending_queue() 364 callback(-EINPROGRESS, areq, info); in process_pending_queue() 370 areq = pentry->areq; in process_pending_queue() 383 callback(res_code, areq, info); in process_pending_queue()
|
/linux/drivers/crypto/xilinx/ |
H A D | zynqmp-aes-gcm.c | 207 struct aead_request *areq = in zynqmp_handle_aes_req() local 211 struct zynqmp_aead_req_ctx *rq_ctx = aead_request_ctx(areq); in zynqmp_handle_aes_req() 216 need_fallback = zynqmp_fallback_check(tfm_ctx, areq); in zynqmp_handle_aes_req() 221 aead_request_set_callback(subreq, areq->base.flags, in zynqmp_handle_aes_req() 223 aead_request_set_crypt(subreq, areq->src, areq->dst, in zynqmp_handle_aes_req() 224 areq->cryptlen, areq->iv); in zynqmp_handle_aes_req() 225 aead_request_set_ad(subreq, areq->assoclen); in zynqmp_handle_aes_req() 231 err = zynqmp_aes_aead_cipher(areq); in zynqmp_handle_aes_req() 235 crypto_finalize_aead_request(engine, areq, err); in zynqmp_handle_aes_req()
|
/linux/drivers/crypto/axis/ |
H A D | artpec6_crypto.c | 1439 sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in artpec6_crypto_prepare_hash() 1680 req_ctx = skcipher_request_ctx(areq); in artpec6_crypto_prepare_crypto() 1803 areq->cryptlen; in artpec6_crypto_prepare_crypto() 1908 input_length = areq->cryptlen; in artpec6_crypto_prepare_aead() 1914 __cpu_to_be64(8*areq->assoclen); in artpec6_crypto_prepare_aead() 1934 count = areq->assoclen; in artpec6_crypto_prepare_aead() 1972 size_t output_len = areq->cryptlen; in artpec6_crypto_prepare_aead() 2197 sg_pcopy_to_buffer(areq->src, in artpec6_crypto_complete_aead() 2198 sg_nents(areq->src), in artpec6_crypto_complete_aead() 2201 areq->assoclen + areq->cryptlen - in artpec6_crypto_complete_aead() [all …]
|