/dports/multimedia/libv4l/linux-5.13-rc2/drivers/crypto/rockchip/ |
H A D | rk3288_crypto.c | 128 sg_init_one(&dev->sg_tmp, dev->addr_vir, count); in rk_load_data() 129 if (!dma_map_sg(dev->dev, &dev->sg_tmp, 1, DMA_TO_DEVICE)) { in rk_load_data() 134 dev->addr_in = sg_dma_address(&dev->sg_tmp); in rk_load_data() 137 if (!dma_map_sg(dev->dev, &dev->sg_tmp, 1, in rk_load_data() 142 dma_unmap_sg(dev->dev, &dev->sg_tmp, 1, in rk_load_data() 146 dev->addr_out = sg_dma_address(&dev->sg_tmp); in rk_load_data() 157 sg_in = dev->aligned ? dev->sg_src : &dev->sg_tmp; in rk_unload_data() 161 sg_out = dev->aligned ? dev->sg_dst : &dev->sg_tmp; in rk_unload_data()
|
H A D | rk3288_crypto.h | 207 struct scatterlist sg_tmp; member
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/crypto/rockchip/ |
H A D | rk3288_crypto.c | 128 sg_init_one(&dev->sg_tmp, dev->addr_vir, count); in rk_load_data() 129 if (!dma_map_sg(dev->dev, &dev->sg_tmp, 1, DMA_TO_DEVICE)) { in rk_load_data() 134 dev->addr_in = sg_dma_address(&dev->sg_tmp); in rk_load_data() 137 if (!dma_map_sg(dev->dev, &dev->sg_tmp, 1, in rk_load_data() 142 dma_unmap_sg(dev->dev, &dev->sg_tmp, 1, in rk_load_data() 146 dev->addr_out = sg_dma_address(&dev->sg_tmp); in rk_load_data() 157 sg_in = dev->aligned ? dev->sg_src : &dev->sg_tmp; in rk_unload_data() 161 sg_out = dev->aligned ? dev->sg_dst : &dev->sg_tmp; in rk_unload_data()
|
H A D | rk3288_crypto.h | 207 struct scatterlist sg_tmp; member
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/crypto/rockchip/ |
H A D | rk3288_crypto.c | 128 sg_init_one(&dev->sg_tmp, dev->addr_vir, count); in rk_load_data() 129 if (!dma_map_sg(dev->dev, &dev->sg_tmp, 1, DMA_TO_DEVICE)) { in rk_load_data() 134 dev->addr_in = sg_dma_address(&dev->sg_tmp); in rk_load_data() 137 if (!dma_map_sg(dev->dev, &dev->sg_tmp, 1, in rk_load_data() 142 dma_unmap_sg(dev->dev, &dev->sg_tmp, 1, in rk_load_data() 146 dev->addr_out = sg_dma_address(&dev->sg_tmp); in rk_load_data() 157 sg_in = dev->aligned ? dev->sg_src : &dev->sg_tmp; in rk_unload_data() 161 sg_out = dev->aligned ? dev->sg_dst : &dev->sg_tmp; in rk_unload_data()
|
H A D | rk3288_crypto.h | 207 struct scatterlist sg_tmp; member
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/mmc/host/ |
H A D | renesas_sdhi_sys_dmac.c | 157 struct scatterlist *sg = host->sg_ptr, *sg_tmp; in renesas_sdhi_sys_dmac_start_dma_rx() local 165 for_each_sg(sg, sg_tmp, host->sg_len, i) { in renesas_sdhi_sys_dmac_start_dma_rx() 166 if (sg_tmp->offset & align) in renesas_sdhi_sys_dmac_start_dma_rx() 168 if (sg_tmp->length & align) { in renesas_sdhi_sys_dmac_start_dma_rx() 229 struct scatterlist *sg = host->sg_ptr, *sg_tmp; in renesas_sdhi_sys_dmac_start_dma_tx() local 237 for_each_sg(sg, sg_tmp, host->sg_len, i) { in renesas_sdhi_sys_dmac_start_dma_tx() 238 if (sg_tmp->offset & align) in renesas_sdhi_sys_dmac_start_dma_tx() 240 if (sg_tmp->length & align) { in renesas_sdhi_sys_dmac_start_dma_tx()
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/mmc/host/ |
H A D | renesas_sdhi_sys_dmac.c | 157 struct scatterlist *sg = host->sg_ptr, *sg_tmp; in renesas_sdhi_sys_dmac_start_dma_rx() local 165 for_each_sg(sg, sg_tmp, host->sg_len, i) { in renesas_sdhi_sys_dmac_start_dma_rx() 166 if (sg_tmp->offset & align) in renesas_sdhi_sys_dmac_start_dma_rx() 168 if (sg_tmp->length & align) { in renesas_sdhi_sys_dmac_start_dma_rx() 229 struct scatterlist *sg = host->sg_ptr, *sg_tmp; in renesas_sdhi_sys_dmac_start_dma_tx() local 237 for_each_sg(sg, sg_tmp, host->sg_len, i) { in renesas_sdhi_sys_dmac_start_dma_tx() 238 if (sg_tmp->offset & align) in renesas_sdhi_sys_dmac_start_dma_tx() 240 if (sg_tmp->length & align) { in renesas_sdhi_sys_dmac_start_dma_tx()
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/mmc/host/ |
H A D | renesas_sdhi_sys_dmac.c | 157 struct scatterlist *sg = host->sg_ptr, *sg_tmp; in renesas_sdhi_sys_dmac_start_dma_rx() local 165 for_each_sg(sg, sg_tmp, host->sg_len, i) { in renesas_sdhi_sys_dmac_start_dma_rx() 166 if (sg_tmp->offset & align) in renesas_sdhi_sys_dmac_start_dma_rx() 168 if (sg_tmp->length & align) { in renesas_sdhi_sys_dmac_start_dma_rx() 229 struct scatterlist *sg = host->sg_ptr, *sg_tmp; in renesas_sdhi_sys_dmac_start_dma_tx() local 237 for_each_sg(sg, sg_tmp, host->sg_len, i) { in renesas_sdhi_sys_dmac_start_dma_tx() 238 if (sg_tmp->offset & align) in renesas_sdhi_sys_dmac_start_dma_tx() 240 if (sg_tmp->length & align) { in renesas_sdhi_sys_dmac_start_dma_tx()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/memstick/core/ |
H A D | ms_block.c | 1603 struct scatterlist sg_tmp[10]; in msb_cache_write() local 1630 sg_init_table(sg_tmp, ARRAY_SIZE(sg_tmp)); in msb_cache_write() 1631 msb_sg_copy(sg, sg_tmp, ARRAY_SIZE(sg_tmp), offset, msb->page_size); in msb_cache_write() 1633 sg_copy_to_buffer(sg_tmp, sg_nents(sg_tmp), in msb_cache_write() 1644 struct scatterlist sg_tmp[10]; in msb_cache_read() local 1653 sg_init_table(sg_tmp, ARRAY_SIZE(sg_tmp)); in msb_cache_read() 1654 msb_sg_copy(sg, sg_tmp, ARRAY_SIZE(sg_tmp), in msb_cache_read() 1656 sg_copy_from_buffer(sg_tmp, sg_nents(sg_tmp), in msb_cache_read()
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/memstick/core/ |
H A D | ms_block.c | 1603 struct scatterlist sg_tmp[10]; in msb_cache_write() local 1630 sg_init_table(sg_tmp, ARRAY_SIZE(sg_tmp)); in msb_cache_write() 1631 msb_sg_copy(sg, sg_tmp, ARRAY_SIZE(sg_tmp), offset, msb->page_size); in msb_cache_write() 1633 sg_copy_to_buffer(sg_tmp, sg_nents(sg_tmp), in msb_cache_write() 1644 struct scatterlist sg_tmp[10]; in msb_cache_read() local 1653 sg_init_table(sg_tmp, ARRAY_SIZE(sg_tmp)); in msb_cache_read() 1654 msb_sg_copy(sg, sg_tmp, ARRAY_SIZE(sg_tmp), in msb_cache_read() 1656 sg_copy_from_buffer(sg_tmp, sg_nents(sg_tmp), in msb_cache_read()
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/memstick/core/ |
H A D | ms_block.c | 1603 struct scatterlist sg_tmp[10]; in msb_cache_write() local 1630 sg_init_table(sg_tmp, ARRAY_SIZE(sg_tmp)); in msb_cache_write() 1631 msb_sg_copy(sg, sg_tmp, ARRAY_SIZE(sg_tmp), offset, msb->page_size); in msb_cache_write() 1633 sg_copy_to_buffer(sg_tmp, sg_nents(sg_tmp), in msb_cache_write() 1644 struct scatterlist sg_tmp[10]; in msb_cache_read() local 1653 sg_init_table(sg_tmp, ARRAY_SIZE(sg_tmp)); in msb_cache_read() 1654 msb_sg_copy(sg, sg_tmp, ARRAY_SIZE(sg_tmp), in msb_cache_read() 1656 sg_copy_from_buffer(sg_tmp, sg_nents(sg_tmp), in msb_cache_read()
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/crypto/keembay/ |
H A D | ocs-aes.c | 1424 struct scatterlist *sg_tmp; in ocs_create_linked_list_from_sg() local 1453 sg_tmp = sg; in ocs_create_linked_list_from_sg() 1456 if (!sg_tmp) in ocs_create_linked_list_from_sg() 1458 tmp += sg_dma_len(sg_tmp); in ocs_create_linked_list_from_sg() 1460 sg_tmp = sg_next(sg_tmp); in ocs_create_linked_list_from_sg()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/crypto/keembay/ |
H A D | ocs-aes.c | 1424 struct scatterlist *sg_tmp; in ocs_create_linked_list_from_sg() local 1453 sg_tmp = sg; in ocs_create_linked_list_from_sg() 1456 if (!sg_tmp) in ocs_create_linked_list_from_sg() 1458 tmp += sg_dma_len(sg_tmp); in ocs_create_linked_list_from_sg() 1460 sg_tmp = sg_next(sg_tmp); in ocs_create_linked_list_from_sg()
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/crypto/keembay/ |
H A D | ocs-aes.c | 1424 struct scatterlist *sg_tmp; in ocs_create_linked_list_from_sg() local 1453 sg_tmp = sg; in ocs_create_linked_list_from_sg() 1456 if (!sg_tmp) in ocs_create_linked_list_from_sg() 1458 tmp += sg_dma_len(sg_tmp); in ocs_create_linked_list_from_sg() 1460 sg_tmp = sg_next(sg_tmp); in ocs_create_linked_list_from_sg()
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/crypto/ |
H A D | s5p-sss.c | 1145 struct scatterlist *sg_tmp = sg; in s5p_hash_prepare_sgs() local 1153 while (nbytes > 0 && sg_tmp) { in s5p_hash_prepare_sgs() 1155 if (skip >= sg_tmp->length) { in s5p_hash_prepare_sgs() 1156 skip -= sg_tmp->length; in s5p_hash_prepare_sgs() 1157 if (!sg_tmp->length) { in s5p_hash_prepare_sgs() 1162 if (!IS_ALIGNED(sg_tmp->length - skip, BUFLEN)) { in s5p_hash_prepare_sgs() 1167 if (nbytes < sg_tmp->length - skip) { in s5p_hash_prepare_sgs() 1172 nbytes -= sg_tmp->length - skip; in s5p_hash_prepare_sgs() 1176 sg_tmp = sg_next(sg_tmp); in s5p_hash_prepare_sgs()
|
H A D | omap-sham.c | 754 struct scatterlist *sg_tmp = sg; in omap_sham_align_sgs() local 787 while (nbytes > 0 && sg_tmp) { in omap_sham_align_sgs() 804 if (page_zonenum(sg_page(sg_tmp)) != ZONE_DMA) { in omap_sham_align_sgs() 810 if (offset < sg_tmp->length) { in omap_sham_align_sgs() 811 if (!IS_ALIGNED(offset + sg_tmp->offset, 4)) { in omap_sham_align_sgs() 816 if (!IS_ALIGNED(sg_tmp->length - offset, bs)) { in omap_sham_align_sgs() 823 offset -= sg_tmp->length; in omap_sham_align_sgs() 829 nbytes -= sg_tmp->length; in omap_sham_align_sgs() 832 sg_tmp = sg_next(sg_tmp); in omap_sham_align_sgs()
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/crypto/ |
H A D | s5p-sss.c | 1145 struct scatterlist *sg_tmp = sg; in s5p_hash_prepare_sgs() local 1153 while (nbytes > 0 && sg_tmp) { in s5p_hash_prepare_sgs() 1155 if (skip >= sg_tmp->length) { in s5p_hash_prepare_sgs() 1156 skip -= sg_tmp->length; in s5p_hash_prepare_sgs() 1157 if (!sg_tmp->length) { in s5p_hash_prepare_sgs() 1162 if (!IS_ALIGNED(sg_tmp->length - skip, BUFLEN)) { in s5p_hash_prepare_sgs() 1167 if (nbytes < sg_tmp->length - skip) { in s5p_hash_prepare_sgs() 1172 nbytes -= sg_tmp->length - skip; in s5p_hash_prepare_sgs() 1176 sg_tmp = sg_next(sg_tmp); in s5p_hash_prepare_sgs()
|
H A D | omap-sham.c | 754 struct scatterlist *sg_tmp = sg; in omap_sham_align_sgs() local 787 while (nbytes > 0 && sg_tmp) { in omap_sham_align_sgs() 804 if (page_zonenum(sg_page(sg_tmp)) != ZONE_DMA) { in omap_sham_align_sgs() 810 if (offset < sg_tmp->length) { in omap_sham_align_sgs() 811 if (!IS_ALIGNED(offset + sg_tmp->offset, 4)) { in omap_sham_align_sgs() 816 if (!IS_ALIGNED(sg_tmp->length - offset, bs)) { in omap_sham_align_sgs() 823 offset -= sg_tmp->length; in omap_sham_align_sgs() 829 nbytes -= sg_tmp->length; in omap_sham_align_sgs() 832 sg_tmp = sg_next(sg_tmp); in omap_sham_align_sgs()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/crypto/ |
H A D | s5p-sss.c | 1145 struct scatterlist *sg_tmp = sg; in s5p_hash_prepare_sgs() local 1153 while (nbytes > 0 && sg_tmp) { in s5p_hash_prepare_sgs() 1155 if (skip >= sg_tmp->length) { in s5p_hash_prepare_sgs() 1156 skip -= sg_tmp->length; in s5p_hash_prepare_sgs() 1157 if (!sg_tmp->length) { in s5p_hash_prepare_sgs() 1162 if (!IS_ALIGNED(sg_tmp->length - skip, BUFLEN)) { in s5p_hash_prepare_sgs() 1167 if (nbytes < sg_tmp->length - skip) { in s5p_hash_prepare_sgs() 1172 nbytes -= sg_tmp->length - skip; in s5p_hash_prepare_sgs() 1176 sg_tmp = sg_next(sg_tmp); in s5p_hash_prepare_sgs()
|
H A D | omap-sham.c | 754 struct scatterlist *sg_tmp = sg; in omap_sham_align_sgs() local 787 while (nbytes > 0 && sg_tmp) { in omap_sham_align_sgs() 804 if (page_zonenum(sg_page(sg_tmp)) != ZONE_DMA) { in omap_sham_align_sgs() 810 if (offset < sg_tmp->length) { in omap_sham_align_sgs() 811 if (!IS_ALIGNED(offset + sg_tmp->offset, 4)) { in omap_sham_align_sgs() 816 if (!IS_ALIGNED(sg_tmp->length - offset, bs)) { in omap_sham_align_sgs() 823 offset -= sg_tmp->length; in omap_sham_align_sgs() 829 nbytes -= sg_tmp->length; in omap_sham_align_sgs() 832 sg_tmp = sg_next(sg_tmp); in omap_sham_align_sgs()
|