| /OK3568_Linux_fs/kernel/drivers/crypto/rockchip/ |
| H A D | rk3288_crypto.c | 68 static int check_alignment(struct scatterlist *sg_src, in check_alignment() argument 74 in = IS_ALIGNED((uint32_t)sg_src->offset, 4) && in check_alignment() 75 IS_ALIGNED((uint32_t)sg_src->length, align_mask); in check_alignment() 82 return (align && (sg_src->length == sg_dst->length)); in check_alignment() 86 struct scatterlist *sg_src, in rk_load_data() argument 92 check_alignment(sg_src, sg_dst, dev->align_size) : in rk_load_data() 95 count = min(dev->left_bytes, sg_src->length); in rk_load_data() 98 if (!dma_map_sg(dev->dev, sg_src, 1, DMA_TO_DEVICE)) { in rk_load_data() 103 dev->addr_in = sg_dma_address(sg_src); in rk_load_data() 110 dma_unmap_sg(dev->dev, sg_src, 1, in rk_load_data() [all …]
|
| H A D | rk_crypto_core.c | 62 struct scatterlist *sg_src, in rk_load_data() argument 109 sg_src = scatterwalk_ffwd(rk_dev->src, alg_ctx->req_src, in rk_load_data() 111 sg_dst = (alg_ctx->req_src == alg_ctx->req_dst) ? sg_src : in rk_load_data() 115 alg_ctx->sg_src = sg_src; in rk_load_data() 117 src_nents = sg_nents_for_len(sg_src, alg_ctx->total); in rk_load_data() 124 alg_ctx->aligned = rk_crypto_check_align(sg_src, src_nents, sg_dst, dst_nents, in rk_load_data() 126 alg_ctx->is_dma = rk_crypto_check_dmafd(sg_src, src_nents) && in rk_load_data() 138 count = rk_crypto_hw_desc_maxlen(sg_src, alg_ctx->left_bytes, &nents); in rk_load_data() 141 count = min_t(unsigned int, alg_ctx->left_bytes, sg_src->length); in rk_load_data() 147 if (!alg_ctx->is_dma && !dma_map_sg(dev, sg_src, nents, DMA_TO_DEVICE)) { in rk_load_data() [all …]
|
| H A D | rk_crypto_v1_skcipher.c | 249 u8 *src_last_blk = page_address(sg_page(alg_ctx->sg_src)) + in rk_set_data_start() 250 alg_ctx->sg_src->offset + alg_ctx->sg_src->length - ivsize; in rk_set_data_start() 261 err = rk_dev->load_data(rk_dev, alg_ctx->sg_src, alg_ctx->sg_dst); in rk_set_data_start() 276 alg_ctx->sg_src = req->src; in rk_ablk_start() 352 if (sg_is_last(alg_ctx->sg_src)) { in rk_ablk_rx() 358 alg_ctx->sg_src = sg_next(alg_ctx->sg_src); in rk_ablk_rx()
|
| H A D | rk_crypto_utils.c | 69 static int check_scatter_align(struct scatterlist *sg_src, in check_scatter_align() argument 75 in = IS_ALIGNED((u32)sg_src->offset, 4) && in check_scatter_align() 76 IS_ALIGNED((u32)sg_src->length, align_mask) && in check_scatter_align() 77 (sg_phys(sg_src) < SZ_4G); in check_scatter_align() 86 return (align && (sg_src->length == sg_dst->length)); in check_scatter_align()
|
| H A D | rk3288_crypto_skcipher.c | 251 u8 *src_last_blk = page_address(sg_page(dev->sg_src)) + in rk_set_data_start() 252 dev->sg_src->offset + dev->sg_src->length - ivsize; in rk_set_data_start() 263 err = dev->load_data(dev, dev->sg_src, dev->sg_dst); in rk_set_data_start() 278 dev->sg_src = req->src; in rk_ablk_start() 357 if (sg_is_last(dev->sg_src)) { in rk_ablk_rx() 363 dev->sg_src = sg_next(dev->sg_src); in rk_ablk_rx()
|
| H A D | rk_crypto_skcipher_utils.c | 185 err = rk_dev->load_data(rk_dev, alg_ctx->sg_src, alg_ctx->sg_dst); in rk_set_data_start() 193 src_sg = alg_ctx->aligned ? alg_ctx->sg_src : &alg_ctx->sg_tmp; in rk_set_data_start() 269 if (sg_is_last(alg_ctx->sg_src)) { in rk_ablk_rx() 275 alg_ctx->sg_src = sg_next(alg_ctx->sg_src); in rk_ablk_rx() 338 alg_ctx->sg_src = req->src; in rk_ablk_start() 452 alg_ctx->sg_src = req->src; in rk_aead_start()
|
| H A D | rk_crypto_ahash_utils.c | 29 alg_ctx->sg_src = 0; in rk_alg_ctx_clear() 73 err = rk_dev->load_data(rk_dev, alg_ctx->sg_src, alg_ctx->sg_dst); in rk_ahash_set_data_start() 372 alg_ctx->sg_src = src_sg; in rk_ahash_start() 414 if (sg_is_last(alg_ctx->sg_src)) { in rk_ahash_crypto_rx() 420 alg_ctx->sg_src = sg_next(alg_ctx->sg_src); in rk_ahash_crypto_rx()
|
| H A D | rk_crypto_v1_ahash.c | 239 err = rk_dev->load_data(rk_dev, alg_ctx->sg_src, NULL); in rk_ahash_set_data_start() 254 alg_ctx->sg_src = req->src; in rk_ahash_start() 293 if (sg_is_last(alg_ctx->sg_src)) { in rk_ahash_crypto_rx() 299 alg_ctx->sg_src = sg_next(alg_ctx->sg_src); in rk_ahash_crypto_rx()
|
| H A D | rk3288_crypto_ahash.c | 188 err = dev->load_data(dev, dev->sg_src, NULL); in rk_ahash_set_data_start() 205 dev->sg_src = req->src; in rk_ahash_start() 239 if (sg_is_last(dev->sg_src)) { in rk_ahash_crypto_rx() 245 dev->sg_src = sg_next(dev->sg_src); in rk_ahash_crypto_rx()
|
| H A D | rk_crypto_v2_skcipher.c | 350 alg_ctx->sg_src, alg_ctx->sg_dst, alg_ctx->count); in crypto_dma_start() 543 struct scatterlist *sg_src, *sg_dst; in rk_aead_crypt() local 562 sg_src = scatterwalk_ffwd(src, req->src, req->assoclen); in rk_aead_crypt() 563 sg_dst = (req->src == req->dst) ? sg_src : scatterwalk_ffwd(dst, req->dst, req->assoclen); in rk_aead_crypt() 565 aligned = rk_crypto_check_align(sg_src, sg_nents_for_len(sg_src, data_len), in rk_aead_crypt() 569 if (sg_nents_for_len(sg_src, data_len) > RK_DEFAULT_LLI_CNT || in rk_aead_crypt()
|
| H A D | rk_crypto_v3_skcipher.c | 349 alg_ctx->sg_src, alg_ctx->sg_dst, alg_ctx->count); in crypto_dma_start() 542 struct scatterlist *sg_src, *sg_dst; in rk_aead_crypt() local 561 sg_src = scatterwalk_ffwd(src, req->src, req->assoclen); in rk_aead_crypt() 562 sg_dst = (req->src == req->dst) ? sg_src : scatterwalk_ffwd(dst, req->dst, req->assoclen); in rk_aead_crypt() 564 aligned = rk_crypto_check_align(sg_src, sg_nents_for_len(sg_src, data_len), in rk_aead_crypt() 568 if (sg_nents_for_len(sg_src, data_len) > RK_DEFAULT_LLI_CNT || in rk_aead_crypt()
|
| H A D | rk3288_crypto.h | 204 struct scatterlist *sg_src; member 225 struct scatterlist *sg_src,
|
| H A D | rk_crypto_core.h | 100 struct scatterlist *sg_src, 139 struct scatterlist *sg_src; member
|
| /OK3568_Linux_fs/kernel/arch/arm/crypto/ |
| H A D | aes-ce-glue.c | 273 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_encrypt() local 301 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_encrypt() 331 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_decrypt() local 359 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_decrypt() 450 struct scatterlist sg_src[2], sg_dst[2]; in xts_encrypt() local 496 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_encrypt() 522 struct scatterlist sg_src[2], sg_dst[2]; in xts_decrypt() local 568 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_decrypt()
|
| /OK3568_Linux_fs/kernel/drivers/crypto/nx/ |
| H A D | nx.c | 150 struct scatterlist *sg_src, in nx_walk_and_build() argument 161 scatterwalk_start(&walk, sg_src); in nx_walk_and_build() 163 if (start < offset + sg_src->length) in nx_walk_and_build() 166 offset += sg_src->length; in nx_walk_and_build() 167 sg_src = sg_next(sg_src); in nx_walk_and_build()
|
| /OK3568_Linux_fs/kernel/arch/arm64/crypto/ |
| H A D | aes-glue.c | 278 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_encrypt() local 305 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_encrypt() 335 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_decrypt() local 362 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_decrypt() 490 struct scatterlist sg_src[2], sg_dst[2]; in xts_encrypt() local 536 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_encrypt() 562 struct scatterlist sg_src[2], sg_dst[2]; in xts_decrypt() local 608 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_decrypt()
|
| H A D | aes-neonbs-glue.c | 278 struct scatterlist sg_src[2], sg_dst[2]; in __xts_crypt() local 348 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in __xts_crypt()
|
| /OK3568_Linux_fs/kernel/lib/crypto/ |
| H A D | chacha20poly1305-selftest.c | 8891 struct scatterlist sg_src[3]; in chacha20poly1305_selftest() local 8926 sg_init_one(sg_src, computed_output, in chacha20poly1305_selftest() 8928 ret = chacha20poly1305_encrypt_sg_inplace(sg_src, in chacha20poly1305_selftest() 8968 sg_init_one(sg_src, computed_output, in chacha20poly1305_selftest() 8970 ret = chacha20poly1305_decrypt_sg_inplace(sg_src, in chacha20poly1305_selftest() 9032 sg_init_table(sg_src, 3); in chacha20poly1305_selftest() 9034 sg_set_buf(&sg_src[k++], input, i); in chacha20poly1305_selftest() 9036 sg_set_buf(&sg_src[k++], input + i, j - i); in chacha20poly1305_selftest() 9038 sg_set_buf(&sg_src[k++], input + j, total_len - j); in chacha20poly1305_selftest() 9039 sg_init_marker(sg_src, k); in chacha20poly1305_selftest() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/video/rockchip/rga/ |
| H A D | rga_drv.c | 762 reg->sg_src = req->sg_src; in rga_reg_init() 890 sgt = (!reg) ? req->sg_src : reg->sg_src; in rga_put_dma_buf() 1007 req->sg_src = NULL; in rga_convert_dma_buf() 1037 req->sg_src = ion_sg_table(rga_drvdata->ion_client, hdl); in rga_convert_dma_buf() 1192 req->sg_src = NULL; in rga_get_dma_buf() 1197 ret = rga_get_img_info(&req->src, mmu_flag, &req->sg_src, in rga_get_dma_buf() 1215 if (req->sg_src && req->attach_src) { in rga_get_dma_buf() 1217 req->sg_src, DMA_BIDIRECTIONAL); in rga_get_dma_buf() 1282 reg1->sg_src = req1->sg_src; in rga_reg_init_2()
|
| H A D | rga.h | 303 struct sg_table *sg_src; member 385 struct sg_table *sg_src; member
|
| /OK3568_Linux_fs/kernel/arch/x86/crypto/ |
| H A D | aegis128-aesni-glue.c | 70 struct aegis_state *state, struct scatterlist *sg_src, in crypto_aegis128_aesni_process_ad() argument 77 scatterwalk_start(&walk, sg_src); in crypto_aegis128_aesni_process_ad()
|
| /OK3568_Linux_fs/kernel/drivers/crypto/ |
| H A D | s5p-sss.c | 309 struct scatterlist *sg_src; member 514 dma_unmap_sg(dev->dev, dev->sg_src, 1, DMA_TO_DEVICE); in s5p_unset_indata() 564 dev->sg_src = sg; in s5p_set_indata() 604 if (!sg_is_last(dev->sg_src)) { in s5p_aes_rx() 605 ret = s5p_set_indata(dev, sg_next(dev->sg_src)); in s5p_aes_rx() 749 s5p_set_dma_indata(dev, dev->sg_src); in s5p_aes_interrupt() 1957 s5p_set_dma_indata(dev, dev->sg_src); in s5p_aes_crypt_start()
|
| /OK3568_Linux_fs/kernel/drivers/crypto/ux500/cryp/ |
| H A D | cryp.h | 215 struct scatterlist *sg_src; member
|
| /OK3568_Linux_fs/kernel/crypto/ |
| H A D | aegis128-core.c | 281 struct scatterlist *sg_src, in crypto_aegis128_process_ad() argument 288 scatterwalk_start(&walk, sg_src); in crypto_aegis128_process_ad()
|
| /OK3568_Linux_fs/kernel/drivers/dma/ |
| H A D | ste_dma40.c | 2111 struct scatterlist *sg_src, struct scatterlist *sg_dst, in d40_prep_sg_log() argument 2120 ret = d40_log_sg_to_lli(sg_src, sg_len, in d40_prep_sg_log() 2139 struct scatterlist *sg_src, struct scatterlist *sg_dst, in d40_prep_sg_phy() argument 2152 ret = d40_phy_sg_to_lli(sg_src, sg_len, src_dev_addr, in d40_prep_sg_phy() 2209 d40_prep_sg(struct dma_chan *dchan, struct scatterlist *sg_src, in d40_prep_sg() argument 2229 desc = d40_prep_desc(chan, sg_src, sg_len, dma_flags); in d40_prep_sg() 2233 if (sg_next(&sg_src[sg_len - 1]) == sg_src) in d40_prep_sg() 2244 ret = d40_prep_sg_log(chan, desc, sg_src, sg_dst, in d40_prep_sg() 2247 ret = d40_prep_sg_phy(chan, desc, sg_src, sg_dst, in d40_prep_sg()
|