Home
last modified time | relevance | path

Searched refs:sg_src (Results 1 – 25 of 29) sorted by relevance

12

/OK3568_Linux_fs/kernel/drivers/crypto/rockchip/
H A Drk3288_crypto.c68 static int check_alignment(struct scatterlist *sg_src, in check_alignment() argument
74 in = IS_ALIGNED((uint32_t)sg_src->offset, 4) && in check_alignment()
75 IS_ALIGNED((uint32_t)sg_src->length, align_mask); in check_alignment()
82 return (align && (sg_src->length == sg_dst->length)); in check_alignment()
86 struct scatterlist *sg_src, in rk_load_data() argument
92 check_alignment(sg_src, sg_dst, dev->align_size) : in rk_load_data()
95 count = min(dev->left_bytes, sg_src->length); in rk_load_data()
98 if (!dma_map_sg(dev->dev, sg_src, 1, DMA_TO_DEVICE)) { in rk_load_data()
103 dev->addr_in = sg_dma_address(sg_src); in rk_load_data()
110 dma_unmap_sg(dev->dev, sg_src, 1, in rk_load_data()
[all …]
H A Drk_crypto_core.c62 struct scatterlist *sg_src, in rk_load_data() argument
109 sg_src = scatterwalk_ffwd(rk_dev->src, alg_ctx->req_src, in rk_load_data()
111 sg_dst = (alg_ctx->req_src == alg_ctx->req_dst) ? sg_src : in rk_load_data()
115 alg_ctx->sg_src = sg_src; in rk_load_data()
117 src_nents = sg_nents_for_len(sg_src, alg_ctx->total); in rk_load_data()
124 alg_ctx->aligned = rk_crypto_check_align(sg_src, src_nents, sg_dst, dst_nents, in rk_load_data()
126 alg_ctx->is_dma = rk_crypto_check_dmafd(sg_src, src_nents) && in rk_load_data()
138 count = rk_crypto_hw_desc_maxlen(sg_src, alg_ctx->left_bytes, &nents); in rk_load_data()
141 count = min_t(unsigned int, alg_ctx->left_bytes, sg_src->length); in rk_load_data()
147 if (!alg_ctx->is_dma && !dma_map_sg(dev, sg_src, nents, DMA_TO_DEVICE)) { in rk_load_data()
[all …]
H A Drk_crypto_v1_skcipher.c249 u8 *src_last_blk = page_address(sg_page(alg_ctx->sg_src)) + in rk_set_data_start()
250 alg_ctx->sg_src->offset + alg_ctx->sg_src->length - ivsize; in rk_set_data_start()
261 err = rk_dev->load_data(rk_dev, alg_ctx->sg_src, alg_ctx->sg_dst); in rk_set_data_start()
276 alg_ctx->sg_src = req->src; in rk_ablk_start()
352 if (sg_is_last(alg_ctx->sg_src)) { in rk_ablk_rx()
358 alg_ctx->sg_src = sg_next(alg_ctx->sg_src); in rk_ablk_rx()
H A Drk_crypto_utils.c69 static int check_scatter_align(struct scatterlist *sg_src, in check_scatter_align() argument
75 in = IS_ALIGNED((u32)sg_src->offset, 4) && in check_scatter_align()
76 IS_ALIGNED((u32)sg_src->length, align_mask) && in check_scatter_align()
77 (sg_phys(sg_src) < SZ_4G); in check_scatter_align()
86 return (align && (sg_src->length == sg_dst->length)); in check_scatter_align()
H A Drk3288_crypto_skcipher.c251 u8 *src_last_blk = page_address(sg_page(dev->sg_src)) + in rk_set_data_start()
252 dev->sg_src->offset + dev->sg_src->length - ivsize; in rk_set_data_start()
263 err = dev->load_data(dev, dev->sg_src, dev->sg_dst); in rk_set_data_start()
278 dev->sg_src = req->src; in rk_ablk_start()
357 if (sg_is_last(dev->sg_src)) { in rk_ablk_rx()
363 dev->sg_src = sg_next(dev->sg_src); in rk_ablk_rx()
H A Drk_crypto_skcipher_utils.c185 err = rk_dev->load_data(rk_dev, alg_ctx->sg_src, alg_ctx->sg_dst); in rk_set_data_start()
193 src_sg = alg_ctx->aligned ? alg_ctx->sg_src : &alg_ctx->sg_tmp; in rk_set_data_start()
269 if (sg_is_last(alg_ctx->sg_src)) { in rk_ablk_rx()
275 alg_ctx->sg_src = sg_next(alg_ctx->sg_src); in rk_ablk_rx()
338 alg_ctx->sg_src = req->src; in rk_ablk_start()
452 alg_ctx->sg_src = req->src; in rk_aead_start()
H A Drk_crypto_ahash_utils.c29 alg_ctx->sg_src = 0; in rk_alg_ctx_clear()
73 err = rk_dev->load_data(rk_dev, alg_ctx->sg_src, alg_ctx->sg_dst); in rk_ahash_set_data_start()
372 alg_ctx->sg_src = src_sg; in rk_ahash_start()
414 if (sg_is_last(alg_ctx->sg_src)) { in rk_ahash_crypto_rx()
420 alg_ctx->sg_src = sg_next(alg_ctx->sg_src); in rk_ahash_crypto_rx()
H A Drk_crypto_v1_ahash.c239 err = rk_dev->load_data(rk_dev, alg_ctx->sg_src, NULL); in rk_ahash_set_data_start()
254 alg_ctx->sg_src = req->src; in rk_ahash_start()
293 if (sg_is_last(alg_ctx->sg_src)) { in rk_ahash_crypto_rx()
299 alg_ctx->sg_src = sg_next(alg_ctx->sg_src); in rk_ahash_crypto_rx()
H A Drk3288_crypto_ahash.c188 err = dev->load_data(dev, dev->sg_src, NULL); in rk_ahash_set_data_start()
205 dev->sg_src = req->src; in rk_ahash_start()
239 if (sg_is_last(dev->sg_src)) { in rk_ahash_crypto_rx()
245 dev->sg_src = sg_next(dev->sg_src); in rk_ahash_crypto_rx()
H A Drk_crypto_v2_skcipher.c350 alg_ctx->sg_src, alg_ctx->sg_dst, alg_ctx->count); in crypto_dma_start()
543 struct scatterlist *sg_src, *sg_dst; in rk_aead_crypt() local
562 sg_src = scatterwalk_ffwd(src, req->src, req->assoclen); in rk_aead_crypt()
563 sg_dst = (req->src == req->dst) ? sg_src : scatterwalk_ffwd(dst, req->dst, req->assoclen); in rk_aead_crypt()
565 aligned = rk_crypto_check_align(sg_src, sg_nents_for_len(sg_src, data_len), in rk_aead_crypt()
569 if (sg_nents_for_len(sg_src, data_len) > RK_DEFAULT_LLI_CNT || in rk_aead_crypt()
H A Drk_crypto_v3_skcipher.c349 alg_ctx->sg_src, alg_ctx->sg_dst, alg_ctx->count); in crypto_dma_start()
542 struct scatterlist *sg_src, *sg_dst; in rk_aead_crypt() local
561 sg_src = scatterwalk_ffwd(src, req->src, req->assoclen); in rk_aead_crypt()
562 sg_dst = (req->src == req->dst) ? sg_src : scatterwalk_ffwd(dst, req->dst, req->assoclen); in rk_aead_crypt()
564 aligned = rk_crypto_check_align(sg_src, sg_nents_for_len(sg_src, data_len), in rk_aead_crypt()
568 if (sg_nents_for_len(sg_src, data_len) > RK_DEFAULT_LLI_CNT || in rk_aead_crypt()
H A Drk3288_crypto.h204 struct scatterlist *sg_src; member
225 struct scatterlist *sg_src,
H A Drk_crypto_core.h100 struct scatterlist *sg_src,
139 struct scatterlist *sg_src; member
/OK3568_Linux_fs/kernel/arch/arm/crypto/
H A Daes-ce-glue.c273 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_encrypt() local
301 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_encrypt()
331 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_decrypt() local
359 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_decrypt()
450 struct scatterlist sg_src[2], sg_dst[2]; in xts_encrypt() local
496 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_encrypt()
522 struct scatterlist sg_src[2], sg_dst[2]; in xts_decrypt() local
568 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_decrypt()
/OK3568_Linux_fs/kernel/drivers/crypto/nx/
H A Dnx.c150 struct scatterlist *sg_src, in nx_walk_and_build() argument
161 scatterwalk_start(&walk, sg_src); in nx_walk_and_build()
163 if (start < offset + sg_src->length) in nx_walk_and_build()
166 offset += sg_src->length; in nx_walk_and_build()
167 sg_src = sg_next(sg_src); in nx_walk_and_build()
/OK3568_Linux_fs/kernel/arch/arm64/crypto/
H A Daes-glue.c278 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_encrypt() local
305 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_encrypt()
335 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_decrypt() local
362 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_decrypt()
490 struct scatterlist sg_src[2], sg_dst[2]; in xts_encrypt() local
536 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_encrypt()
562 struct scatterlist sg_src[2], sg_dst[2]; in xts_decrypt() local
608 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_decrypt()
H A Daes-neonbs-glue.c278 struct scatterlist sg_src[2], sg_dst[2]; in __xts_crypt() local
348 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in __xts_crypt()
/OK3568_Linux_fs/kernel/lib/crypto/
H A Dchacha20poly1305-selftest.c8891 struct scatterlist sg_src[3]; in chacha20poly1305_selftest() local
8926 sg_init_one(sg_src, computed_output, in chacha20poly1305_selftest()
8928 ret = chacha20poly1305_encrypt_sg_inplace(sg_src, in chacha20poly1305_selftest()
8968 sg_init_one(sg_src, computed_output, in chacha20poly1305_selftest()
8970 ret = chacha20poly1305_decrypt_sg_inplace(sg_src, in chacha20poly1305_selftest()
9032 sg_init_table(sg_src, 3); in chacha20poly1305_selftest()
9034 sg_set_buf(&sg_src[k++], input, i); in chacha20poly1305_selftest()
9036 sg_set_buf(&sg_src[k++], input + i, j - i); in chacha20poly1305_selftest()
9038 sg_set_buf(&sg_src[k++], input + j, total_len - j); in chacha20poly1305_selftest()
9039 sg_init_marker(sg_src, k); in chacha20poly1305_selftest()
[all …]
/OK3568_Linux_fs/kernel/drivers/video/rockchip/rga/
H A Drga_drv.c762 reg->sg_src = req->sg_src; in rga_reg_init()
890 sgt = (!reg) ? req->sg_src : reg->sg_src; in rga_put_dma_buf()
1007 req->sg_src = NULL; in rga_convert_dma_buf()
1037 req->sg_src = ion_sg_table(rga_drvdata->ion_client, hdl); in rga_convert_dma_buf()
1192 req->sg_src = NULL; in rga_get_dma_buf()
1197 ret = rga_get_img_info(&req->src, mmu_flag, &req->sg_src, in rga_get_dma_buf()
1215 if (req->sg_src && req->attach_src) { in rga_get_dma_buf()
1217 req->sg_src, DMA_BIDIRECTIONAL); in rga_get_dma_buf()
1282 reg1->sg_src = req1->sg_src; in rga_reg_init_2()
H A Drga.h303 struct sg_table *sg_src; member
385 struct sg_table *sg_src; member
/OK3568_Linux_fs/kernel/arch/x86/crypto/
H A Daegis128-aesni-glue.c70 struct aegis_state *state, struct scatterlist *sg_src, in crypto_aegis128_aesni_process_ad() argument
77 scatterwalk_start(&walk, sg_src); in crypto_aegis128_aesni_process_ad()
/OK3568_Linux_fs/kernel/drivers/crypto/
H A Ds5p-sss.c309 struct scatterlist *sg_src; member
514 dma_unmap_sg(dev->dev, dev->sg_src, 1, DMA_TO_DEVICE); in s5p_unset_indata()
564 dev->sg_src = sg; in s5p_set_indata()
604 if (!sg_is_last(dev->sg_src)) { in s5p_aes_rx()
605 ret = s5p_set_indata(dev, sg_next(dev->sg_src)); in s5p_aes_rx()
749 s5p_set_dma_indata(dev, dev->sg_src); in s5p_aes_interrupt()
1957 s5p_set_dma_indata(dev, dev->sg_src); in s5p_aes_crypt_start()
/OK3568_Linux_fs/kernel/drivers/crypto/ux500/cryp/
H A Dcryp.h215 struct scatterlist *sg_src; member
/OK3568_Linux_fs/kernel/crypto/
H A Daegis128-core.c281 struct scatterlist *sg_src, in crypto_aegis128_process_ad() argument
288 scatterwalk_start(&walk, sg_src); in crypto_aegis128_process_ad()
/OK3568_Linux_fs/kernel/drivers/dma/
H A Dste_dma40.c2111 struct scatterlist *sg_src, struct scatterlist *sg_dst, in d40_prep_sg_log() argument
2120 ret = d40_log_sg_to_lli(sg_src, sg_len, in d40_prep_sg_log()
2139 struct scatterlist *sg_src, struct scatterlist *sg_dst, in d40_prep_sg_phy() argument
2152 ret = d40_phy_sg_to_lli(sg_src, sg_len, src_dev_addr, in d40_prep_sg_phy()
2209 d40_prep_sg(struct dma_chan *dchan, struct scatterlist *sg_src, in d40_prep_sg() argument
2229 desc = d40_prep_desc(chan, sg_src, sg_len, dma_flags); in d40_prep_sg()
2233 if (sg_next(&sg_src[sg_len - 1]) == sg_src) in d40_prep_sg()
2244 ret = d40_prep_sg_log(chan, desc, sg_src, sg_dst, in d40_prep_sg()
2247 ret = d40_prep_sg_phy(chan, desc, sg_src, sg_dst, in d40_prep_sg()

12