| /OK3568_Linux_fs/kernel/drivers/s390/scsi/ |
| H A D | zfcp_fsf.c | 80 static void zfcp_fsf_class_not_supp(struct zfcp_fsf_req *req) in zfcp_fsf_class_not_supp() argument 82 dev_err(&req->adapter->ccw_device->dev, "FCP device not " in zfcp_fsf_class_not_supp() 84 zfcp_erp_adapter_shutdown(req->adapter, 0, "fscns_1"); in zfcp_fsf_class_not_supp() 85 req->status |= ZFCP_STATUS_FSFREQ_ERROR; in zfcp_fsf_class_not_supp() 90 * @req: pointer to struct zfcp_fsf_req 92 void zfcp_fsf_req_free(struct zfcp_fsf_req *req) in zfcp_fsf_req_free() argument 94 if (likely(req->pool)) { in zfcp_fsf_req_free() 95 if (likely(!zfcp_fsf_req_is_status_read_buffer(req))) in zfcp_fsf_req_free() 96 mempool_free(req->qtcb, req->adapter->pool.qtcb_pool); in zfcp_fsf_req_free() 97 mempool_free(req, req->pool); in zfcp_fsf_req_free() [all …]
|
| /OK3568_Linux_fs/external/xserver/glx/ |
| H A D | glxcmdsswap.c | 59 xGLXCreateContextReq *req = (xGLXCreateContextReq *) pc; in __glXDispSwap_CreateContext() local 63 __GLX_SWAP_SHORT(&req->length); in __glXDispSwap_CreateContext() 64 __GLX_SWAP_INT(&req->context); in __glXDispSwap_CreateContext() 65 __GLX_SWAP_INT(&req->visual); in __glXDispSwap_CreateContext() 66 __GLX_SWAP_INT(&req->screen); in __glXDispSwap_CreateContext() 67 __GLX_SWAP_INT(&req->shareList); in __glXDispSwap_CreateContext() 75 xGLXCreateNewContextReq *req = (xGLXCreateNewContextReq *) pc; in __glXDispSwap_CreateNewContext() local 79 __GLX_SWAP_SHORT(&req->length); in __glXDispSwap_CreateNewContext() 80 __GLX_SWAP_INT(&req->context); in __glXDispSwap_CreateNewContext() 81 __GLX_SWAP_INT(&req->fbconfig); in __glXDispSwap_CreateNewContext() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/staging/greybus/ |
| H A D | audio_apbridgea.c | 16 struct audio_apbridgea_set_config_request req; in gb_audio_apbridgea_set_config() local 18 req.hdr.type = AUDIO_APBRIDGEA_TYPE_SET_CONFIG; in gb_audio_apbridgea_set_config() 19 req.hdr.i2s_port = cpu_to_le16(i2s_port); in gb_audio_apbridgea_set_config() 20 req.format = cpu_to_le32(format); in gb_audio_apbridgea_set_config() 21 req.rate = cpu_to_le32(rate); in gb_audio_apbridgea_set_config() 22 req.mclk_freq = cpu_to_le32(mclk_freq); in gb_audio_apbridgea_set_config() 24 return gb_hd_output(connection->hd, &req, sizeof(req), in gb_audio_apbridgea_set_config() 33 struct audio_apbridgea_register_cport_request req; in gb_audio_apbridgea_register_cport() local 36 req.hdr.type = AUDIO_APBRIDGEA_TYPE_REGISTER_CPORT; in gb_audio_apbridgea_register_cport() 37 req.hdr.i2s_port = cpu_to_le16(i2s_port); in gb_audio_apbridgea_register_cport() [all …]
|
| H A D | audio_gb.c | 50 struct gb_audio_get_control_request req; in gb_audio_gb_get_control() local 54 req.control_id = control_id; in gb_audio_gb_get_control() 55 req.index = index; in gb_audio_gb_get_control() 58 &req, sizeof(req), &resp, sizeof(resp)); in gb_audio_gb_get_control() 72 struct gb_audio_set_control_request req; in gb_audio_gb_set_control() local 74 req.control_id = control_id; in gb_audio_gb_set_control() 75 req.index = index; in gb_audio_gb_set_control() 76 memcpy(&req.value, value, sizeof(req.value)); in gb_audio_gb_set_control() 79 &req, sizeof(req), NULL, 0); in gb_audio_gb_set_control() 86 struct gb_audio_enable_widget_request req; in gb_audio_gb_enable_widget() local [all …]
|
| /OK3568_Linux_fs/kernel/drivers/nvme/target/ |
| H A D | io-cmd-file.c | 102 static ssize_t nvmet_file_submit_bvec(struct nvmet_req *req, loff_t pos, in nvmet_file_submit_bvec() argument 105 struct kiocb *iocb = &req->f.iocb; in nvmet_file_submit_bvec() 110 if (req->cmd->rw.opcode == nvme_cmd_write) { in nvmet_file_submit_bvec() 111 if (req->cmd->rw.control & cpu_to_le16(NVME_RW_FUA)) in nvmet_file_submit_bvec() 113 call_iter = req->ns->file->f_op->write_iter; in nvmet_file_submit_bvec() 116 call_iter = req->ns->file->f_op->read_iter; in nvmet_file_submit_bvec() 120 iov_iter_bvec(&iter, rw, req->f.bvec, nr_segs, count); in nvmet_file_submit_bvec() 123 iocb->ki_filp = req->ns->file; in nvmet_file_submit_bvec() 124 iocb->ki_flags = ki_flags | iocb_flags(req->ns->file); in nvmet_file_submit_bvec() 131 struct nvmet_req *req = container_of(iocb, struct nvmet_req, f.iocb); in nvmet_file_io_done() local [all …]
|
| H A D | admin-cmd.c | 28 static u32 nvmet_feat_data_len(struct nvmet_req *req, u32 cdw10) in nvmet_feat_data_len() argument 32 return sizeof(req->sq->ctrl->hostid); in nvmet_feat_data_len() 43 static void nvmet_execute_get_log_page_noop(struct nvmet_req *req) in nvmet_execute_get_log_page_noop() argument 45 nvmet_req_complete(req, nvmet_zero_sgl(req, 0, req->transfer_len)); in nvmet_execute_get_log_page_noop() 48 static void nvmet_execute_get_log_page_error(struct nvmet_req *req) in nvmet_execute_get_log_page_error() argument 50 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_execute_get_log_page_error() 60 if (nvmet_copy_to_sgl(req, offset, &ctrl->slots[slot], in nvmet_execute_get_log_page_error() 71 nvmet_req_complete(req, 0); in nvmet_execute_get_log_page_error() 74 static u16 nvmet_get_smart_log_nsid(struct nvmet_req *req, in nvmet_get_smart_log_nsid() argument 80 ns = nvmet_find_namespace(req->sq->ctrl, req->cmd->get_log_page.nsid); in nvmet_get_smart_log_nsid() [all …]
|
| H A D | io-cmd-bdev.c | 105 static u16 blk_to_nvme_status(struct nvmet_req *req, blk_status_t blk_sts) in blk_to_nvme_status() argument 120 req->error_loc = offsetof(struct nvme_rw_command, length); in blk_to_nvme_status() 124 req->error_loc = offsetof(struct nvme_rw_command, slba); in blk_to_nvme_status() 127 req->error_loc = offsetof(struct nvme_common_command, opcode); in blk_to_nvme_status() 128 switch (req->cmd->common.opcode) { in blk_to_nvme_status() 139 req->error_loc = offsetof(struct nvme_rw_command, nsid); in blk_to_nvme_status() 144 req->error_loc = offsetof(struct nvme_common_command, opcode); in blk_to_nvme_status() 147 switch (req->cmd->common.opcode) { in blk_to_nvme_status() 150 req->error_slba = le64_to_cpu(req->cmd->rw.slba); in blk_to_nvme_status() 153 req->error_slba = in blk_to_nvme_status() [all …]
|
| H A D | fabrics-cmd.c | 10 static void nvmet_execute_prop_set(struct nvmet_req *req) in nvmet_execute_prop_set() argument 12 u64 val = le64_to_cpu(req->cmd->prop_set.value); in nvmet_execute_prop_set() 15 if (!nvmet_check_transfer_len(req, 0)) in nvmet_execute_prop_set() 18 if (req->cmd->prop_set.attrib & 1) { in nvmet_execute_prop_set() 19 req->error_loc = in nvmet_execute_prop_set() 25 switch (le32_to_cpu(req->cmd->prop_set.offset)) { in nvmet_execute_prop_set() 27 nvmet_update_cc(req->sq->ctrl, val); in nvmet_execute_prop_set() 30 req->error_loc = in nvmet_execute_prop_set() 35 nvmet_req_complete(req, status); in nvmet_execute_prop_set() 38 static void nvmet_execute_prop_get(struct nvmet_req *req) in nvmet_execute_prop_get() argument [all …]
|
| /OK3568_Linux_fs/kernel/drivers/media/mc/ |
| H A D | mc-request.c | 39 static void media_request_clean(struct media_request *req) in media_request_clean() argument 44 WARN_ON(req->state != MEDIA_REQUEST_STATE_CLEANING); in media_request_clean() 45 WARN_ON(req->updating_count); in media_request_clean() 46 WARN_ON(req->access_count); in media_request_clean() 48 list_for_each_entry_safe(obj, obj_safe, &req->objects, list) { in media_request_clean() 53 req->updating_count = 0; in media_request_clean() 54 req->access_count = 0; in media_request_clean() 55 WARN_ON(req->num_incomplete_objects); in media_request_clean() 56 req->num_incomplete_objects = 0; in media_request_clean() 57 wake_up_interruptible_all(&req->poll_wait); in media_request_clean() [all …]
|
| /OK3568_Linux_fs/kernel/crypto/ |
| H A D | chacha20poly1305.c | 42 struct ahash_request req; /* must be last member */ member 48 struct skcipher_request req; /* must be last member */ member 70 static inline void async_done_continue(struct aead_request *req, int err, in async_done_continue() argument 74 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); in async_done_continue() 77 err = cont(req); in async_done_continue() 81 aead_request_complete(req, err); in async_done_continue() 84 static void chacha_iv(u8 *iv, struct aead_request *req, u32 icb) in chacha_iv() argument 86 struct chachapoly_ctx *ctx = crypto_aead_ctx(crypto_aead_reqtfm(req)); in chacha_iv() 91 memcpy(iv + sizeof(leicb) + ctx->saltlen, req->iv, in chacha_iv() 95 static int poly_verify_tag(struct aead_request *req) in poly_verify_tag() argument [all …]
|
| H A D | gcm.c | 60 int (*complete)(struct aead_request *req, u32 flags); 82 static int crypto_rfc4543_copy_src_to_dst(struct aead_request *req, bool enc); 85 struct aead_request *req) in crypto_gcm_reqctx() argument 87 unsigned long align = crypto_aead_alignmask(crypto_aead_reqtfm(req)); in crypto_gcm_reqctx() 89 return (void *)PTR_ALIGN((u8 *)aead_request_ctx(req), align + 1); in crypto_gcm_reqctx() 105 struct skcipher_request req; in crypto_gcm_setkey() member 123 skcipher_request_set_tfm(&data->req, ctr); in crypto_gcm_setkey() 124 skcipher_request_set_callback(&data->req, CRYPTO_TFM_REQ_MAY_SLEEP | in crypto_gcm_setkey() 128 skcipher_request_set_crypt(&data->req, data->sg, data->sg, in crypto_gcm_setkey() 131 err = crypto_wait_req(crypto_skcipher_encrypt(&data->req), in crypto_gcm_setkey() [all …]
|
| H A D | seqiv.c | 21 static void seqiv_aead_encrypt_complete2(struct aead_request *req, int err) in seqiv_aead_encrypt_complete2() argument 23 struct aead_request *subreq = aead_request_ctx(req); in seqiv_aead_encrypt_complete2() 32 geniv = crypto_aead_reqtfm(req); in seqiv_aead_encrypt_complete2() 33 memcpy(req->iv, subreq->iv, crypto_aead_ivsize(geniv)); in seqiv_aead_encrypt_complete2() 42 struct aead_request *req = base->data; in seqiv_aead_encrypt_complete() local 44 seqiv_aead_encrypt_complete2(req, err); in seqiv_aead_encrypt_complete() 45 aead_request_complete(req, err); in seqiv_aead_encrypt_complete() 48 static int seqiv_aead_encrypt(struct aead_request *req) in seqiv_aead_encrypt() argument 50 struct crypto_aead *geniv = crypto_aead_reqtfm(req); in seqiv_aead_encrypt() 52 struct aead_request *subreq = aead_request_ctx(req); in seqiv_aead_encrypt() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/crypto/inside-secure/ |
| H A D | safexcel_hash.c | 69 static inline u64 safexcel_queued_len(struct safexcel_ahash_req *req) in safexcel_queued_len() argument 71 return req->len - req->processed; in safexcel_queued_len() 108 struct safexcel_ahash_req *req, in safexcel_context_control() argument 122 if (unlikely(req->digest == CONTEXT_CONTROL_DIGEST_XCM)) { in safexcel_context_control() 123 if (req->xcbcmac) in safexcel_context_control() 126 memcpy(ctx->base.ctxr->data, req->state, req->state_sz); in safexcel_context_control() 128 if (!req->finish && req->xcbcmac) in safexcel_context_control() 133 CONTEXT_CONTROL_SIZE(req->state_sz / in safexcel_context_control() 139 CONTEXT_CONTROL_SIZE(req->state_sz / in safexcel_context_control() 142 } else if (!req->processed) { in safexcel_context_control() [all …]
|
| /OK3568_Linux_fs/external/xserver/hw/dmx/glxProxy/ |
| H A D | glxcmdsswap.c | 55 xGLXCreateContextReq *req = (xGLXCreateContextReq *) pc; in __glXSwapCreateContext() local 59 __GLX_SWAP_SHORT(&req->length); in __glXSwapCreateContext() 60 __GLX_SWAP_INT(&req->context); in __glXSwapCreateContext() 61 __GLX_SWAP_INT(&req->visual); in __glXSwapCreateContext() 62 __GLX_SWAP_INT(&req->screen); in __glXSwapCreateContext() 63 __GLX_SWAP_INT(&req->shareList); in __glXSwapCreateContext() 71 xGLXCreateNewContextReq *req = (xGLXCreateNewContextReq *) pc; in __glXSwapCreateNewContext() local 75 __GLX_SWAP_SHORT(&req->length); in __glXSwapCreateNewContext() 76 __GLX_SWAP_INT(&req->context); in __glXSwapCreateNewContext() 77 __GLX_SWAP_INT(&req->fbconfig); in __glXSwapCreateNewContext() [all …]
|
| H A D | glxvendor.c | 63 #define GetReqVendorPrivate(name, req) \ argument 65 if ((dpy->bufptr + SIZEOF(x##name##Req)) > dpy->bufmax)\ 67 req = (x##name##Req *)(dpy->last_req = dpy->bufptr);\ 68 req->length = (SIZEOF(x##name##Req))>>2;\ 69 dpy->bufptr += SIZEOF(x##name##Req);\ 73 #define GetReqVendorPrivate(name, req) \ argument 75 if ((dpy->bufptr + SIZEOF(x/**/name/**/Req)) > dpy->bufmax)\ 77 req = (x/**/name/**/Req *)(dpy->last_req = dpy->bufptr);\ 78 req->length = (SIZEOF(x/**/name/**/Req))>>2;\ 79 dpy->bufptr += SIZEOF(x/**/name/**/Req);\ [all …]
|
| /OK3568_Linux_fs/kernel/drivers/block/drbd/ |
| H A D | drbd_req.c | 26 struct drbd_request *req; in drbd_req_new() local 28 req = mempool_alloc(&drbd_request_mempool, GFP_NOIO); in drbd_req_new() 29 if (!req) in drbd_req_new() 31 memset(req, 0, sizeof(*req)); in drbd_req_new() 33 drbd_req_make_private_bio(req, bio_src); in drbd_req_new() 34 req->rq_state = (bio_data_dir(bio_src) == WRITE ? RQ_WRITE : 0) in drbd_req_new() 38 req->device = device; in drbd_req_new() 39 req->master_bio = bio_src; in drbd_req_new() 40 req->epoch = 0; in drbd_req_new() 42 drbd_clear_interval(&req->i); in drbd_req_new() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/video/rockchip/rga2/ |
| H A D | rga2_mmu_info.c | 367 int rga2_get_dma_info(struct rga2_reg *reg, struct rga2_req *req) in rga2_get_dma_info() argument 385 src0 = &req->src; in rga2_get_dma_info() 386 src1 = &req->src1; in rga2_get_dma_info() 387 dst = &req->dst; in rga2_get_dma_info() 388 els = &req->pat; in rga2_get_dma_info() 391 mmu_flag = req->mmu_info.src0_mmu_flag; in rga2_get_dma_info() 414 mmu_flag = req->mmu_info.src1_mmu_flag; in rga2_get_dma_info() 438 mmu_flag = req->mmu_info.dst_mmu_flag; in rga2_get_dma_info() 462 mmu_flag = req->mmu_info.els_mmu_flag; in rga2_get_dma_info() 505 static int rga2_get_dma_info(struct rga2_reg *reg, struct rga2_req *req) in rga2_get_dma_info() argument [all …]
|
| /OK3568_Linux_fs/kernel/drivers/infiniband/hw/hfi1/ |
| H A D | user_sdma.c | 80 static int user_sdma_send_pkts(struct user_sdma_request *req, u16 maxpkts); 83 static void user_sdma_free_request(struct user_sdma_request *req, bool unpin); 84 static int pin_vector_pages(struct user_sdma_request *req, 88 static int check_header_template(struct user_sdma_request *req, 91 static int set_txreq_header(struct user_sdma_request *req, 93 static int set_txreq_header_ahg(struct user_sdma_request *req, 355 struct user_sdma_request *req; in hfi1_user_sdma_process_request() local 362 if (iovec[idx].iov_len < sizeof(info) + sizeof(req->hdr)) { in hfi1_user_sdma_process_request() 367 iovec[idx].iov_len, sizeof(info) + sizeof(req->hdr)); in hfi1_user_sdma_process_request() 417 req = pq->reqs + info.comp_idx; in hfi1_user_sdma_process_request() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/crypto/amcc/ |
| H A D | crypto4xx_alg.c | 69 static inline int crypto4xx_crypt(struct skcipher_request *req, in crypto4xx_crypt() argument 73 struct crypto_skcipher *cipher = crypto_skcipher_reqtfm(req); in crypto4xx_crypt() 77 if (check_blocksize && !IS_ALIGNED(req->cryptlen, AES_BLOCK_SIZE)) in crypto4xx_crypt() 81 crypto4xx_memcpy_to_le32(iv, req->iv, ivlen); in crypto4xx_crypt() 83 return crypto4xx_build_pd(&req->base, ctx, req->src, req->dst, in crypto4xx_crypt() 84 req->cryptlen, iv, ivlen, decrypt ? ctx->sa_in : ctx->sa_out, in crypto4xx_crypt() 88 int crypto4xx_encrypt_noiv_block(struct skcipher_request *req) in crypto4xx_encrypt_noiv_block() argument 90 return crypto4xx_crypt(req, 0, false, true); in crypto4xx_encrypt_noiv_block() 93 int crypto4xx_encrypt_iv_stream(struct skcipher_request *req) in crypto4xx_encrypt_iv_stream() argument 95 return crypto4xx_crypt(req, AES_IV_SIZE, false, false); in crypto4xx_encrypt_iv_stream() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/crypto/marvell/cesa/ |
| H A D | cipher.c | 42 struct skcipher_request *req) in mv_cesa_skcipher_req_iter_init() argument 44 mv_cesa_req_dma_iter_init(&iter->base, req->cryptlen); in mv_cesa_skcipher_req_iter_init() 45 mv_cesa_sg_dma_iter_init(&iter->src, req->src, DMA_TO_DEVICE); in mv_cesa_skcipher_req_iter_init() 46 mv_cesa_sg_dma_iter_init(&iter->dst, req->dst, DMA_FROM_DEVICE); in mv_cesa_skcipher_req_iter_init() 59 mv_cesa_skcipher_dma_cleanup(struct skcipher_request *req) in mv_cesa_skcipher_dma_cleanup() argument 61 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); in mv_cesa_skcipher_dma_cleanup() 63 if (req->dst != req->src) { in mv_cesa_skcipher_dma_cleanup() 64 dma_unmap_sg(cesa_dev->dev, req->dst, creq->dst_nents, in mv_cesa_skcipher_dma_cleanup() 66 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_skcipher_dma_cleanup() 69 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_skcipher_dma_cleanup() [all …]
|
| H A D | hash.c | 27 struct ahash_request *req) in mv_cesa_ahash_req_iter_init() argument 29 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); in mv_cesa_ahash_req_iter_init() 30 unsigned int len = req->nbytes + creq->cache_ptr; in mv_cesa_ahash_req_iter_init() 36 mv_cesa_sg_dma_iter_init(&iter->src, req->src, DMA_TO_DEVICE); in mv_cesa_ahash_req_iter_init() 49 mv_cesa_ahash_dma_alloc_cache(struct mv_cesa_ahash_dma_req *req, gfp_t flags) in mv_cesa_ahash_dma_alloc_cache() argument 51 req->cache = dma_pool_alloc(cesa_dev->dma->cache_pool, flags, in mv_cesa_ahash_dma_alloc_cache() 52 &req->cache_dma); in mv_cesa_ahash_dma_alloc_cache() 53 if (!req->cache) in mv_cesa_ahash_dma_alloc_cache() 60 mv_cesa_ahash_dma_free_cache(struct mv_cesa_ahash_dma_req *req) in mv_cesa_ahash_dma_free_cache() argument 62 if (!req->cache) in mv_cesa_ahash_dma_free_cache() [all …]
|
| /OK3568_Linux_fs/kernel/fs/nfs/ |
| H A D | pagelist.c | 64 hdr->req = nfs_list_entry(mirror->pg_list.next); in nfs_pgheader_init() 66 hdr->cred = nfs_req_openctx(hdr->req)->cred; in nfs_pgheader_init() 67 hdr->io_start = req_offset(hdr->req); in nfs_pgheader_init() 153 * @req: any member of the page group 156 nfs_page_group_lock_head(struct nfs_page *req) in nfs_page_group_lock_head() argument 158 struct nfs_page *head = req->wb_head; in nfs_page_group_lock_head() 165 if (head != req) in nfs_page_group_lock_head() 171 * nfs_unroll_locks - unlock all newly locked reqs and wait on @req 173 * @req: request that couldn't lock and needs to wait on the req bit lock 179 nfs_unroll_locks(struct nfs_page *head, struct nfs_page *req) in nfs_unroll_locks() argument [all …]
|
| /OK3568_Linux_fs/kernel/drivers/video/rockchip/rga/ |
| H A D | rga_drv.c | 110 static int rga_blit_async(rga_session *session, struct rga_req *req); 299 static void print_debug_info(struct rga_req *req) in print_debug_info() argument 302 rga_get_render_mode_str(req->render_mode), in print_debug_info() 303 rga_get_rotate_mode_str(req), req->bsfilter_flag); in print_debug_info() 305 req->src.yrgb_addr, req->src.uv_addr, req->src.v_addr, in print_debug_info() 306 rga_get_format_name(req->src.format), in print_debug_info() 307 req->src.act_w, req->src.act_h, req->src.vir_w, req->src.vir_h, in print_debug_info() 308 req->src.x_offset, req->src.y_offset); in print_debug_info() 310 req->dst.yrgb_addr, req->dst.uv_addr, req->dst.v_addr, in print_debug_info() 311 rga_get_format_name(req->dst.format), in print_debug_info() [all …]
|
| /OK3568_Linux_fs/kernel/io_uring/ |
| H A D | io_uring.c | 518 struct io_kiocb *req; member 814 typedef void (*io_req_tw_func_t)(struct io_kiocb *req, bool *locked); 909 struct io_kiocb *req; member 914 /* needs req->file assigned */ 1083 static bool io_disarm_next(struct io_kiocb *req); 1090 static void io_fill_cqe_req(struct io_kiocb *req, s32 res, u32 cflags); 1092 static void io_put_req(struct io_kiocb *req); 1093 static void io_put_req_deferred(struct io_kiocb *req); 1094 static void io_dismantle_req(struct io_kiocb *req); 1095 static void io_queue_linked_timeout(struct io_kiocb *req); [all …]
|
| /OK3568_Linux_fs/kernel/arch/arm/crypto/ |
| H A D | ghash-ce-core.S | 14 SHASH .req q0 15 T1 .req q1 16 XL .req q2 17 XM .req q3 18 XH .req q4 19 IN1 .req q4 21 SHASH_L .req d0 22 SHASH_H .req d1 23 T1_L .req d2 24 T1_H .req d3 [all …]
|