Lines Matching refs:hwreq
354 static int add_td_to_list(struct ci_hw_ep *hwep, struct ci_hw_req *hwreq, in add_td_to_list() argument
375 u32 mul = hwreq->req.length / hwep->ep.maxpacket; in add_td_to_list()
377 if (hwreq->req.length == 0 in add_td_to_list()
378 || hwreq->req.length % hwep->ep.maxpacket) in add_td_to_list()
384 temp = (u32) (sg_dma_address(s) + hwreq->req.actual); in add_td_to_list()
387 temp = (u32) (hwreq->req.dma + hwreq->req.actual); in add_td_to_list()
399 hwreq->req.actual += length; in add_td_to_list()
401 if (!list_empty(&hwreq->tds)) { in add_td_to_list()
403 lastnode = list_entry(hwreq->tds.prev, in add_td_to_list()
409 list_add_tail(&node->td, &hwreq->tds); in add_td_to_list()
424 struct ci_hw_req *hwreq) in prepare_td_for_non_sg() argument
426 unsigned int rest = hwreq->req.length; in prepare_td_for_non_sg()
431 ret = add_td_to_list(hwep, hwreq, 0, NULL); in prepare_td_for_non_sg()
440 if (hwreq->req.dma % PAGE_SIZE) in prepare_td_for_non_sg()
444 unsigned int count = min(hwreq->req.length - hwreq->req.actual, in prepare_td_for_non_sg()
447 ret = add_td_to_list(hwep, hwreq, count, NULL); in prepare_td_for_non_sg()
454 if (hwreq->req.zero && hwreq->req.length && hwep->dir == TX in prepare_td_for_non_sg()
455 && (hwreq->req.length % hwep->ep.maxpacket == 0)) { in prepare_td_for_non_sg()
456 ret = add_td_to_list(hwep, hwreq, 0, NULL); in prepare_td_for_non_sg()
464 static int prepare_td_per_sg(struct ci_hw_ep *hwep, struct ci_hw_req *hwreq, in prepare_td_per_sg() argument
470 hwreq->req.actual = 0; in prepare_td_per_sg()
475 ret = add_td_to_list(hwep, hwreq, count, s); in prepare_td_per_sg()
504 static int prepare_td_for_sg(struct ci_hw_ep *hwep, struct ci_hw_req *hwreq) in prepare_td_for_sg() argument
506 struct usb_request *req = &hwreq->req; in prepare_td_for_sg()
526 ret = prepare_td_per_sg(hwep, hwreq, s); in prepare_td_for_sg()
530 node = list_entry(hwreq->tds.prev, in prepare_td_for_sg()
547 static int _hardware_enqueue(struct ci_hw_ep *hwep, struct ci_hw_req *hwreq) in _hardware_enqueue() argument
554 if (hwreq->req.status == -EALREADY) in _hardware_enqueue()
557 hwreq->req.status = -EALREADY; in _hardware_enqueue()
560 &hwreq->req, hwep->dir); in _hardware_enqueue()
564 if (hwreq->req.num_mapped_sgs) in _hardware_enqueue()
565 ret = prepare_td_for_sg(hwep, hwreq); in _hardware_enqueue()
567 ret = prepare_td_for_non_sg(hwep, hwreq); in _hardware_enqueue()
572 firstnode = list_first_entry(&hwreq->tds, struct td_node, td); in _hardware_enqueue()
574 lastnode = list_entry(hwreq->tds.prev, in _hardware_enqueue()
578 if (!hwreq->req.no_interrupt) in _hardware_enqueue()
582 hwreq->req.actual = 0; in _hardware_enqueue()
614 u32 mul = hwreq->req.length / hwep->ep.maxpacket; in _hardware_enqueue()
616 if (hwreq->req.length == 0 in _hardware_enqueue()
617 || hwreq->req.length % hwep->ep.maxpacket) in _hardware_enqueue()
659 static int _hardware_dequeue(struct ci_hw_ep *hwep, struct ci_hw_req *hwreq) in _hardware_dequeue() argument
664 unsigned actual = hwreq->req.length; in _hardware_dequeue()
667 if (hwreq->req.status != -EALREADY) in _hardware_dequeue()
670 hwreq->req.status = 0; in _hardware_dequeue()
672 list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { in _hardware_dequeue()
680 hwreq->req.status = -EALREADY; in _hardware_dequeue()
688 hwreq->req.status = tmptoken & TD_STATUS; in _hardware_dequeue()
689 if ((TD_STATUS_HALTED & hwreq->req.status)) { in _hardware_dequeue()
690 hwreq->req.status = -EPIPE; in _hardware_dequeue()
692 } else if ((TD_STATUS_DT_ERR & hwreq->req.status)) { in _hardware_dequeue()
693 hwreq->req.status = -EPROTO; in _hardware_dequeue()
695 } else if ((TD_STATUS_TR_ERR & hwreq->req.status)) { in _hardware_dequeue()
696 hwreq->req.status = -EILSEQ; in _hardware_dequeue()
702 hwreq->req.status = -EPROTO; in _hardware_dequeue()
719 &hwreq->req, hwep->dir); in _hardware_dequeue()
721 hwreq->req.actual += actual; in _hardware_dequeue()
723 if (hwreq->req.status) in _hardware_dequeue()
724 return hwreq->req.status; in _hardware_dequeue()
726 return hwreq->req.actual; in _hardware_dequeue()
749 struct ci_hw_req *hwreq = list_entry(hwep->qh.queue.next, in _ep_nuke() local
752 list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { in _ep_nuke()
759 list_del_init(&hwreq->queue); in _ep_nuke()
760 hwreq->req.status = -ESHUTDOWN; in _ep_nuke()
762 if (hwreq->req.complete != NULL) { in _ep_nuke()
764 usb_gadget_giveback_request(&hwep->ep, &hwreq->req); in _ep_nuke()
918 struct ci_hw_req *hwreq = container_of(req, struct ci_hw_req, req); in _ep_queue() local
937 hwreq->req.length > hwep->ep.mult * hwep->ep.maxpacket) { in _ep_queue()
943 if (!list_empty(&hwreq->queue)) { in _ep_queue()
949 hwreq->req.status = -EINPROGRESS; in _ep_queue()
950 hwreq->req.actual = 0; in _ep_queue()
952 retval = _hardware_enqueue(hwep, hwreq); in _ep_queue()
957 list_add_tail(&hwreq->queue, &hwep->qh.queue); in _ep_queue()
1090 struct ci_hw_req *hwreq, *hwreqtemp; in isr_tr_complete_low() local
1094 list_for_each_entry_safe(hwreq, hwreqtemp, &hwep->qh.queue, in isr_tr_complete_low()
1096 retval = _hardware_dequeue(hwep, hwreq); in isr_tr_complete_low()
1099 list_del_init(&hwreq->queue); in isr_tr_complete_low()
1100 if (hwreq->req.complete != NULL) { in isr_tr_complete_low()
1103 hwreq->req.length) in isr_tr_complete_low()
1105 usb_gadget_giveback_request(&hweptemp->ep, &hwreq->req); in isr_tr_complete_low()
1452 struct ci_hw_req *hwreq = NULL; in ep_alloc_request() local
1457 hwreq = kzalloc(sizeof(struct ci_hw_req), gfp_flags); in ep_alloc_request()
1458 if (hwreq != NULL) { in ep_alloc_request()
1459 INIT_LIST_HEAD(&hwreq->queue); in ep_alloc_request()
1460 INIT_LIST_HEAD(&hwreq->tds); in ep_alloc_request()
1463 return (hwreq == NULL) ? NULL : &hwreq->req; in ep_alloc_request()
1474 struct ci_hw_req *hwreq = container_of(req, struct ci_hw_req, req); in ep_free_request() local
1480 } else if (!list_empty(&hwreq->queue)) { in ep_free_request()
1487 list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { in ep_free_request()
1494 kfree(hwreq); in ep_free_request()
1532 struct ci_hw_req *hwreq = container_of(req, struct ci_hw_req, req); in ep_dequeue() local
1536 if (ep == NULL || req == NULL || hwreq->req.status != -EALREADY || in ep_dequeue()
1537 hwep->ep.desc == NULL || list_empty(&hwreq->queue) || in ep_dequeue()
1545 list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { in ep_dequeue()
1552 list_del_init(&hwreq->queue); in ep_dequeue()
1558 if (hwreq->req.complete != NULL) { in ep_dequeue()
1560 usb_gadget_giveback_request(&hwep->ep, &hwreq->req); in ep_dequeue()