Lines Matching refs:dma

85 	struct dma_controller *dma = musb->dma_controller;  in map_dma_buffer()  local
89 if (!is_dma_capable() || !musb_ep->dma) in map_dma_buffer()
96 if (dma->is_compatible) in map_dma_buffer()
97 compatible = dma->is_compatible(musb_ep->dma, in map_dma_buffer()
103 if (request->request.dma == DMA_ADDR_INVALID) { in map_dma_buffer()
104 request->request.dma = dma_map_single( in map_dma_buffer()
114 request->request.dma, in map_dma_buffer()
130 if (request->request.dma == DMA_ADDR_INVALID) { in unmap_dma_buffer()
137 request->request.dma, in unmap_dma_buffer()
142 request->request.dma = DMA_ADDR_INVALID; in unmap_dma_buffer()
145 request->request.dma, in unmap_dma_buffer()
221 if (is_dma_capable() && ep->dma) { in nuke()
242 value = c->channel_abort(ep->dma); in nuke()
245 c->channel_release(ep->dma); in nuke()
246 ep->dma = NULL; in nuke()
333 if (dma_channel_status(musb_ep->dma) == MUSB_DMA_STATUS_BUSY) { in txstate()
368 musb_ep->dma->max_len); in txstate()
370 use_dma = (request->dma != DMA_ADDR_INVALID); in txstate()
377 musb_ep->dma->desired_mode = 0; in txstate()
379 musb_ep->dma->desired_mode = 1; in txstate()
382 musb_ep->dma, musb_ep->packet_sz, in txstate()
383 musb_ep->dma->desired_mode, in txstate()
384 request->dma + request->actual, request_size); in txstate()
386 if (musb_ep->dma->desired_mode == 0) { in txstate()
437 musb_ep->dma, musb_ep->packet_sz, in txstate()
439 request->dma + request->actual, in txstate()
442 c->channel_release(musb_ep->dma); in txstate()
443 musb_ep->dma = NULL; in txstate()
450 musb_ep->dma, musb_ep->packet_sz, in txstate()
452 request->dma + request->actual, in txstate()
494 struct dma_channel *dma; in musb_g_tx() local
503 dma = is_dma_capable() ? musb_ep->dma : NULL; in musb_g_tx()
525 if (dma_channel_status(dma) == MUSB_DMA_STATUS_BUSY) { in musb_g_tx()
537 if (dma && (csr & MUSB_TXCSR_DMAENAB)) { in musb_g_tx()
545 request->actual += musb_ep->dma->actual_len; in musb_g_tx()
547 epnum, csr, musb_ep->dma->actual_len, request); in musb_g_tx()
558 || (is_dma && (!dma->desired_mode || in musb_g_tx()
662 if (dma_channel_status(musb_ep->dma) == MUSB_DMA_STATUS_BUSY) { in rxstate()
675 struct dma_channel *channel = musb_ep->dma; in rxstate()
685 request->dma + request->actual, in rxstate()
722 channel = musb_ep->dma; in rxstate()
774 musb_ep->dma->desired_mode = 1; in rxstate()
778 musb_ep->dma->desired_mode = 0; in rxstate()
785 request->dma in rxstate()
802 channel = musb_ep->dma; in rxstate()
823 musb_ep->dma->desired_mode = 0; in rxstate()
825 musb_ep->dma->desired_mode = 1; in rxstate()
834 request->dma in rxstate()
853 struct dma_channel *channel = musb_ep->dma; in rxstate()
854 u32 dma_addr = request->dma + request->actual; in rxstate()
913 struct dma_channel *dma; in musb_g_rx() local
930 dma = is_dma_capable() ? musb_ep->dma : NULL; in musb_g_rx()
933 csr, dma ? " (dma)" : "", request); in musb_g_rx()
956 if (dma_channel_status(dma) == MUSB_DMA_STATUS_BUSY) { in musb_g_rx()
963 if (dma && (csr & MUSB_RXCSR_DMAENAB)) { in musb_g_rx()
970 request->actual += musb_ep->dma->actual_len; in musb_g_rx()
975 musb_ep->dma->actual_len, request); in musb_g_rx()
980 if ((dma->desired_mode == 0 && !hw_ep->rx_double_buffered) in musb_g_rx()
981 || (dma->actual_len in musb_g_rx()
990 && (musb_ep->dma->actual_len in musb_g_rx()
1181 musb_ep->dma = c->channel_alloc(c, hw_ep, in musb_gadget_enable()
1184 musb_ep->dma = NULL; in musb_gadget_enable()
1199 musb_ep->dma ? "dma, " : "", in musb_gadget_enable()
1275 request->request.dma = DMA_ADDR_INVALID; in musb_alloc_request()
1297 dma_addr_t dma; member
1400 else if (is_dma_capable() && musb_ep->dma) { in musb_gadget_dequeue()
1405 status = c->channel_abort(musb_ep->dma); in musb_gadget_dequeue()