Lines Matching refs:pd_chan

164 struct pch_dma_desc *pdc_first_active(struct pch_dma_chan *pd_chan)  in pdc_first_active()  argument
166 return list_first_entry(&pd_chan->active_list, in pdc_first_active()
171 struct pch_dma_desc *pdc_first_queued(struct pch_dma_chan *pd_chan) in pdc_first_queued() argument
173 return list_first_entry(&pd_chan->queue, in pdc_first_queued()
203 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pdc_set_dir() local
217 if (pd_chan->dir == DMA_MEM_TO_DEV) in pdc_set_dir()
235 if (pd_chan->dir == DMA_MEM_TO_DEV) in pdc_set_dir()
283 static u32 pdc_get_status0(struct pch_dma_chan *pd_chan) in pdc_get_status0() argument
285 struct pch_dma *pd = to_pd(pd_chan->chan.device); in pdc_get_status0()
290 DMA_STATUS_BITS_PER_CH * pd_chan->chan.chan_id)); in pdc_get_status0()
293 static u32 pdc_get_status2(struct pch_dma_chan *pd_chan) in pdc_get_status2() argument
295 struct pch_dma *pd = to_pd(pd_chan->chan.device); in pdc_get_status2()
300 DMA_STATUS_BITS_PER_CH * (pd_chan->chan.chan_id - 8))); in pdc_get_status2()
303 static bool pdc_is_idle(struct pch_dma_chan *pd_chan) in pdc_is_idle() argument
307 if (pd_chan->chan.chan_id < 8) in pdc_is_idle()
308 sts = pdc_get_status0(pd_chan); in pdc_is_idle()
310 sts = pdc_get_status2(pd_chan); in pdc_is_idle()
319 static void pdc_dostart(struct pch_dma_chan *pd_chan, struct pch_dma_desc* desc) in pdc_dostart() argument
321 if (!pdc_is_idle(pd_chan)) { in pdc_dostart()
322 dev_err(chan2dev(&pd_chan->chan), in pdc_dostart()
327 dev_dbg(chan2dev(&pd_chan->chan), "chan %d -> dev_addr: %x\n", in pdc_dostart()
328 pd_chan->chan.chan_id, desc->regs.dev_addr); in pdc_dostart()
329 dev_dbg(chan2dev(&pd_chan->chan), "chan %d -> mem_addr: %x\n", in pdc_dostart()
330 pd_chan->chan.chan_id, desc->regs.mem_addr); in pdc_dostart()
331 dev_dbg(chan2dev(&pd_chan->chan), "chan %d -> size: %x\n", in pdc_dostart()
332 pd_chan->chan.chan_id, desc->regs.size); in pdc_dostart()
333 dev_dbg(chan2dev(&pd_chan->chan), "chan %d -> next: %x\n", in pdc_dostart()
334 pd_chan->chan.chan_id, desc->regs.next); in pdc_dostart()
337 channel_writel(pd_chan, DEV_ADDR, desc->regs.dev_addr); in pdc_dostart()
338 channel_writel(pd_chan, MEM_ADDR, desc->regs.mem_addr); in pdc_dostart()
339 channel_writel(pd_chan, SIZE, desc->regs.size); in pdc_dostart()
340 channel_writel(pd_chan, NEXT, desc->regs.next); in pdc_dostart()
341 pdc_set_mode(&pd_chan->chan, DMA_CTL0_ONESHOT); in pdc_dostart()
343 channel_writel(pd_chan, NEXT, desc->txd.phys); in pdc_dostart()
344 pdc_set_mode(&pd_chan->chan, DMA_CTL0_SG); in pdc_dostart()
348 static void pdc_chain_complete(struct pch_dma_chan *pd_chan, in pdc_chain_complete() argument
355 list_splice_init(&desc->tx_list, &pd_chan->free_list); in pdc_chain_complete()
356 list_move(&desc->desc_node, &pd_chan->free_list); in pdc_chain_complete()
361 static void pdc_complete_all(struct pch_dma_chan *pd_chan) in pdc_complete_all() argument
366 BUG_ON(!pdc_is_idle(pd_chan)); in pdc_complete_all()
368 if (!list_empty(&pd_chan->queue)) in pdc_complete_all()
369 pdc_dostart(pd_chan, pdc_first_queued(pd_chan)); in pdc_complete_all()
371 list_splice_init(&pd_chan->active_list, &list); in pdc_complete_all()
372 list_splice_init(&pd_chan->queue, &pd_chan->active_list); in pdc_complete_all()
375 pdc_chain_complete(pd_chan, desc); in pdc_complete_all()
378 static void pdc_handle_error(struct pch_dma_chan *pd_chan) in pdc_handle_error() argument
382 bad_desc = pdc_first_active(pd_chan); in pdc_handle_error()
385 list_splice_init(&pd_chan->queue, pd_chan->active_list.prev); in pdc_handle_error()
387 if (!list_empty(&pd_chan->active_list)) in pdc_handle_error()
388 pdc_dostart(pd_chan, pdc_first_active(pd_chan)); in pdc_handle_error()
390 dev_crit(chan2dev(&pd_chan->chan), "Bad descriptor submitted\n"); in pdc_handle_error()
391 dev_crit(chan2dev(&pd_chan->chan), "descriptor cookie: %d\n", in pdc_handle_error()
394 pdc_chain_complete(pd_chan, bad_desc); in pdc_handle_error()
397 static void pdc_advance_work(struct pch_dma_chan *pd_chan) in pdc_advance_work() argument
399 if (list_empty(&pd_chan->active_list) || in pdc_advance_work()
400 list_is_singular(&pd_chan->active_list)) { in pdc_advance_work()
401 pdc_complete_all(pd_chan); in pdc_advance_work()
403 pdc_chain_complete(pd_chan, pdc_first_active(pd_chan)); in pdc_advance_work()
404 pdc_dostart(pd_chan, pdc_first_active(pd_chan)); in pdc_advance_work()
411 struct pch_dma_chan *pd_chan = to_pd_chan(txd->chan); in pd_tx_submit() local
413 spin_lock(&pd_chan->lock); in pd_tx_submit()
415 if (list_empty(&pd_chan->active_list)) { in pd_tx_submit()
416 list_add_tail(&desc->desc_node, &pd_chan->active_list); in pd_tx_submit()
417 pdc_dostart(pd_chan, desc); in pd_tx_submit()
419 list_add_tail(&desc->desc_node, &pd_chan->queue); in pd_tx_submit()
422 spin_unlock(&pd_chan->lock); in pd_tx_submit()
444 static struct pch_dma_desc *pdc_desc_get(struct pch_dma_chan *pd_chan) in pdc_desc_get() argument
450 spin_lock(&pd_chan->lock); in pdc_desc_get()
451 list_for_each_entry_safe(desc, _d, &pd_chan->free_list, desc_node) { in pdc_desc_get()
458 dev_dbg(chan2dev(&pd_chan->chan), "desc %p not ACKed\n", desc); in pdc_desc_get()
460 spin_unlock(&pd_chan->lock); in pdc_desc_get()
461 dev_dbg(chan2dev(&pd_chan->chan), "scanned %d descriptors\n", i); in pdc_desc_get()
464 ret = pdc_alloc_desc(&pd_chan->chan, GFP_ATOMIC); in pdc_desc_get()
466 spin_lock(&pd_chan->lock); in pdc_desc_get()
467 pd_chan->descs_allocated++; in pdc_desc_get()
468 spin_unlock(&pd_chan->lock); in pdc_desc_get()
470 dev_err(chan2dev(&pd_chan->chan), in pdc_desc_get()
478 static void pdc_desc_put(struct pch_dma_chan *pd_chan, in pdc_desc_put() argument
482 spin_lock(&pd_chan->lock); in pdc_desc_put()
483 list_splice_init(&desc->tx_list, &pd_chan->free_list); in pdc_desc_put()
484 list_add(&desc->desc_node, &pd_chan->free_list); in pdc_desc_put()
485 spin_unlock(&pd_chan->lock); in pdc_desc_put()
491 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pd_alloc_chan_resources() local
496 if (!pdc_is_idle(pd_chan)) { in pd_alloc_chan_resources()
501 if (!list_empty(&pd_chan->free_list)) in pd_alloc_chan_resources()
502 return pd_chan->descs_allocated; in pd_alloc_chan_resources()
516 spin_lock_irq(&pd_chan->lock); in pd_alloc_chan_resources()
517 list_splice(&tmp_list, &pd_chan->free_list); in pd_alloc_chan_resources()
518 pd_chan->descs_allocated = i; in pd_alloc_chan_resources()
520 spin_unlock_irq(&pd_chan->lock); in pd_alloc_chan_resources()
524 return pd_chan->descs_allocated; in pd_alloc_chan_resources()
529 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pd_free_chan_resources() local
534 BUG_ON(!pdc_is_idle(pd_chan)); in pd_free_chan_resources()
535 BUG_ON(!list_empty(&pd_chan->active_list)); in pd_free_chan_resources()
536 BUG_ON(!list_empty(&pd_chan->queue)); in pd_free_chan_resources()
538 spin_lock_irq(&pd_chan->lock); in pd_free_chan_resources()
539 list_splice_init(&pd_chan->free_list, &tmp_list); in pd_free_chan_resources()
540 pd_chan->descs_allocated = 0; in pd_free_chan_resources()
541 spin_unlock_irq(&pd_chan->lock); in pd_free_chan_resources()
557 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pd_issue_pending() local
559 if (pdc_is_idle(pd_chan)) { in pd_issue_pending()
560 spin_lock(&pd_chan->lock); in pd_issue_pending()
561 pdc_advance_work(pd_chan); in pd_issue_pending()
562 spin_unlock(&pd_chan->lock); in pd_issue_pending()
571 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pd_prep_slave_sg() local
592 pd_chan->dir = direction; in pd_prep_slave_sg()
596 desc = pdc_desc_get(pd_chan); in pd_prep_slave_sg()
648 pdc_desc_put(pd_chan, first); in pd_prep_slave_sg()
654 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pd_device_terminate_all() local
658 spin_lock_irq(&pd_chan->lock); in pd_device_terminate_all()
660 pdc_set_mode(&pd_chan->chan, DMA_CTL0_DISABLE); in pd_device_terminate_all()
662 list_splice_init(&pd_chan->active_list, &list); in pd_device_terminate_all()
663 list_splice_init(&pd_chan->queue, &list); in pd_device_terminate_all()
666 pdc_chain_complete(pd_chan, desc); in pd_device_terminate_all()
668 spin_unlock_irq(&pd_chan->lock); in pd_device_terminate_all()
675 struct pch_dma_chan *pd_chan = from_tasklet(pd_chan, t, tasklet); in pdc_tasklet() local
678 if (!pdc_is_idle(pd_chan)) { in pdc_tasklet()
679 dev_err(chan2dev(&pd_chan->chan), in pdc_tasklet()
684 spin_lock_irqsave(&pd_chan->lock, flags); in pdc_tasklet()
685 if (test_and_clear_bit(0, &pd_chan->err_status)) in pdc_tasklet()
686 pdc_handle_error(pd_chan); in pdc_tasklet()
688 pdc_advance_work(pd_chan); in pdc_tasklet()
689 spin_unlock_irqrestore(&pd_chan->lock, flags); in pdc_tasklet()
695 struct pch_dma_chan *pd_chan; in pd_irq() local
708 pd_chan = &pd->channels[i]; in pd_irq()
713 set_bit(0, &pd_chan->err_status); in pd_irq()
715 tasklet_schedule(&pd_chan->tasklet); in pd_irq()
721 set_bit(0, &pd_chan->err_status); in pd_irq()
723 tasklet_schedule(&pd_chan->tasklet); in pd_irq()
740 struct pch_dma_chan *pd_chan; in pch_dma_save_regs() local
750 pd_chan = to_pd_chan(chan); in pch_dma_save_regs()
752 pd->ch_regs[i].dev_addr = channel_readl(pd_chan, DEV_ADDR); in pch_dma_save_regs()
753 pd->ch_regs[i].mem_addr = channel_readl(pd_chan, MEM_ADDR); in pch_dma_save_regs()
754 pd->ch_regs[i].size = channel_readl(pd_chan, SIZE); in pch_dma_save_regs()
755 pd->ch_regs[i].next = channel_readl(pd_chan, NEXT); in pch_dma_save_regs()
763 struct pch_dma_chan *pd_chan; in pch_dma_restore_regs() local
773 pd_chan = to_pd_chan(chan); in pch_dma_restore_regs()
775 channel_writel(pd_chan, DEV_ADDR, pd->ch_regs[i].dev_addr); in pch_dma_restore_regs()
776 channel_writel(pd_chan, MEM_ADDR, pd->ch_regs[i].mem_addr); in pch_dma_restore_regs()
777 channel_writel(pd_chan, SIZE, pd->ch_regs[i].size); in pch_dma_restore_regs()
778 channel_writel(pd_chan, NEXT, pd->ch_regs[i].next); in pch_dma_restore_regs()
872 struct pch_dma_chan *pd_chan = &pd->channels[i]; in pch_dma_probe() local
874 pd_chan->chan.device = &pd->dma; in pch_dma_probe()
875 dma_cookie_init(&pd_chan->chan); in pch_dma_probe()
877 pd_chan->membase = &regs->desc[i]; in pch_dma_probe()
879 spin_lock_init(&pd_chan->lock); in pch_dma_probe()
881 INIT_LIST_HEAD(&pd_chan->active_list); in pch_dma_probe()
882 INIT_LIST_HEAD(&pd_chan->queue); in pch_dma_probe()
883 INIT_LIST_HEAD(&pd_chan->free_list); in pch_dma_probe()
885 tasklet_setup(&pd_chan->tasklet, pdc_tasklet); in pch_dma_probe()
886 list_add_tail(&pd_chan->chan.device_node, &pd->dma.channels); in pch_dma_probe()
926 struct pch_dma_chan *pd_chan; in pch_dma_remove() local
936 pd_chan = to_pd_chan(chan); in pch_dma_remove()
938 tasklet_kill(&pd_chan->tasklet); in pch_dma_remove()