Lines Matching refs:pd_chan
172 struct pch_dma_desc *pdc_first_active(struct pch_dma_chan *pd_chan) in pdc_first_active() argument
174 return list_first_entry(&pd_chan->active_list, in pdc_first_active()
179 struct pch_dma_desc *pdc_first_queued(struct pch_dma_chan *pd_chan) in pdc_first_queued() argument
181 return list_first_entry(&pd_chan->queue, in pdc_first_queued()
211 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pdc_set_dir() local
225 if (pd_chan->dir == DMA_MEM_TO_DEV) in pdc_set_dir()
243 if (pd_chan->dir == DMA_MEM_TO_DEV) in pdc_set_dir()
291 static u32 pdc_get_status0(struct pch_dma_chan *pd_chan) in pdc_get_status0() argument
293 struct pch_dma *pd = to_pd(pd_chan->chan.device); in pdc_get_status0()
298 DMA_STATUS_BITS_PER_CH * pd_chan->chan.chan_id)); in pdc_get_status0()
301 static u32 pdc_get_status2(struct pch_dma_chan *pd_chan) in pdc_get_status2() argument
303 struct pch_dma *pd = to_pd(pd_chan->chan.device); in pdc_get_status2()
308 DMA_STATUS_BITS_PER_CH * (pd_chan->chan.chan_id - 8))); in pdc_get_status2()
311 static bool pdc_is_idle(struct pch_dma_chan *pd_chan) in pdc_is_idle() argument
315 if (pd_chan->chan.chan_id < 8) in pdc_is_idle()
316 sts = pdc_get_status0(pd_chan); in pdc_is_idle()
318 sts = pdc_get_status2(pd_chan); in pdc_is_idle()
327 static void pdc_dostart(struct pch_dma_chan *pd_chan, struct pch_dma_desc* desc) in pdc_dostart() argument
329 if (!pdc_is_idle(pd_chan)) { in pdc_dostart()
330 dev_err(chan2dev(&pd_chan->chan), in pdc_dostart()
335 dev_dbg(chan2dev(&pd_chan->chan), "chan %d -> dev_addr: %x\n", in pdc_dostart()
336 pd_chan->chan.chan_id, desc->regs.dev_addr); in pdc_dostart()
337 dev_dbg(chan2dev(&pd_chan->chan), "chan %d -> mem_addr: %x\n", in pdc_dostart()
338 pd_chan->chan.chan_id, desc->regs.mem_addr); in pdc_dostart()
339 dev_dbg(chan2dev(&pd_chan->chan), "chan %d -> size: %x\n", in pdc_dostart()
340 pd_chan->chan.chan_id, desc->regs.size); in pdc_dostart()
341 dev_dbg(chan2dev(&pd_chan->chan), "chan %d -> next: %x\n", in pdc_dostart()
342 pd_chan->chan.chan_id, desc->regs.next); in pdc_dostart()
345 channel_writel(pd_chan, DEV_ADDR, desc->regs.dev_addr); in pdc_dostart()
346 channel_writel(pd_chan, MEM_ADDR, desc->regs.mem_addr); in pdc_dostart()
347 channel_writel(pd_chan, SIZE, desc->regs.size); in pdc_dostart()
348 channel_writel(pd_chan, NEXT, desc->regs.next); in pdc_dostart()
349 pdc_set_mode(&pd_chan->chan, DMA_CTL0_ONESHOT); in pdc_dostart()
351 channel_writel(pd_chan, NEXT, desc->txd.phys); in pdc_dostart()
352 pdc_set_mode(&pd_chan->chan, DMA_CTL0_SG); in pdc_dostart()
356 static void pdc_chain_complete(struct pch_dma_chan *pd_chan, in pdc_chain_complete() argument
363 list_splice_init(&desc->tx_list, &pd_chan->free_list); in pdc_chain_complete()
364 list_move(&desc->desc_node, &pd_chan->free_list); in pdc_chain_complete()
370 static void pdc_complete_all(struct pch_dma_chan *pd_chan) in pdc_complete_all() argument
375 BUG_ON(!pdc_is_idle(pd_chan)); in pdc_complete_all()
377 if (!list_empty(&pd_chan->queue)) in pdc_complete_all()
378 pdc_dostart(pd_chan, pdc_first_queued(pd_chan)); in pdc_complete_all()
380 list_splice_init(&pd_chan->active_list, &list); in pdc_complete_all()
381 list_splice_init(&pd_chan->queue, &pd_chan->active_list); in pdc_complete_all()
384 pdc_chain_complete(pd_chan, desc); in pdc_complete_all()
387 static void pdc_handle_error(struct pch_dma_chan *pd_chan) in pdc_handle_error() argument
391 bad_desc = pdc_first_active(pd_chan); in pdc_handle_error()
394 list_splice_init(&pd_chan->queue, pd_chan->active_list.prev); in pdc_handle_error()
396 if (!list_empty(&pd_chan->active_list)) in pdc_handle_error()
397 pdc_dostart(pd_chan, pdc_first_active(pd_chan)); in pdc_handle_error()
399 dev_crit(chan2dev(&pd_chan->chan), "Bad descriptor submitted\n"); in pdc_handle_error()
400 dev_crit(chan2dev(&pd_chan->chan), "descriptor cookie: %d\n", in pdc_handle_error()
403 pdc_chain_complete(pd_chan, bad_desc); in pdc_handle_error()
406 static void pdc_advance_work(struct pch_dma_chan *pd_chan) in pdc_advance_work() argument
408 if (list_empty(&pd_chan->active_list) || in pdc_advance_work()
409 list_is_singular(&pd_chan->active_list)) { in pdc_advance_work()
410 pdc_complete_all(pd_chan); in pdc_advance_work()
412 pdc_chain_complete(pd_chan, pdc_first_active(pd_chan)); in pdc_advance_work()
413 pdc_dostart(pd_chan, pdc_first_active(pd_chan)); in pdc_advance_work()
420 struct pch_dma_chan *pd_chan = to_pd_chan(txd->chan); in pd_tx_submit() local
423 spin_lock(&pd_chan->lock); in pd_tx_submit()
426 if (list_empty(&pd_chan->active_list)) { in pd_tx_submit()
427 list_add_tail(&desc->desc_node, &pd_chan->active_list); in pd_tx_submit()
428 pdc_dostart(pd_chan, desc); in pd_tx_submit()
430 list_add_tail(&desc->desc_node, &pd_chan->queue); in pd_tx_submit()
433 spin_unlock(&pd_chan->lock); in pd_tx_submit()
456 static struct pch_dma_desc *pdc_desc_get(struct pch_dma_chan *pd_chan) in pdc_desc_get() argument
462 spin_lock(&pd_chan->lock); in pdc_desc_get()
463 list_for_each_entry_safe(desc, _d, &pd_chan->free_list, desc_node) { in pdc_desc_get()
470 dev_dbg(chan2dev(&pd_chan->chan), "desc %p not ACKed\n", desc); in pdc_desc_get()
472 spin_unlock(&pd_chan->lock); in pdc_desc_get()
473 dev_dbg(chan2dev(&pd_chan->chan), "scanned %d descriptors\n", i); in pdc_desc_get()
476 ret = pdc_alloc_desc(&pd_chan->chan, GFP_ATOMIC); in pdc_desc_get()
478 spin_lock(&pd_chan->lock); in pdc_desc_get()
479 pd_chan->descs_allocated++; in pdc_desc_get()
480 spin_unlock(&pd_chan->lock); in pdc_desc_get()
482 dev_err(chan2dev(&pd_chan->chan), in pdc_desc_get()
490 static void pdc_desc_put(struct pch_dma_chan *pd_chan, in pdc_desc_put() argument
494 spin_lock(&pd_chan->lock); in pdc_desc_put()
495 list_splice_init(&desc->tx_list, &pd_chan->free_list); in pdc_desc_put()
496 list_add(&desc->desc_node, &pd_chan->free_list); in pdc_desc_put()
497 spin_unlock(&pd_chan->lock); in pdc_desc_put()
503 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pd_alloc_chan_resources() local
508 if (!pdc_is_idle(pd_chan)) { in pd_alloc_chan_resources()
513 if (!list_empty(&pd_chan->free_list)) in pd_alloc_chan_resources()
514 return pd_chan->descs_allocated; in pd_alloc_chan_resources()
528 spin_lock_irq(&pd_chan->lock); in pd_alloc_chan_resources()
529 list_splice(&tmp_list, &pd_chan->free_list); in pd_alloc_chan_resources()
530 pd_chan->descs_allocated = i; in pd_alloc_chan_resources()
532 spin_unlock_irq(&pd_chan->lock); in pd_alloc_chan_resources()
536 return pd_chan->descs_allocated; in pd_alloc_chan_resources()
541 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pd_free_chan_resources() local
546 BUG_ON(!pdc_is_idle(pd_chan)); in pd_free_chan_resources()
547 BUG_ON(!list_empty(&pd_chan->active_list)); in pd_free_chan_resources()
548 BUG_ON(!list_empty(&pd_chan->queue)); in pd_free_chan_resources()
550 spin_lock_irq(&pd_chan->lock); in pd_free_chan_resources()
551 list_splice_init(&pd_chan->free_list, &tmp_list); in pd_free_chan_resources()
552 pd_chan->descs_allocated = 0; in pd_free_chan_resources()
553 spin_unlock_irq(&pd_chan->lock); in pd_free_chan_resources()
569 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pd_issue_pending() local
571 if (pdc_is_idle(pd_chan)) { in pd_issue_pending()
572 spin_lock(&pd_chan->lock); in pd_issue_pending()
573 pdc_advance_work(pd_chan); in pd_issue_pending()
574 spin_unlock(&pd_chan->lock); in pd_issue_pending()
583 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pd_prep_slave_sg() local
604 pd_chan->dir = direction; in pd_prep_slave_sg()
608 desc = pdc_desc_get(pd_chan); in pd_prep_slave_sg()
660 pdc_desc_put(pd_chan, first); in pd_prep_slave_sg()
666 struct pch_dma_chan *pd_chan = to_pd_chan(chan); in pd_device_terminate_all() local
670 spin_lock_irq(&pd_chan->lock); in pd_device_terminate_all()
672 pdc_set_mode(&pd_chan->chan, DMA_CTL0_DISABLE); in pd_device_terminate_all()
674 list_splice_init(&pd_chan->active_list, &list); in pd_device_terminate_all()
675 list_splice_init(&pd_chan->queue, &list); in pd_device_terminate_all()
678 pdc_chain_complete(pd_chan, desc); in pd_device_terminate_all()
680 spin_unlock_irq(&pd_chan->lock); in pd_device_terminate_all()
687 struct pch_dma_chan *pd_chan = (struct pch_dma_chan *)data; in pdc_tasklet() local
690 if (!pdc_is_idle(pd_chan)) { in pdc_tasklet()
691 dev_err(chan2dev(&pd_chan->chan), in pdc_tasklet()
696 spin_lock_irqsave(&pd_chan->lock, flags); in pdc_tasklet()
697 if (test_and_clear_bit(0, &pd_chan->err_status)) in pdc_tasklet()
698 pdc_handle_error(pd_chan); in pdc_tasklet()
700 pdc_advance_work(pd_chan); in pdc_tasklet()
701 spin_unlock_irqrestore(&pd_chan->lock, flags); in pdc_tasklet()
707 struct pch_dma_chan *pd_chan; in pd_irq() local
720 pd_chan = &pd->channels[i]; in pd_irq()
725 set_bit(0, &pd_chan->err_status); in pd_irq()
727 tasklet_schedule(&pd_chan->tasklet); in pd_irq()
733 set_bit(0, &pd_chan->err_status); in pd_irq()
735 tasklet_schedule(&pd_chan->tasklet); in pd_irq()
753 struct pch_dma_chan *pd_chan; in pch_dma_save_regs() local
763 pd_chan = to_pd_chan(chan); in pch_dma_save_regs()
765 pd->ch_regs[i].dev_addr = channel_readl(pd_chan, DEV_ADDR); in pch_dma_save_regs()
766 pd->ch_regs[i].mem_addr = channel_readl(pd_chan, MEM_ADDR); in pch_dma_save_regs()
767 pd->ch_regs[i].size = channel_readl(pd_chan, SIZE); in pch_dma_save_regs()
768 pd->ch_regs[i].next = channel_readl(pd_chan, NEXT); in pch_dma_save_regs()
776 struct pch_dma_chan *pd_chan; in pch_dma_restore_regs() local
786 pd_chan = to_pd_chan(chan); in pch_dma_restore_regs()
788 channel_writel(pd_chan, DEV_ADDR, pd->ch_regs[i].dev_addr); in pch_dma_restore_regs()
789 channel_writel(pd_chan, MEM_ADDR, pd->ch_regs[i].mem_addr); in pch_dma_restore_regs()
790 channel_writel(pd_chan, SIZE, pd->ch_regs[i].size); in pch_dma_restore_regs()
791 channel_writel(pd_chan, NEXT, pd->ch_regs[i].next); in pch_dma_restore_regs()
900 struct pch_dma_chan *pd_chan = &pd->channels[i]; in pch_dma_probe() local
902 pd_chan->chan.device = &pd->dma; in pch_dma_probe()
903 dma_cookie_init(&pd_chan->chan); in pch_dma_probe()
905 pd_chan->membase = ®s->desc[i]; in pch_dma_probe()
907 spin_lock_init(&pd_chan->lock); in pch_dma_probe()
909 INIT_LIST_HEAD(&pd_chan->active_list); in pch_dma_probe()
910 INIT_LIST_HEAD(&pd_chan->queue); in pch_dma_probe()
911 INIT_LIST_HEAD(&pd_chan->free_list); in pch_dma_probe()
913 tasklet_init(&pd_chan->tasklet, pdc_tasklet, in pch_dma_probe()
914 (unsigned long)pd_chan); in pch_dma_probe()
915 list_add_tail(&pd_chan->chan.device_node, &pd->dma.channels); in pch_dma_probe()
955 struct pch_dma_chan *pd_chan; in pch_dma_remove() local
965 pd_chan = to_pd_chan(chan); in pch_dma_remove()
967 tasklet_kill(&pd_chan->tasklet); in pch_dma_remove()