Lines Matching refs:chan
96 struct dma_chan chan; member
140 container_of(dchan, struct mmp_pdma_chan, chan)
242 struct mmp_pdma_device *pdev = to_mmp_pdma_dev(pchan->chan.device); in lookup_phy()
275 struct mmp_pdma_device *pdev = to_mmp_pdma_dev(pchan->chan.device); in mmp_pdma_free_phy()
296 static void start_pending_queue(struct mmp_pdma_chan *chan) in start_pending_queue() argument
301 if (!chan->idle) { in start_pending_queue()
302 dev_dbg(chan->dev, "DMA controller still busy\n"); in start_pending_queue()
306 if (list_empty(&chan->chain_pending)) { in start_pending_queue()
308 mmp_pdma_free_phy(chan); in start_pending_queue()
309 dev_dbg(chan->dev, "no pending list\n"); in start_pending_queue()
313 if (!chan->phy) { in start_pending_queue()
314 chan->phy = lookup_phy(chan); in start_pending_queue()
315 if (!chan->phy) { in start_pending_queue()
316 dev_dbg(chan->dev, "no free dma channel\n"); in start_pending_queue()
325 desc = list_first_entry(&chan->chain_pending, in start_pending_queue()
327 list_splice_tail_init(&chan->chain_pending, &chan->chain_running); in start_pending_queue()
333 set_desc(chan->phy, desc->async_tx.phys); in start_pending_queue()
334 enable_chan(chan->phy); in start_pending_queue()
335 chan->idle = false; in start_pending_queue()
342 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(tx->chan); in mmp_pdma_tx_submit() local
348 spin_lock_irqsave(&chan->desc_lock, flags); in mmp_pdma_tx_submit()
355 list_splice_tail_init(&desc->tx_list, &chan->chain_pending); in mmp_pdma_tx_submit()
357 spin_unlock_irqrestore(&chan->desc_lock, flags); in mmp_pdma_tx_submit()
363 mmp_pdma_alloc_descriptor(struct mmp_pdma_chan *chan) in mmp_pdma_alloc_descriptor() argument
368 desc = dma_pool_alloc(chan->desc_pool, GFP_ATOMIC, &pdesc); in mmp_pdma_alloc_descriptor()
370 dev_err(chan->dev, "out of memory for link descriptor\n"); in mmp_pdma_alloc_descriptor()
376 dma_async_tx_descriptor_init(&desc->async_tx, &chan->chan); in mmp_pdma_alloc_descriptor()
394 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan); in mmp_pdma_alloc_chan_resources() local
396 if (chan->desc_pool) in mmp_pdma_alloc_chan_resources()
399 chan->desc_pool = dma_pool_create(dev_name(&dchan->dev->device), in mmp_pdma_alloc_chan_resources()
400 chan->dev, in mmp_pdma_alloc_chan_resources()
404 if (!chan->desc_pool) { in mmp_pdma_alloc_chan_resources()
405 dev_err(chan->dev, "unable to allocate descriptor pool\n"); in mmp_pdma_alloc_chan_resources()
409 mmp_pdma_free_phy(chan); in mmp_pdma_alloc_chan_resources()
410 chan->idle = true; in mmp_pdma_alloc_chan_resources()
411 chan->dev_addr = 0; in mmp_pdma_alloc_chan_resources()
415 static void mmp_pdma_free_desc_list(struct mmp_pdma_chan *chan, in mmp_pdma_free_desc_list() argument
422 dma_pool_free(chan->desc_pool, desc, desc->async_tx.phys); in mmp_pdma_free_desc_list()
428 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan); in mmp_pdma_free_chan_resources() local
431 spin_lock_irqsave(&chan->desc_lock, flags); in mmp_pdma_free_chan_resources()
432 mmp_pdma_free_desc_list(chan, &chan->chain_pending); in mmp_pdma_free_chan_resources()
433 mmp_pdma_free_desc_list(chan, &chan->chain_running); in mmp_pdma_free_chan_resources()
434 spin_unlock_irqrestore(&chan->desc_lock, flags); in mmp_pdma_free_chan_resources()
436 dma_pool_destroy(chan->desc_pool); in mmp_pdma_free_chan_resources()
437 chan->desc_pool = NULL; in mmp_pdma_free_chan_resources()
438 chan->idle = true; in mmp_pdma_free_chan_resources()
439 chan->dev_addr = 0; in mmp_pdma_free_chan_resources()
440 mmp_pdma_free_phy(chan); in mmp_pdma_free_chan_resources()
449 struct mmp_pdma_chan *chan; in mmp_pdma_prep_memcpy() local
459 chan = to_mmp_pdma_chan(dchan); in mmp_pdma_prep_memcpy()
460 chan->byte_align = false; in mmp_pdma_prep_memcpy()
462 if (!chan->dir) { in mmp_pdma_prep_memcpy()
463 chan->dir = DMA_MEM_TO_MEM; in mmp_pdma_prep_memcpy()
464 chan->dcmd = DCMD_INCTRGADDR | DCMD_INCSRCADDR; in mmp_pdma_prep_memcpy()
465 chan->dcmd |= DCMD_BURST32; in mmp_pdma_prep_memcpy()
470 new = mmp_pdma_alloc_descriptor(chan); in mmp_pdma_prep_memcpy()
472 dev_err(chan->dev, "no memory for desc\n"); in mmp_pdma_prep_memcpy()
478 chan->byte_align = true; in mmp_pdma_prep_memcpy()
480 new->desc.dcmd = chan->dcmd | (DCMD_LENGTH & copy); in mmp_pdma_prep_memcpy()
495 if (chan->dir == DMA_MEM_TO_DEV) { in mmp_pdma_prep_memcpy()
497 } else if (chan->dir == DMA_DEV_TO_MEM) { in mmp_pdma_prep_memcpy()
499 } else if (chan->dir == DMA_MEM_TO_MEM) { in mmp_pdma_prep_memcpy()
515 chan->cyclic_first = NULL; in mmp_pdma_prep_memcpy()
521 mmp_pdma_free_desc_list(chan, &first->tx_list); in mmp_pdma_prep_memcpy()
530 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan); in mmp_pdma_prep_slave_sg() local
540 chan->byte_align = false; in mmp_pdma_prep_slave_sg()
549 chan->byte_align = true; in mmp_pdma_prep_slave_sg()
552 new = mmp_pdma_alloc_descriptor(chan); in mmp_pdma_prep_slave_sg()
554 dev_err(chan->dev, "no memory for desc\n"); in mmp_pdma_prep_slave_sg()
558 new->desc.dcmd = chan->dcmd | (DCMD_LENGTH & len); in mmp_pdma_prep_slave_sg()
561 new->desc.dtadr = chan->dev_addr; in mmp_pdma_prep_slave_sg()
563 new->desc.dsadr = chan->dev_addr; in mmp_pdma_prep_slave_sg()
592 chan->dir = dir; in mmp_pdma_prep_slave_sg()
593 chan->cyclic_first = NULL; in mmp_pdma_prep_slave_sg()
599 mmp_pdma_free_desc_list(chan, &first->tx_list); in mmp_pdma_prep_slave_sg()
609 struct mmp_pdma_chan *chan; in mmp_pdma_prep_dma_cyclic() local
623 chan = to_mmp_pdma_chan(dchan); in mmp_pdma_prep_dma_cyclic()
628 dma_dst = chan->dev_addr; in mmp_pdma_prep_dma_cyclic()
632 dma_src = chan->dev_addr; in mmp_pdma_prep_dma_cyclic()
635 dev_err(chan->dev, "Unsupported direction for cyclic DMA\n"); in mmp_pdma_prep_dma_cyclic()
639 chan->dir = direction; in mmp_pdma_prep_dma_cyclic()
643 new = mmp_pdma_alloc_descriptor(chan); in mmp_pdma_prep_dma_cyclic()
645 dev_err(chan->dev, "no memory for desc\n"); in mmp_pdma_prep_dma_cyclic()
649 new->desc.dcmd = (chan->dcmd | DCMD_ENDIRQEN | in mmp_pdma_prep_dma_cyclic()
665 if (chan->dir == DMA_MEM_TO_DEV) in mmp_pdma_prep_dma_cyclic()
679 chan->cyclic_first = first; in mmp_pdma_prep_dma_cyclic()
685 mmp_pdma_free_desc_list(chan, &first->tx_list); in mmp_pdma_prep_dma_cyclic()
692 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan); in mmp_pdma_config() local
700 chan->dcmd = DCMD_INCTRGADDR | DCMD_FLOWSRC; in mmp_pdma_config()
705 chan->dcmd = DCMD_INCSRCADDR | DCMD_FLOWTRG; in mmp_pdma_config()
712 chan->dcmd |= DCMD_WIDTH1; in mmp_pdma_config()
714 chan->dcmd |= DCMD_WIDTH2; in mmp_pdma_config()
716 chan->dcmd |= DCMD_WIDTH4; in mmp_pdma_config()
719 chan->dcmd |= DCMD_BURST8; in mmp_pdma_config()
721 chan->dcmd |= DCMD_BURST16; in mmp_pdma_config()
723 chan->dcmd |= DCMD_BURST32; in mmp_pdma_config()
725 chan->dir = cfg->direction; in mmp_pdma_config()
726 chan->dev_addr = addr; in mmp_pdma_config()
732 chan->drcmr = cfg->slave_id; in mmp_pdma_config()
739 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan); in mmp_pdma_terminate_all() local
745 disable_chan(chan->phy); in mmp_pdma_terminate_all()
746 mmp_pdma_free_phy(chan); in mmp_pdma_terminate_all()
747 spin_lock_irqsave(&chan->desc_lock, flags); in mmp_pdma_terminate_all()
748 mmp_pdma_free_desc_list(chan, &chan->chain_pending); in mmp_pdma_terminate_all()
749 mmp_pdma_free_desc_list(chan, &chan->chain_running); in mmp_pdma_terminate_all()
750 spin_unlock_irqrestore(&chan->desc_lock, flags); in mmp_pdma_terminate_all()
751 chan->idle = true; in mmp_pdma_terminate_all()
756 static unsigned int mmp_pdma_residue(struct mmp_pdma_chan *chan, in mmp_pdma_residue() argument
762 bool cyclic = chan->cyclic_first != NULL; in mmp_pdma_residue()
768 if (!chan->phy) in mmp_pdma_residue()
771 if (chan->dir == DMA_DEV_TO_MEM) in mmp_pdma_residue()
772 curr = readl(chan->phy->base + DTADR(chan->phy->idx)); in mmp_pdma_residue()
774 curr = readl(chan->phy->base + DSADR(chan->phy->idx)); in mmp_pdma_residue()
776 list_for_each_entry(sw, &chan->chain_running, node) { in mmp_pdma_residue()
779 if (chan->dir == DMA_DEV_TO_MEM) in mmp_pdma_residue()
834 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan); in mmp_pdma_tx_status() local
839 dma_set_residue(txstate, mmp_pdma_residue(chan, cookie)); in mmp_pdma_tx_status()
850 struct mmp_pdma_chan *chan = to_mmp_pdma_chan(dchan); in mmp_pdma_issue_pending() local
853 spin_lock_irqsave(&chan->desc_lock, flags); in mmp_pdma_issue_pending()
854 start_pending_queue(chan); in mmp_pdma_issue_pending()
855 spin_unlock_irqrestore(&chan->desc_lock, flags); in mmp_pdma_issue_pending()
865 struct mmp_pdma_chan *chan = (struct mmp_pdma_chan *)data; in dma_do_tasklet() local
870 if (chan->cyclic_first) { in dma_do_tasklet()
874 spin_lock_irqsave(&chan->desc_lock, flags); in dma_do_tasklet()
875 desc = chan->cyclic_first; in dma_do_tasklet()
878 spin_unlock_irqrestore(&chan->desc_lock, flags); in dma_do_tasklet()
887 spin_lock_irqsave(&chan->desc_lock, flags); in dma_do_tasklet()
889 list_for_each_entry_safe(desc, _desc, &chan->chain_running, node) { in dma_do_tasklet()
904 dev_dbg(chan->dev, "completed_cookie=%d\n", cookie); in dma_do_tasklet()
913 chan->idle = list_empty(&chan->chain_running); in dma_do_tasklet()
916 start_pending_queue(chan); in dma_do_tasklet()
917 spin_unlock_irqrestore(&chan->desc_lock, flags); in dma_do_tasklet()
929 dma_pool_free(chan->desc_pool, desc, txd->phys); in dma_do_tasklet()
944 struct mmp_pdma_chan *chan; in mmp_pdma_chan_init() local
947 chan = devm_kzalloc(pdev->dev, sizeof(*chan), GFP_KERNEL); in mmp_pdma_chan_init()
948 if (chan == NULL) in mmp_pdma_chan_init()
963 spin_lock_init(&chan->desc_lock); in mmp_pdma_chan_init()
964 chan->dev = pdev->dev; in mmp_pdma_chan_init()
965 chan->chan.device = &pdev->device; in mmp_pdma_chan_init()
966 tasklet_init(&chan->tasklet, dma_do_tasklet, (unsigned long)chan); in mmp_pdma_chan_init()
967 INIT_LIST_HEAD(&chan->chain_pending); in mmp_pdma_chan_init()
968 INIT_LIST_HEAD(&chan->chain_running); in mmp_pdma_chan_init()
971 list_add_tail(&chan->chan.device_node, &pdev->device.channels); in mmp_pdma_chan_init()
986 struct dma_chan *chan; in mmp_pdma_dma_xlate() local
988 chan = dma_get_any_slave_channel(&d->device); in mmp_pdma_dma_xlate()
989 if (!chan) in mmp_pdma_dma_xlate()
992 to_mmp_pdma_chan(chan)->drcmr = dma_spec->args[0]; in mmp_pdma_dma_xlate()
994 return chan; in mmp_pdma_dma_xlate()
1121 bool mmp_pdma_filter_fn(struct dma_chan *chan, void *param) in mmp_pdma_filter_fn() argument
1123 struct mmp_pdma_chan *c = to_mmp_pdma_chan(chan); in mmp_pdma_filter_fn()
1125 if (chan->device->dev->driver != &mmp_pdma_driver.driver) in mmp_pdma_filter_fn()