/linux-4.1.27/drivers/dma/ |
D | dma-jz4780.c | 116 struct virt_dma_desc vdesc; member 159 struct virt_dma_desc *vdesc) in to_jz4780_dma_desc() argument 161 return container_of(vdesc, struct jz4780_dma_desc, vdesc); in to_jz4780_dma_desc() 208 static void jz4780_dma_desc_free(struct virt_dma_desc *vdesc) in jz4780_dma_desc_free() argument 210 struct jz4780_dma_desc *desc = to_jz4780_dma_desc(vdesc); in jz4780_dma_desc_free() 211 struct jz4780_dma_chan *jzchan = to_jz4780_dma_chan(vdesc->tx.chan); in jz4780_dma_desc_free() 333 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); in jz4780_dma_prep_slave_sg() 383 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); in jz4780_dma_prep_dma_cyclic() 412 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); in jz4780_dma_prep_dma_memcpy() 418 struct virt_dma_desc *vdesc; in jz4780_dma_begin() local [all …]
|
D | dma-jz4740.c | 106 struct virt_dma_desc vdesc; member 146 static struct jz4740_dma_desc *to_jz4740_dma_desc(struct virt_dma_desc *vdesc) in to_jz4740_dma_desc() argument 148 return container_of(vdesc, struct jz4740_dma_desc, vdesc); in to_jz4740_dma_desc() 293 struct virt_dma_desc *vdesc; in jz4740_dma_start_transfer() local 300 vdesc = vchan_next_desc(&chan->vchan); in jz4740_dma_start_transfer() 301 if (!vdesc) in jz4740_dma_start_transfer() 303 chan->desc = to_jz4740_dma_desc(vdesc); in jz4740_dma_start_transfer() 343 vchan_cyclic_callback(&chan->desc->vdesc); in jz4740_dma_chan_irq() 346 list_del(&chan->desc->vdesc.node); in jz4740_dma_chan_irq() 347 vchan_cookie_complete(&chan->desc->vdesc); in jz4740_dma_chan_irq() [all …]
|
D | edma.c | 69 struct virt_dma_desc vdesc; member 141 return container_of(tx, struct edma_desc, vdesc.tx); in to_edma_desc() 144 static void edma_desc_free(struct virt_dma_desc *vdesc) in edma_desc_free() argument 146 kfree(container_of(vdesc, struct edma_desc, vdesc)); in edma_desc_free() 152 struct virt_dma_desc *vdesc; in edma_execute() local 161 vdesc = vchan_next_desc(&echan->vchan); in edma_execute() 162 if (!vdesc) { in edma_execute() 166 list_del(&vdesc->node); in edma_execute() 167 echan->edesc = to_edma_desc(&vdesc->tx); in edma_execute() 268 edma_desc_free(&echan->edesc->vdesc); in edma_terminate_all() [all …]
|
D | fsl-edma.c | 157 struct virt_dma_desc vdesc; member 225 return container_of(vd, struct fsl_edma_desc, vdesc); in to_fsl_edma_desc() 280 static void fsl_edma_free_desc(struct virt_dma_desc *vdesc) in fsl_edma_free_desc() argument 285 fsl_desc = to_fsl_edma_desc(vdesc); in fsl_edma_free_desc() 358 struct virt_dma_desc *vdesc, bool in_progress) in fsl_edma_desc_residue() argument 404 struct virt_dma_desc *vdesc; in fsl_edma_tx_status() local 416 vdesc = vchan_find_desc(&fsl_chan->vchan, cookie); in fsl_edma_tx_status() 417 if (fsl_chan->edesc && cookie == fsl_chan->edesc->vdesc.tx.cookie) in fsl_edma_tx_status() 418 txstate->residue = fsl_edma_desc_residue(fsl_chan, vdesc, true); in fsl_edma_tx_status() 419 else if (vdesc) in fsl_edma_tx_status() [all …]
|
D | img-mdc-dma.c | 178 struct virt_dma_desc *vdesc = container_of(t, struct virt_dma_desc, tx); in to_mdc_desc() local 180 return container_of(vdesc, struct mdc_tx_desc, vd); in to_mdc_desc()
|
/linux-4.1.27/fs/nilfs2/ |
D | ioctl.c | 575 struct nilfs_vdesc *vdesc, in nilfs_ioctl_move_inode_block() argument 581 if (vdesc->vd_flags == 0) in nilfs_ioctl_move_inode_block() 583 inode, vdesc->vd_offset, vdesc->vd_blocknr, in nilfs_ioctl_move_inode_block() 584 vdesc->vd_vblocknr, &bh); in nilfs_ioctl_move_inode_block() 587 inode, vdesc->vd_blocknr, vdesc->vd_vblocknr, &bh); in nilfs_ioctl_move_inode_block() 595 __func__, vdesc->vd_flags ? "node" : "data", in nilfs_ioctl_move_inode_block() 596 (unsigned long long)vdesc->vd_ino, in nilfs_ioctl_move_inode_block() 597 (unsigned long long)vdesc->vd_cno, in nilfs_ioctl_move_inode_block() 598 (unsigned long long)vdesc->vd_offset, in nilfs_ioctl_move_inode_block() 599 (unsigned long long)vdesc->vd_blocknr, in nilfs_ioctl_move_inode_block() [all …]
|
/linux-4.1.27/drivers/dma/hsu/ |
D | hsu.c | 121 struct virt_dma_desc *vdesc; in hsu_dma_start_transfer() local 124 vdesc = vchan_next_desc(&hsuc->vchan); in hsu_dma_start_transfer() 125 if (!vdesc) { in hsu_dma_start_transfer() 130 list_del(&vdesc->node); in hsu_dma_start_transfer() 131 hsuc->desc = to_hsu_dma_desc(vdesc); in hsu_dma_start_transfer() 186 vchan_cookie_complete(&desc->vdesc); in hsu_dma_irq() 214 static void hsu_dma_desc_free(struct virt_dma_desc *vdesc) in hsu_dma_desc_free() argument 216 struct hsu_dma_desc *desc = to_hsu_dma_desc(vdesc); in hsu_dma_desc_free() 246 return vchan_tx_prep(&hsuc->vchan, &desc->vdesc, flags); in hsu_dma_prep_slave_sg() 292 struct virt_dma_desc *vdesc; in hsu_dma_tx_status() local [all …]
|
D | hsu.h | 67 struct virt_dma_desc vdesc; member 75 static inline struct hsu_dma_desc *to_hsu_dma_desc(struct virt_dma_desc *vdesc) in to_hsu_dma_desc() argument 77 return container_of(vdesc, struct hsu_dma_desc, vdesc); in to_hsu_dma_desc()
|