Lines Matching refs:cur_p

583 	struct axidma_bd *cur_p;  in axienet_start_xmit_done()  local
586 cur_p = &lp->tx_bd_v[lp->tx_bd_ci]; in axienet_start_xmit_done()
587 status = cur_p->status; in axienet_start_xmit_done()
589 dma_unmap_single(ndev->dev.parent, cur_p->phys, in axienet_start_xmit_done()
590 (cur_p->cntrl & XAXIDMA_BD_CTRL_LENGTH_MASK), in axienet_start_xmit_done()
592 if (cur_p->app4) in axienet_start_xmit_done()
593 dev_kfree_skb_irq((struct sk_buff *)cur_p->app4); in axienet_start_xmit_done()
595 cur_p->app0 = 0; in axienet_start_xmit_done()
596 cur_p->app1 = 0; in axienet_start_xmit_done()
597 cur_p->app2 = 0; in axienet_start_xmit_done()
598 cur_p->app4 = 0; in axienet_start_xmit_done()
599 cur_p->status = 0; in axienet_start_xmit_done()
606 cur_p = &lp->tx_bd_v[lp->tx_bd_ci]; in axienet_start_xmit_done()
607 status = cur_p->status; in axienet_start_xmit_done()
631 struct axidma_bd *cur_p; in axienet_check_tx_bd_space() local
632 cur_p = &lp->tx_bd_v[(lp->tx_bd_tail + num_frag) % TX_BD_NUM]; in axienet_check_tx_bd_space()
633 if (cur_p->status & XAXIDMA_BD_STS_ALL_MASK) in axienet_check_tx_bd_space()
660 struct axidma_bd *cur_p; in axienet_start_xmit() local
663 cur_p = &lp->tx_bd_v[lp->tx_bd_tail]; in axienet_start_xmit()
674 cur_p->app0 |= 2; in axienet_start_xmit()
679 cur_p->app0 |= 1; in axienet_start_xmit()
680 cur_p->app1 = (csum_start_off << 16) | csum_index_off; in axienet_start_xmit()
683 cur_p->app0 |= 2; /* Tx Full Checksum Offload Enabled */ in axienet_start_xmit()
686 cur_p->cntrl = skb_headlen(skb) | XAXIDMA_BD_CTRL_TXSOF_MASK; in axienet_start_xmit()
687 cur_p->phys = dma_map_single(ndev->dev.parent, skb->data, in axienet_start_xmit()
693 cur_p = &lp->tx_bd_v[lp->tx_bd_tail]; in axienet_start_xmit()
695 cur_p->phys = dma_map_single(ndev->dev.parent, in axienet_start_xmit()
699 cur_p->cntrl = skb_frag_size(frag); in axienet_start_xmit()
702 cur_p->cntrl |= XAXIDMA_BD_CTRL_TXEOF_MASK; in axienet_start_xmit()
703 cur_p->app4 = (unsigned long)skb; in axienet_start_xmit()
732 struct axidma_bd *cur_p; in axienet_recv() local
735 cur_p = &lp->rx_bd_v[lp->rx_bd_ci]; in axienet_recv()
737 while ((cur_p->status & XAXIDMA_BD_STS_COMPLETE_MASK)) { in axienet_recv()
738 skb = (struct sk_buff *) (cur_p->sw_id_offset); in axienet_recv()
739 length = cur_p->app4 & 0x0000FFFF; in axienet_recv()
741 dma_unmap_single(ndev->dev.parent, cur_p->phys, in axienet_recv()
752 csumstatus = (cur_p->app2 & in axienet_recv()
761 skb->csum = be32_to_cpu(cur_p->app3 & 0xFFFF); in axienet_recv()
774 cur_p->phys = dma_map_single(ndev->dev.parent, new_skb->data, in axienet_recv()
777 cur_p->cntrl = lp->max_frm_size; in axienet_recv()
778 cur_p->status = 0; in axienet_recv()
779 cur_p->sw_id_offset = (u32) new_skb; in axienet_recv()
783 cur_p = &lp->rx_bd_v[lp->rx_bd_ci]; in axienet_recv()
1348 struct axidma_bd *cur_p; in axienet_dma_err_handler() local
1368 cur_p = &lp->tx_bd_v[i]; in axienet_dma_err_handler()
1369 if (cur_p->phys) in axienet_dma_err_handler()
1370 dma_unmap_single(ndev->dev.parent, cur_p->phys, in axienet_dma_err_handler()
1371 (cur_p->cntrl & in axienet_dma_err_handler()
1374 if (cur_p->app4) in axienet_dma_err_handler()
1375 dev_kfree_skb_irq((struct sk_buff *) cur_p->app4); in axienet_dma_err_handler()
1376 cur_p->phys = 0; in axienet_dma_err_handler()
1377 cur_p->cntrl = 0; in axienet_dma_err_handler()
1378 cur_p->status = 0; in axienet_dma_err_handler()
1379 cur_p->app0 = 0; in axienet_dma_err_handler()
1380 cur_p->app1 = 0; in axienet_dma_err_handler()
1381 cur_p->app2 = 0; in axienet_dma_err_handler()
1382 cur_p->app3 = 0; in axienet_dma_err_handler()
1383 cur_p->app4 = 0; in axienet_dma_err_handler()
1384 cur_p->sw_id_offset = 0; in axienet_dma_err_handler()
1388 cur_p = &lp->rx_bd_v[i]; in axienet_dma_err_handler()
1389 cur_p->status = 0; in axienet_dma_err_handler()
1390 cur_p->app0 = 0; in axienet_dma_err_handler()
1391 cur_p->app1 = 0; in axienet_dma_err_handler()
1392 cur_p->app2 = 0; in axienet_dma_err_handler()
1393 cur_p->app3 = 0; in axienet_dma_err_handler()
1394 cur_p->app4 = 0; in axienet_dma_err_handler()