Lines Matching refs:txq
1874 static void netif_setup_tc(struct net_device *dev, unsigned int txq) in netif_setup_tc() argument
1880 if (tc->offset + tc->count > txq) { in netif_setup_tc()
1891 if (tc->offset + tc->count > txq) { in netif_setup_tc()
2128 int netif_set_real_num_tx_queues(struct net_device *dev, unsigned int txq) in netif_set_real_num_tx_queues() argument
2132 if (txq < 1 || txq > dev->num_tx_queues) in netif_set_real_num_tx_queues()
2140 txq); in netif_set_real_num_tx_queues()
2145 netif_setup_tc(dev, txq); in netif_set_real_num_tx_queues()
2147 if (txq < dev->real_num_tx_queues) { in netif_set_real_num_tx_queues()
2148 qdisc_reset_all_tx_gt(dev, txq); in netif_set_real_num_tx_queues()
2150 netif_reset_xps_queues_gt(dev, txq); in netif_set_real_num_tx_queues()
2155 dev->real_num_tx_queues = txq; in netif_set_real_num_tx_queues()
2235 void netif_schedule_queue(struct netdev_queue *txq) in netif_schedule_queue() argument
2238 if (!(txq->state & QUEUE_STATE_ANY_XOFF)) { in netif_schedule_queue()
2239 struct Qdisc *q = rcu_dereference(txq->qdisc); in netif_schedule_queue()
2256 struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index); in netif_wake_subqueue() local
2258 if (test_and_clear_bit(__QUEUE_STATE_DRV_XOFF, &txq->state)) { in netif_wake_subqueue()
2262 q = rcu_dereference(txq->qdisc); in netif_wake_subqueue()
2645 struct netdev_queue *txq, bool more) in xmit_one() argument
2655 rc = netdev_start_xmit(skb, dev, txq, more); in xmit_one()
2662 struct netdev_queue *txq, int *ret) in dev_hard_start_xmit() argument
2671 rc = xmit_one(skb, dev, txq, next != NULL); in dev_hard_start_xmit()
2678 if (netif_xmit_stopped(txq) && skb) { in dev_hard_start_xmit()
2809 struct netdev_queue *txq) in __dev_xmit_skb() argument
2841 if (sch_direct_xmit(skb, q, dev, txq, root_lock, true)) { in __dev_xmit_skb()
2934 struct netdev_queue *txq; in __dev_queue_xmit() local
2958 txq = netdev_pick_tx(dev, skb, accel_priv); in __dev_queue_xmit()
2959 q = rcu_dereference_bh(txq->qdisc); in __dev_queue_xmit()
2966 rc = __dev_xmit_skb(skb, q, dev, txq); in __dev_queue_xmit()
2985 if (txq->xmit_lock_owner != cpu) { in __dev_queue_xmit()
2994 HARD_TX_LOCK(dev, txq, cpu); in __dev_queue_xmit()
2996 if (!netif_xmit_stopped(txq)) { in __dev_queue_xmit()
2998 skb = dev_hard_start_xmit(skb, dev, txq, &rc); in __dev_queue_xmit()
3001 HARD_TX_UNLOCK(dev, txq); in __dev_queue_xmit()
3005 HARD_TX_UNLOCK(dev, txq); in __dev_queue_xmit()