Lines Matching refs:txq

1906 static void netif_setup_tc(struct net_device *dev, unsigned int txq)  in netif_setup_tc()  argument
1912 if (tc->offset + tc->count > txq) { in netif_setup_tc()
1923 if (tc->offset + tc->count > txq) { in netif_setup_tc()
2160 int netif_set_real_num_tx_queues(struct net_device *dev, unsigned int txq) in netif_set_real_num_tx_queues() argument
2164 if (txq < 1 || txq > dev->num_tx_queues) in netif_set_real_num_tx_queues()
2172 txq); in netif_set_real_num_tx_queues()
2177 netif_setup_tc(dev, txq); in netif_set_real_num_tx_queues()
2179 if (txq < dev->real_num_tx_queues) { in netif_set_real_num_tx_queues()
2180 qdisc_reset_all_tx_gt(dev, txq); in netif_set_real_num_tx_queues()
2182 netif_reset_xps_queues_gt(dev, txq); in netif_set_real_num_tx_queues()
2187 dev->real_num_tx_queues = txq; in netif_set_real_num_tx_queues()
2267 void netif_schedule_queue(struct netdev_queue *txq) in netif_schedule_queue() argument
2270 if (!(txq->state & QUEUE_STATE_ANY_XOFF)) { in netif_schedule_queue()
2271 struct Qdisc *q = rcu_dereference(txq->qdisc); in netif_schedule_queue()
2288 struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index); in netif_wake_subqueue() local
2290 if (test_and_clear_bit(__QUEUE_STATE_DRV_XOFF, &txq->state)) { in netif_wake_subqueue()
2294 q = rcu_dereference(txq->qdisc); in netif_wake_subqueue()
2708 struct netdev_queue *txq, bool more) in xmit_one() argument
2718 rc = netdev_start_xmit(skb, dev, txq, more); in xmit_one()
2725 struct netdev_queue *txq, int *ret) in dev_hard_start_xmit() argument
2734 rc = xmit_one(skb, dev, txq, next != NULL); in dev_hard_start_xmit()
2741 if (netif_xmit_stopped(txq) && skb) { in dev_hard_start_xmit()
2872 struct netdev_queue *txq) in __dev_xmit_skb() argument
2904 if (sch_direct_xmit(skb, q, dev, txq, root_lock, true)) { in __dev_xmit_skb()
3078 struct netdev_queue *txq; in __dev_queue_xmit() local
3112 txq = netdev_pick_tx(dev, skb, accel_priv); in __dev_queue_xmit()
3113 q = rcu_dereference_bh(txq->qdisc); in __dev_queue_xmit()
3120 rc = __dev_xmit_skb(skb, q, dev, txq); in __dev_queue_xmit()
3139 if (txq->xmit_lock_owner != cpu) { in __dev_queue_xmit()
3148 HARD_TX_LOCK(dev, txq, cpu); in __dev_queue_xmit()
3150 if (!netif_xmit_stopped(txq)) { in __dev_queue_xmit()
3152 skb = dev_hard_start_xmit(skb, dev, txq, &rc); in __dev_queue_xmit()
3155 HARD_TX_UNLOCK(dev, txq); in __dev_queue_xmit()
3159 HARD_TX_UNLOCK(dev, txq); in __dev_queue_xmit()
6595 struct netdev_queue *txq = netdev_get_tx_queue(dev, i); in netif_tx_stop_all_queues() local
6596 netif_tx_stop_queue(txq); in netif_tx_stop_all_queues()