Lines Matching refs:unack
59 struct sk_buff_head unack; /* Unack'ed packets queue */ member
308 spin_lock_irqsave_nested(&bcsp->unack.lock, flags, SINGLE_DEPTH_NESTING); in bcsp_dequeue()
310 if (bcsp->unack.qlen < BCSP_TXWINSIZE) { in bcsp_dequeue()
316 __skb_queue_tail(&bcsp->unack, skb); in bcsp_dequeue()
318 spin_unlock_irqrestore(&bcsp->unack.lock, flags); in bcsp_dequeue()
327 spin_unlock_irqrestore(&bcsp->unack.lock, flags); in bcsp_dequeue()
358 spin_lock_irqsave(&bcsp->unack.lock, flags); in bcsp_pkt_cull()
360 pkts_to_be_removed = skb_queue_len(&bcsp->unack); in bcsp_pkt_cull()
374 pkts_to_be_removed, skb_queue_len(&bcsp->unack), in bcsp_pkt_cull()
378 skb_queue_walk_safe(&bcsp->unack, skb, tmp) { in bcsp_pkt_cull()
383 __skb_unlink(skb, &bcsp->unack); in bcsp_pkt_cull()
387 if (skb_queue_empty(&bcsp->unack)) in bcsp_pkt_cull()
390 spin_unlock_irqrestore(&bcsp->unack.lock, flags); in bcsp_pkt_cull()
677 BT_DBG("hu %p retransmitting %u pkts", hu, bcsp->unack.qlen); in bcsp_timed_event()
679 spin_lock_irqsave_nested(&bcsp->unack.lock, flags, SINGLE_DEPTH_NESTING); in bcsp_timed_event()
681 while ((skb = __skb_dequeue_tail(&bcsp->unack)) != NULL) { in bcsp_timed_event()
686 spin_unlock_irqrestore(&bcsp->unack.lock, flags); in bcsp_timed_event()
702 skb_queue_head_init(&bcsp->unack); in bcsp_open()
728 skb_queue_purge(&bcsp->unack); in bcsp_close()