queued            175 block/bfq-cgroup.c 	if (blkg_rwstat_total(&stats->queued))
queued            217 block/bfq-cgroup.c 		      blkg_rwstat_total(&stats->queued));
queued            225 block/bfq-cgroup.c 	blkg_rwstat_add(&bfqg->stats.queued, op, 1);
queued            233 block/bfq-cgroup.c 	blkg_rwstat_add(&bfqg->stats.queued, op, -1);
queued            438 block/bfq-cgroup.c 	blkg_rwstat_exit(&stats->queued);
queued            455 block/bfq-cgroup.c 	    blkg_rwstat_init(&stats->queued, gfp) ||
queued           1290 block/bfq-cgroup.c 		.private = offsetof(struct bfq_group, stats.queued),
queued           1333 block/bfq-cgroup.c 		.private = offsetof(struct bfq_group, stats.queued),
queued            423 block/bfq-iosched.c 	if (bfqd->queued != 0) {
queued           1873 block/bfq-iosched.c 	bfqq->queued[rq_is_sync(rq)]++;
queued           1874 block/bfq-iosched.c 	bfqd->queued++;
queued           2166 block/bfq-iosched.c 	bfqq->queued[sync]--;
queued           2167 block/bfq-iosched.c 	bfqd->queued--;
queued           5356 block/bfq-iosched.c 		bool small_req = bfqq->queued[rq_is_sync(rq)] == 1 &&
queued           5571 block/bfq-iosched.c 	if (bfqd->rq_in_driver + bfqd->queued <= BFQ_HW_QUEUE_THRESHOLD)
queued           5580 block/bfq-iosched.c 	    bfqq->dispatched + bfqq->queued[0] + bfqq->queued[1] <
queued           6240 block/bfq-iosched.c 	else if (bfqq->queued[0] == 0 && bfqq->queued[1] == 0)
queued            261 block/bfq-iosched.h 	int queued[2];
queued            519 block/bfq-iosched.h 	int queued;
queued            820 block/bfq-iosched.h 	struct blkg_rwstat		queued;
queued            579 block/blk-mq-debugfs.c 	seq_printf(m, "%lu\n", hctx->queued);
queued            588 block/blk-mq-debugfs.c 	hctx->queued = 0;
queued            418 block/blk-mq.c 	data->hctx->queued++;
queued           1048 block/blk-mq.c static inline unsigned int queued_to_index(unsigned int queued)
queued           1050 block/blk-mq.c 	if (!queued)
queued           1053 block/blk-mq.c 	return min(BLK_MQ_MAX_DISPATCH_ORDER - 1, ilog2(queued) + 1);
queued           1217 block/blk-mq.c 	int errors, queued;
queued           1228 block/blk-mq.c 	errors = queued = 0;
queued           1297 block/blk-mq.c 		queued++;
queued           1300 block/blk-mq.c 	hctx->dispatched[queued_to_index(queued)]++;
queued           1364 block/blk-mq.c 	return (queued + errors) != 0;
queued             80 block/blk-throttle.c 	struct list_head	queued[2];	/* throtl_qnode [READ/WRITE] */
queued            407 block/blk-throttle.c 				 struct list_head *queued)
queued            411 block/blk-throttle.c 		list_add_tail(&qn->node, queued);
queued            420 block/blk-throttle.c static struct bio *throtl_peek_queued(struct list_head *queued)
queued            422 block/blk-throttle.c 	struct throtl_qnode *qn = list_first_entry(queued, struct throtl_qnode, node);
queued            425 block/blk-throttle.c 	if (list_empty(queued))
queued            447 block/blk-throttle.c static struct bio *throtl_pop_queued(struct list_head *queued,
queued            450 block/blk-throttle.c 	struct throtl_qnode *qn = list_first_entry(queued, struct throtl_qnode, node);
queued            453 block/blk-throttle.c 	if (list_empty(queued))
queued            466 block/blk-throttle.c 		list_move_tail(&qn->node, queued);
queued            475 block/blk-throttle.c 	INIT_LIST_HEAD(&sq->queued[0]);
queued            476 block/blk-throttle.c 	INIT_LIST_HEAD(&sq->queued[1]);
queued            980 block/blk-throttle.c 	       bio != throtl_peek_queued(&tg->service_queue.queued[rw]));
queued           1072 block/blk-throttle.c 	throtl_qnode_add_bio(bio, qn, &sq->queued[rw]);
queued           1084 block/blk-throttle.c 	bio = throtl_peek_queued(&sq->queued[READ]);
queued           1088 block/blk-throttle.c 	bio = throtl_peek_queued(&sq->queued[WRITE]);
queued           1128 block/blk-throttle.c 	bio = throtl_pop_queued(&sq->queued[rw], &tg_to_put);
queued           1145 block/blk-throttle.c 				     &parent_sq->queued[rw]);
queued           1166 block/blk-throttle.c 	while ((bio = throtl_peek_queued(&sq->queued[READ])) &&
queued           1176 block/blk-throttle.c 	while ((bio = throtl_peek_queued(&sq->queued[WRITE])) &&
queued           1318 block/blk-throttle.c 		while ((bio = throtl_pop_queued(&td_sq->queued[rw], NULL)))
queued           2323 block/blk-throttle.c 		while ((bio = throtl_peek_queued(&sq->queued[READ])))
queued           2325 block/blk-throttle.c 		while ((bio = throtl_peek_queued(&sq->queued[WRITE])))
queued           2364 block/blk-throttle.c 		while ((bio = throtl_pop_queued(&td->service_queue.queued[rw],
queued            153 drivers/atm/eni.c static int tx_complete = 0,dma_complete = 0,queued = 0,requeued = 0,
queued           1179 drivers/atm/eni.c 	queued++;
queued           1466 drivers/atm/eni.c   tx_complete,dma_complete,queued,requeued,submitted,backlogged,
queued            635 drivers/block/drbd/drbd_worker.c 			int queued = sk->sk_wmem_queued;
queued            637 drivers/block/drbd/drbd_worker.c 			if (queued > sndbuf / 2) {
queued            567 drivers/crypto/inside-secure/safexcel_cipher.c 	int queued, i, ret = 0;
queued            608 drivers/crypto/inside-secure/safexcel_cipher.c 	queued = totlen_src;
queued            649 drivers/crypto/inside-secure/safexcel_cipher.c 		if (queued - len < 0)
queued            650 drivers/crypto/inside-secure/safexcel_cipher.c 			len = queued;
queued            653 drivers/crypto/inside-secure/safexcel_cipher.c 					   !(queued - len),
queued            667 drivers/crypto/inside-secure/safexcel_cipher.c 		queued -= len;
queued            668 drivers/crypto/inside-secure/safexcel_cipher.c 		if (!queued)
queued            265 drivers/crypto/inside-secure/safexcel_hash.c 	u64 queued, len, cache_len;
queued            267 drivers/crypto/inside-secure/safexcel_hash.c 	queued = len = safexcel_queued_len(req);
queued            268 drivers/crypto/inside-secure/safexcel_hash.c 	if (queued <= HASH_CACHE_SIZE)
queued            269 drivers/crypto/inside-secure/safexcel_hash.c 		cache_len = queued;
queued            271 drivers/crypto/inside-secure/safexcel_hash.c 		cache_len = queued - areq->nbytes;
queued            277 drivers/crypto/inside-secure/safexcel_hash.c 		extra = queued & (HASH_CACHE_SIZE - 1);
queued            289 drivers/crypto/inside-secure/safexcel_hash.c 		queued -= extra;
queued            292 drivers/crypto/inside-secure/safexcel_hash.c 		if (!queued) {
queued            317 drivers/crypto/inside-secure/safexcel_hash.c 		queued -= cache_len;
queued            318 drivers/crypto/inside-secure/safexcel_hash.c 		if (!queued)
queued            340 drivers/crypto/inside-secure/safexcel_hash.c 		if (queued < sglen)
queued            341 drivers/crypto/inside-secure/safexcel_hash.c 			sglen = queued;
queued            344 drivers/crypto/inside-secure/safexcel_hash.c 					   !(queued - sglen),
queued            356 drivers/crypto/inside-secure/safexcel_hash.c 		queued -= sglen;
queued            357 drivers/crypto/inside-secure/safexcel_hash.c 		if (!queued)
queued            197 drivers/dma/mpc512x_dma.c 	struct list_head		queued;
queued            262 drivers/dma/mpc512x_dma.c 	while (!list_empty(&mchan->queued)) {
queued            263 drivers/dma/mpc512x_dma.c 		mdesc = list_first_entry(&mchan->queued,
queued            341 drivers/dma/mpc512x_dma.c 		if (!list_empty(&mchan->queued))
queued            472 drivers/dma/mpc512x_dma.c 	list_move_tail(&mdesc->node, &mchan->queued);
queued            558 drivers/dma/mpc512x_dma.c 	BUG_ON(!list_empty(&mchan->queued));
queued            889 drivers/dma/mpc512x_dma.c 	list_splice_tail_init(&mchan->queued, &mchan->free);
queued           1004 drivers/dma/mpc512x_dma.c 		INIT_LIST_HEAD(&mchan->queued);
queued            214 drivers/dma/nbpfaxi.c 	struct list_head queued;
queued            603 drivers/dma/nbpfaxi.c 	if (list_empty(&chan->queued))
queued            606 drivers/dma/nbpfaxi.c 	list_splice_tail_init(&chan->queued, &chan->active);
queued            647 drivers/dma/nbpfaxi.c 				list_for_each_entry(desc, &chan->queued, node)
queued            675 drivers/dma/nbpfaxi.c 	list_add_tail(&desc->node, &chan->queued);
queued            843 drivers/dma/nbpfaxi.c 	list_splice_init(&chan->queued, &head);
queued           1040 drivers/dma/nbpfaxi.c 	INIT_LIST_HEAD(&chan->queued);
queued            178 drivers/dma/qcom/hidma.c 	bool queued = false;
queued            184 drivers/dma/qcom/hidma.c 		queued = true;
queued            194 drivers/dma/qcom/hidma.c 	if (queued) {
queued            219 drivers/dma/qcom/hidma.c 	INIT_LIST_HEAD(&mchan->queued);
queued            244 drivers/dma/qcom/hidma.c 	list_for_each_entry_safe(qdesc, next, &mchan->queued, node) {
queued            338 drivers/dma/qcom/hidma.c 	list_move_tail(&mdesc->node, &mchan->queued);
queued            483 drivers/dma/qcom/hidma.c 	list_splice_init(&mchan->queued, &list);
queued             30 drivers/dma/qcom/hidma.h 	bool queued;			/* flag whether this is pending     */
queued            102 drivers/dma/qcom/hidma.h 	struct list_head		queued;
queued             31 drivers/dma/qcom/hidma_dbg.c 	seq_printf(s, "queued = 0x%x\n", tre->queued);
queued            160 drivers/dma/qcom/hidma_ll.c 	tre->queued = 0;
queued            223 drivers/dma/qcom/hidma_ll.c 	tre->queued = 0;
queued            539 drivers/dma/qcom/hidma_ll.c 	tre->queued = 1;
queued            101 drivers/dma/sirf-dma.c 	struct list_head		queued;
queued            248 drivers/dma/sirf-dma.c 	sdesc = list_first_entry(&schan->queued, struct sirfsoc_dma_desc,
queued            291 drivers/dma/sirf-dma.c 				if (!list_empty(&schan->queued))
queued            317 drivers/dma/sirf-dma.c 				if (!list_empty(&schan->queued))
queued            416 drivers/dma/sirf-dma.c 	list_move_tail(&sdesc->node, &schan->queued);
queued            481 drivers/dma/sirf-dma.c 	list_splice_tail_init(&schan->queued, &schan->free);
queued            608 drivers/dma/sirf-dma.c 	BUG_ON(!list_empty(&schan->queued));
queued            632 drivers/dma/sirf-dma.c 	if (list_empty(&schan->active) && !list_empty(&schan->queued))
queued            933 drivers/dma/sirf-dma.c 		INIT_LIST_HEAD(&schan->queued);
queued             63 drivers/gpu/drm/drm_flip_work.c 	list_add_tail(&task->node, &work->queued);
queued            107 drivers/gpu/drm/drm_flip_work.c 	list_splice_tail(&work->queued, &work->commited);
queued            108 drivers/gpu/drm/drm_flip_work.c 	INIT_LIST_HEAD(&work->queued);
queued            151 drivers/gpu/drm/drm_flip_work.c 	INIT_LIST_HEAD(&work->queued);
queued            168 drivers/gpu/drm/drm_flip_work.c 	WARN_ON(!list_empty(&work->queued) || !list_empty(&work->commited));
queued            746 drivers/infiniband/hw/hfi1/user_sdma.c 	u32 queued = *queued_ptr, data_sent = *data_sent_ptr;
queued            755 drivers/infiniband/hw/hfi1/user_sdma.c 	len = min((datalen - queued), len);
queued            763 drivers/infiniband/hw/hfi1/user_sdma.c 	queued += len;
queued            765 drivers/infiniband/hw/hfi1/user_sdma.c 	if (unlikely(queued < datalen && pageidx == iovec->npages &&
queued            772 drivers/infiniband/hw/hfi1/user_sdma.c 	*queued_ptr = queued;
queued            809 drivers/infiniband/hw/hfi1/user_sdma.c 		u32 datalen = 0, queued = 0, data_sent = 0;
queued            907 drivers/infiniband/hw/hfi1/user_sdma.c 		while (queued < datalen &&
queued            910 drivers/infiniband/hw/hfi1/user_sdma.c 					      &queued, &data_sent, &iov_offset);
queued             26 drivers/md/dm-cache-background-tracker.c 	struct list_head queued;
queued             47 drivers/md/dm-cache-background-tracker.c 	INIT_LIST_HEAD(&b->queued);
queued            205 drivers/md/dm-cache-background-tracker.c 		list_add(&w->list, &b->queued);
queued            219 drivers/md/dm-cache-background-tracker.c 	if (list_empty(&b->queued))
queued            222 drivers/md/dm-cache-background-tracker.c 	w = list_first_entry(&b->queued, struct bt_work, list);
queued           2407 drivers/media/common/videobuf2/videobuf2-core.c 	unsigned int queued:1;
queued           2537 drivers/media/common/videobuf2/videobuf2-core.c 			fileio->bufs[i].queued = 1;
queued           2657 drivers/media/common/videobuf2/videobuf2-core.c 		buf->queued = 0;
queued           2728 drivers/media/common/videobuf2/videobuf2-core.c 		buf->queued = 1;
queued             89 drivers/media/platform/fsl-viu.c 	struct list_head	queued;
queued            314 drivers/media/platform/fsl-viu.c 		if (list_empty(&vidq->queued))
queued            316 drivers/media/platform/fsl-viu.c 		buf = list_entry(vidq->queued.next, struct viu_buf, vb.queue);
queued            500 drivers/media/platform/fsl-viu.c 	if (!list_empty(&vidq->queued)) {
queued            503 drivers/media/platform/fsl-viu.c 				vidq, &vidq->queued);
queued            505 drivers/media/platform/fsl-viu.c 			dev, &vidq->queued, vidq->queued.next,
queued            506 drivers/media/platform/fsl-viu.c 			vidq->queued.prev);
queued            507 drivers/media/platform/fsl-viu.c 		list_add_tail(&buf->vb.queue, &vidq->queued);
queued            531 drivers/media/platform/fsl-viu.c 			list_add_tail(&buf->vb.queue, &vidq->queued);
queued            967 drivers/media/platform/fsl-viu.c 	} else if (!list_empty(&vidq->queued)) {
queued            968 drivers/media/platform/fsl-viu.c 		buf = list_entry(vidq->queued.next, struct viu_buf,
queued           1204 drivers/media/platform/fsl-viu.c 		list_empty(&dev->vidq.queued));
queued           1434 drivers/media/platform/fsl-viu.c 	INIT_LIST_HEAD(&viu_dev->vidq.queued);
queued            224 drivers/media/platform/vsp1/vsp1_dl.c 	struct vsp1_dl_list *queued;
queued            839 drivers/media/platform/vsp1/vsp1_dl.c 	if (!dlm->queued)
queued            897 drivers/media/platform/vsp1/vsp1_dl.c 	__vsp1_dl_list_put(dlm->queued);
queued            898 drivers/media/platform/vsp1/vsp1_dl.c 	dlm->queued = dl;
queued           1018 drivers/media/platform/vsp1/vsp1_dl.c 	if (dlm->queued) {
queued           1019 drivers/media/platform/vsp1/vsp1_dl.c 		if (dlm->queued->flags & VSP1_DL_FRAME_END_INTERNAL)
queued           1021 drivers/media/platform/vsp1/vsp1_dl.c 		dlm->queued->flags &= ~VSP1_DL_FRAME_END_INTERNAL;
queued           1024 drivers/media/platform/vsp1/vsp1_dl.c 		dlm->active = dlm->queued;
queued           1025 drivers/media/platform/vsp1/vsp1_dl.c 		dlm->queued = NULL;
queued           1036 drivers/media/platform/vsp1/vsp1_dl.c 		dlm->queued = dlm->pending;
queued           1072 drivers/media/platform/vsp1/vsp1_dl.c 	__vsp1_dl_list_put(dlm->queued);
queued           1078 drivers/media/platform/vsp1/vsp1_dl.c 	dlm->queued = NULL;
queued           1484 drivers/media/usb/cx231xx/cx231xx-cards.c 	INIT_LIST_HEAD(&dev->video_mode.vidq.queued);
queued           1488 drivers/media/usb/cx231xx/cx231xx-cards.c 	INIT_LIST_HEAD(&dev->vbi_mode.vidq.queued);
queued            240 drivers/media/usb/cx231xx/cx231xx.h 	struct list_head queued;
queued           1348 drivers/media/usb/tm6000/tm6000-video.c 		list_empty(&dev->vidq.queued));
queued           1640 drivers/media/usb/tm6000/tm6000-video.c 	INIT_LIST_HEAD(&dev->vidq.queued);
queued             81 drivers/media/usb/tm6000/tm6000.h 	struct list_head       queued;
queued            572 drivers/mtd/ftl.c     int queued, ret;
queued            579 drivers/mtd/ftl.c 	queued = 0;
queued            590 drivers/mtd/ftl.c 		queued = 1;
queued            610 drivers/mtd/ftl.c 	    if (queued) {
queued           3195 drivers/net/wireless/ath/ath6kl/cfg80211.c 	bool more_data, queued;
queued           3233 drivers/net/wireless/ath/ath6kl/cfg80211.c 		queued = ath6kl_mgmt_powersave_ap(vif, id, freq, wait, buf, len,
queued           3235 drivers/net/wireless/ath/ath6kl/cfg80211.c 		if (queued)
queued            598 drivers/net/wireless/ath/ath9k/debug.c 	PR("MPDUs Queued:    ", queued);
queued           1329 drivers/net/wireless/ath/ath9k/debug.c 	AWDATA(queued);
queued            171 drivers/net/wireless/ath/ath9k/debug.h 	u32 queued;
queued           2063 drivers/net/wireless/ath/ath9k/xmit.c 	TX_STAT_INC(sc, txq->axq_qnum, queued);
queued           2389 drivers/net/wireless/ath/ath9k/xmit.c 	TX_STAT_INC(sc, txctl.txq->axq_qnum, queued);
queued           1432 drivers/net/wireless/ath/carl9170/tx.c 		goto queued;
queued           1440 drivers/net/wireless/ath/carl9170/tx.c 			goto queued;
queued           1445 drivers/net/wireless/ath/carl9170/tx.c queued:
queued             42 drivers/net/wireless/mediatek/mt76/debugfs.c 			   i, q->q->queued, q->q->head, q->q->tail,
queued             85 drivers/net/wireless/mediatek/mt76/dma.c 		q->queued++;
queued            157 drivers/net/wireless/mediatek/mt76/dma.c 	while ((q->queued > n_queued) && q->tail != last) {
queued            180 drivers/net/wireless/mediatek/mt76/dma.c 	q->queued -= n_queued;
queued            192 drivers/net/wireless/mediatek/mt76/dma.c 	       qid < IEEE80211_NUM_ACS && q->queued < q->ndesc - 8;
queued            196 drivers/net/wireless/mediatek/mt76/dma.c 	if (!q->queued)
queued            238 drivers/net/wireless/mediatek/mt76/dma.c 	if (!q->queued)
queued            245 drivers/net/wireless/mediatek/mt76/dma.c 	q->queued--;
queued            339 drivers/net/wireless/mediatek/mt76/dma.c 	if (q->queued + (tx_info.nbuf + 1) / 2 >= q->ndesc - 1) {
queued            372 drivers/net/wireless/mediatek/mt76/dma.c 	while (q->queued < q->ndesc - 1) {
queued            393 drivers/net/wireless/mediatek/mt76/mac80211.c 		if (q && q->queued)
queued            119 drivers/net/wireless/mediatek/mt76/mt76.h 	int queued;
queued            139 drivers/net/wireless/mediatek/mt76/mt7603/beacon.c 	if (dev->mt76.q_tx[MT_TXQ_BEACON].q->queued >
queued           1400 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 		if (!q->queued)
queued            366 drivers/net/wireless/mediatek/mt76/mt76x02_mmio.c 		if (!q->queued)
queued            284 drivers/net/wireless/mediatek/mt76/tx.c 	if (q->queued > q->ndesc - 8 && !q->stopped) {
queued            398 drivers/net/wireless/mediatek/mt76/usb.c 	if (q->queued > 0) {
queued            401 drivers/net/wireless/mediatek/mt76/usb.c 		q->queued--;
queued            521 drivers/net/wireless/mediatek/mt76/usb.c 	q->queued++;
queued            576 drivers/net/wireless/mediatek/mt76/usb.c 	q->queued = 0;
queued            667 drivers/net/wireless/mediatek/mt76/usb.c 		while (q->queued > n_dequeued) {
queued            687 drivers/net/wireless/mediatek/mt76/usb.c 		q->queued -= n_dequeued;
queued            689 drivers/net/wireless/mediatek/mt76/usb.c 		wake = q->stopped && q->queued < q->ndesc - 8;
queued            693 drivers/net/wireless/mediatek/mt76/usb.c 		if (!q->queued)
queued            779 drivers/net/wireless/mediatek/mt76/usb.c 	if (q->queued == q->ndesc)
queued            797 drivers/net/wireless/mediatek/mt76/usb.c 	q->queued++;
queued            901 drivers/net/wireless/mediatek/mt76/usb.c 			while (q->queued) {
queued            904 drivers/net/wireless/mediatek/mt76/usb.c 				q->queued--;
queued           1586 drivers/net/wireless/st/cw1200/wsm.c 	int queued;
queued           1591 drivers/net/wireless/st/cw1200/wsm.c 		queued = cw1200_queue_get_num_queued(&priv->tx_queue[i],
queued           1593 drivers/net/wireless/st/cw1200/wsm.c 		if (!queued)
queued           1595 drivers/net/wireless/st/cw1200/wsm.c 		*total += queued;
queued           1568 drivers/scsi/ncr53c8xx.c 	u_char		queued;
queued           4457 drivers/scsi/ncr53c8xx.c 	cp->queued = 1;
queued           5093 drivers/scsi/ncr53c8xx.c 		if (cp->queued) {
queued           5097 drivers/scsi/ncr53c8xx.c 	if (cp->queued) {
queued           5099 drivers/scsi/ncr53c8xx.c 		cp->queued = 0;
queued           7283 drivers/scsi/ncr53c8xx.c 		if (cp->queued) {
queued           7289 drivers/scsi/ncr53c8xx.c 	if (cp->queued) {
queued           7291 drivers/scsi/ncr53c8xx.c 		cp->queued = 0;
queued            292 drivers/scsi/qla2xxx/qla_target.c 	uint8_t queued = 0;
queued            318 drivers/scsi/qla2xxx/qla_target.c 			if (!queued) {
queued            319 drivers/scsi/qla2xxx/qla_target.c 				queued = 1;
queued           1665 drivers/spi/spi.c 	ctlr->queued = true;
queued           2590 drivers/spi/spi.c 	if (ctlr->queued) {
queued           2612 drivers/spi/spi.c 	if (!ctlr->queued)
queued           2627 drivers/spi/spi.c 	if (!ctlr->queued)
queued            464 drivers/tty/hvc/hvc_iucv.c 	int queued;
queued            473 drivers/tty/hvc/hvc_iucv.c 	queued = hvc_iucv_queue(priv, buf, count);
queued            476 drivers/tty/hvc/hvc_iucv.c 	return queued;
queued            255 drivers/tty/serial/amba-pl011.c 	bool			queued;
queued            556 drivers/tty/serial/amba-pl011.c 	if (uap->dmatx.queued)
queued            575 drivers/tty/serial/amba-pl011.c 		uap->dmatx.queued = false;
queued            615 drivers/tty/serial/amba-pl011.c 		uap->dmatx.queued = false;
queued            647 drivers/tty/serial/amba-pl011.c 		uap->dmatx.queued = false;
queued            656 drivers/tty/serial/amba-pl011.c 		uap->dmatx.queued = false;
queued            677 drivers/tty/serial/amba-pl011.c 	uap->dmatx.queued = true;
queued            710 drivers/tty/serial/amba-pl011.c 	if (uap->dmatx.queued) {
queued            736 drivers/tty/serial/amba-pl011.c 	if (uap->dmatx.queued) {
queued            761 drivers/tty/serial/amba-pl011.c 		if (!uap->dmatx.queued) {
queued            818 drivers/tty/serial/amba-pl011.c 	if (uap->dmatx.queued) {
queued            821 drivers/tty/serial/amba-pl011.c 		uap->dmatx.queued = false;
queued           1204 drivers/tty/serial/amba-pl011.c 		if (uap->dmatx.queued) {
queued           1207 drivers/tty/serial/amba-pl011.c 			uap->dmatx.queued = false;
queued            224 drivers/usb/dwc3/trace.h 		__field(u32, queued)
queued            130 drivers/usb/mon/mon_bin.c 	u32 queued;
queued           1120 drivers/usb/mon/mon_bin.c 		if (put_user(nevents, &sp->queued))
queued           1798 drivers/usb/serial/io_ti.c 	int queued;
queued           1800 drivers/usb/serial/io_ti.c 	queued = tty_insert_flip_string(&port->port, data, length);
queued           1801 drivers/usb/serial/io_ti.c 	if (queued < length)
queued           1803 drivers/usb/serial/io_ti.c 			__func__, length - queued);
queued            302 fs/f2fs/f2fs.h 	unsigned char queued;		/* queued discard */
queued            954 fs/f2fs/segment.c 	dc->queued = 0;
queued            987 fs/f2fs/segment.c 		atomic_sub(dc->queued, &dcc->queued_discard);
queued           1194 fs/f2fs/segment.c 		dc->queued++;
queued            517 fs/fuse/dev.c  	bool queued = false;
queued            536 fs/fuse/dev.c  		queued = true;
queued            540 fs/fuse/dev.c  	return queued;
queued            816 fs/ocfs2/cluster/heartbeat.c 	int queued = 0;
queued            834 fs/ocfs2/cluster/heartbeat.c 			queued = 1;
queued            839 fs/ocfs2/cluster/heartbeat.c 	if (queued)
queued            900 fs/ocfs2/cluster/heartbeat.c 	int queued = 0;
queued            994 fs/ocfs2/cluster/heartbeat.c 			queued = 1;
queued           1046 fs/ocfs2/cluster/heartbeat.c 			queued = 1;
queued           1062 fs/ocfs2/cluster/heartbeat.c 	if (queued)
queued            102 fs/xfs/xfs_mru_cache.c 	unsigned int		queued;	   /* work has been queued */
queued            204 fs/xfs/xfs_mru_cache.c 		if (!mru->queued) {
queued            205 fs/xfs/xfs_mru_cache.c 			mru->queued = 1;
queued            280 fs/xfs/xfs_mru_cache.c 	mru->queued = next;
queued            281 fs/xfs/xfs_mru_cache.c 	if ((mru->queued > 0)) {
queued            388 fs/xfs/xfs_mru_cache.c 	if (mru->queued) {
queued             77 include/drm/drm_flip_work.h 	struct list_head queued;
queued             51 include/linux/blk-mq.h 	unsigned long		queued;
queued            301 include/linux/kvm_host.h 		u32 queued;
queued            550 include/linux/spi/spi.h 	bool				queued;
queued           1589 kernel/sched/core.c 	bool queued, running;
queued           1593 kernel/sched/core.c 	queued = task_on_rq_queued(p);
queued           1596 kernel/sched/core.c 	if (queued) {
queued           1609 kernel/sched/core.c 	if (queued)
queued           1889 kernel/sched/core.c 	int running, queued;
queued           1928 kernel/sched/core.c 		queued = task_on_rq_queued(p);
queued           1960 kernel/sched/core.c 		if (unlikely(queued)) {
queued           4372 kernel/sched/core.c 	int prio, oldprio, queued, running, queue_flag =
queued           4432 kernel/sched/core.c 	queued = task_on_rq_queued(p);
queued           4434 kernel/sched/core.c 	if (queued)
queued           4472 kernel/sched/core.c 	if (queued)
queued           4495 kernel/sched/core.c 	bool queued, running;
queued           4519 kernel/sched/core.c 	queued = task_on_rq_queued(p);
queued           4521 kernel/sched/core.c 	if (queued)
queued           4532 kernel/sched/core.c 	if (queued) {
queued           4755 kernel/sched/core.c 	int retval, oldprio, oldpolicy = -1, queued, running;
queued           4967 kernel/sched/core.c 	queued = task_on_rq_queued(p);
queued           4969 kernel/sched/core.c 	if (queued)
queued           4979 kernel/sched/core.c 	if (queued) {
queued           6147 kernel/sched/core.c 	bool queued, running;
queued           6152 kernel/sched/core.c 	queued = task_on_rq_queued(p);
queued           6155 kernel/sched/core.c 	if (queued)
queued           6162 kernel/sched/core.c 	if (queued)
queued           7035 kernel/sched/core.c 	int queued, running, queue_flags =
queued           7044 kernel/sched/core.c 	queued = task_on_rq_queued(tsk);
queued           7046 kernel/sched/core.c 	if (queued)
queued           7053 kernel/sched/core.c 	if (queued)
queued           1812 kernel/sched/deadline.c static void task_tick_dl(struct rq *rq, struct task_struct *p, int queued)
queued           1822 kernel/sched/deadline.c 	if (hrtick_enabled(rq) && queued && p->dl.runtime > 0 &&
queued           4281 kernel/sched/fair.c entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr, int queued)
queued           4299 kernel/sched/fair.c 	if (queued) {
queued           9979 kernel/sched/fair.c static void task_tick_fair(struct rq *rq, struct task_struct *curr, int queued)
queued           9986 kernel/sched/fair.c 		entity_tick(cfs_rq, se, queued);
queued            428 kernel/sched/idle.c static void task_tick_idle(struct rq *rq, struct task_struct *curr, int queued)
queued           2315 kernel/sched/rt.c static void task_tick_rt(struct rq *rq, struct task_struct *p, int queued)
queued           1753 kernel/sched/sched.h 	void (*task_tick)(struct rq *rq, struct task_struct *p, int queued);
queued             93 kernel/sched/stop_task.c static void task_tick_stop(struct rq *rq, struct task_struct *curr, int queued)
queued            379 kernel/stop_machine.c 	bool queued = false;
queued            395 kernel/stop_machine.c 			queued = true;
queued            401 kernel/stop_machine.c 	return queued;
queued            147 net/ax25/ax25_ds_in.c 	int queued = 0;
queued            240 net/ax25/ax25_ds_in.c 			queued = ax25_rx_iframe(ax25, skb);
queued            273 net/ax25/ax25_ds_in.c 	return queued;
queued            281 net/ax25/ax25_ds_in.c 	int queued = 0, frametype, ns, nr, pf;
queued            287 net/ax25/ax25_ds_in.c 		queued = ax25_ds_state1_machine(ax25, skb, frametype, pf, type);
queued            290 net/ax25/ax25_ds_in.c 		queued = ax25_ds_state2_machine(ax25, skb, frametype, pf, type);
queued            293 net/ax25/ax25_ds_in.c 		queued = ax25_ds_state3_machine(ax25, skb, frametype, ns, nr, pf, type);
queued            297 net/ax25/ax25_ds_in.c 	return queued;
queued            103 net/ax25/ax25_in.c 	int queued = 0;
queued            145 net/ax25/ax25_in.c 				queued = 1;
queued            151 net/ax25/ax25_in.c 	return queued;
queued            159 net/ax25/ax25_in.c 	int queued = 0;
queued            167 net/ax25/ax25_in.c 		queued = ax25_std_frame_in(ax25, skb, type);
queued            173 net/ax25/ax25_in.c 			queued = ax25_ds_frame_in(ax25, skb, type);
queued            175 net/ax25/ax25_in.c 			queued = ax25_std_frame_in(ax25, skb, type);
queued            180 net/ax25/ax25_in.c 	return queued;
queued            143 net/ax25/ax25_std_in.c 	int queued = 0;
queued            225 net/ax25/ax25_std_in.c 			queued = ax25_rx_iframe(ax25, skb);
queued            258 net/ax25/ax25_std_in.c 	return queued;
queued            268 net/ax25/ax25_std_in.c 	int queued = 0;
queued            380 net/ax25/ax25_std_in.c 			queued = ax25_rx_iframe(ax25, skb);
queued            413 net/ax25/ax25_std_in.c 	return queued;
queued            421 net/ax25/ax25_std_in.c 	int queued = 0, frametype, ns, nr, pf;
queued            427 net/ax25/ax25_std_in.c 		queued = ax25_std_state1_machine(ax25, skb, frametype, pf, type);
queued            430 net/ax25/ax25_std_in.c 		queued = ax25_std_state2_machine(ax25, skb, frametype, pf, type);
queued            433 net/ax25/ax25_std_in.c 		queued = ax25_std_state3_machine(ax25, skb, frametype, ns, nr, pf, type);
queued            436 net/ax25/ax25_std_in.c 		queued = ax25_std_state4_machine(ax25, skb, frametype, ns, nr, pf, type);
queued            442 net/ax25/ax25_std_in.c 	return queued;
queued            130 net/caif/caif_socket.c 	bool queued = false;
queued            153 net/caif/caif_socket.c 	queued = !sock_flag(sk, SOCK_DEAD);
queued            154 net/caif/caif_socket.c 	if (queued)
queued            158 net/caif/caif_socket.c 	if (queued)
queued             45 net/dccp/input.c 	int queued = 0;
queued             76 net/dccp/input.c 		queued = 1;
queued             86 net/dccp/input.c 	return queued;
queued             91 net/dccp/input.c 	int queued = 0;
queued            101 net/dccp/input.c 		return queued;
queued            113 net/dccp/input.c 		queued = 1;
queued            120 net/dccp/input.c 	return queued;
queued            524 net/dccp/input.c 	int queued = 0;
queued            562 net/dccp/input.c 			queued = 1; /* packet was queued
queued            568 net/dccp/input.c 	return queued;
queued            578 net/dccp/input.c 	int queued = 0;
queued            676 net/dccp/input.c 		queued = dccp_rcv_request_sent_state_process(sk, skb, dh, len);
queued            677 net/dccp/input.c 		if (queued >= 0)
queued            678 net/dccp/input.c 			return queued;
queued            689 net/dccp/input.c 		queued = dccp_rcv_respond_partopen_state_process(sk, skb,
queued            707 net/dccp/input.c 	if (!queued) {
queued            609 net/decnet/dn_nsp_in.c 	int queued = 0;
queued            622 net/decnet/dn_nsp_in.c 			queued = 1;
queued            628 net/decnet/dn_nsp_in.c 	if (!queued)
queued            634 net/decnet/dn_nsp_in.c 	int queued = 0;
queued            648 net/decnet/dn_nsp_in.c 			queued = 1;
queued            659 net/decnet/dn_nsp_in.c 	if (!queued)
queued           6141 net/ipv4/tcp_input.c 	int queued = 0;
queued           6177 net/ipv4/tcp_input.c 		queued = tcp_rcv_synsent_state_process(sk, skb, th);
queued           6178 net/ipv4/tcp_input.c 		if (queued >= 0)
queued           6179 net/ipv4/tcp_input.c 			return queued;
queued           6358 net/ipv4/tcp_input.c 		queued = 1;
queued           6368 net/ipv4/tcp_input.c 	if (!queued) {
queued            249 net/lapb/lapb_in.c 	int queued = 0;
queued            403 net/lapb/lapb_in.c 			queued = 1;
queued            464 net/lapb/lapb_in.c 	if (!queued)
queued            555 net/mac80211/offchannel.c 	bool queued = false, combine_started = true;
queued            634 net/mac80211/offchannel.c 			queued = true;
queued            662 net/mac80211/offchannel.c 			queued = true;
queued            669 net/mac80211/offchannel.c 		queued = ieee80211_coalesce_hw_started_roc(local, roc, tmp);
queued            670 net/mac80211/offchannel.c 		if (queued)
queued            680 net/mac80211/offchannel.c 	if (!queued)
queued           1087 net/mac80211/tx.c 	bool queued = false;
queued           1127 net/mac80211/tx.c 			queued = true;
queued           1151 net/mac80211/tx.c 	return queued;
queued           1211 net/mac80211/tx.c 			bool queued;
queued           1213 net/mac80211/tx.c 			queued = ieee80211_tx_prep_agg(tx, skb, info,
queued           1216 net/mac80211/tx.c 			if (unlikely(queued))
queued            779 net/netfilter/nfnetlink_queue.c 	unsigned int queued;
queued            816 net/netfilter/nfnetlink_queue.c 	queued = 0;
queued            824 net/netfilter/nfnetlink_queue.c 			queued++;
queued            830 net/netfilter/nfnetlink_queue.c 	if (queued) {
queued            153 net/netrom/nr_in.c 	int queued = 0;
queued            226 net/netrom/nr_in.c 		queued = 1;
queued            273 net/netrom/nr_in.c 	return queued;
queued            280 net/netrom/nr_in.c 	int queued = 0, frametype;
queued            289 net/netrom/nr_in.c 		queued = nr_state1_machine(sk, skb, frametype);
queued            292 net/netrom/nr_in.c 		queued = nr_state2_machine(sk, skb, frametype);
queued            295 net/netrom/nr_in.c 		queued = nr_state3_machine(sk, skb, frametype);
queued            301 net/netrom/nr_in.c 	return queued;
queued            821 net/rds/send.c 			     __be16 dport, int *queued)
queued            826 net/rds/send.c 	if (*queued)
queued            877 net/rds/send.c 		*queued = 1;
queued            882 net/rds/send.c 	return *queued;
queued           1111 net/rds/send.c 	int queued = 0, allocated_mr = 0;
queued           1351 net/rds/send.c 				  dport, &queued)) {
queued           1363 net/rds/send.c 							  &queued),
queued           1365 net/rds/send.c 		rdsdebug("sendmsg woke queued %d timeo %ld\n", queued, timeo);
queued            104 net/rose/rose_in.c 	int queued = 0;
queued            167 net/rose/rose_in.c 				queued = 1;
queued            204 net/rose/rose_in.c 	return queued;
queued            265 net/rose/rose_in.c 	int queued = 0, frametype, ns, nr, q, d, m;
queued            274 net/rose/rose_in.c 		queued = rose_state1_machine(sk, skb, frametype);
queued            277 net/rose/rose_in.c 		queued = rose_state2_machine(sk, skb, frametype);
queued            280 net/rose/rose_in.c 		queued = rose_state3_machine(sk, skb, frametype, ns, nr, q, d, m);
queued            283 net/rose/rose_in.c 		queued = rose_state4_machine(sk, skb, frametype);
queued            286 net/rose/rose_in.c 		queued = rose_state5_machine(sk, skb, frametype);
queued            292 net/rose/rose_in.c 	return queued;
queued             51 net/x25/x25_dev.c 		int queued = 1;
queued             56 net/x25/x25_dev.c 			queued = x25_process_rx_frame(sk, skb);
queued             58 net/x25/x25_dev.c 			queued = !sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf));
queued             62 net/x25/x25_dev.c 		return queued;
queued            210 net/x25/x25_in.c 	int queued = 0;
queued            277 net/x25/x25_in.c 					queued = 1;
queued            315 net/x25/x25_in.c 				queued = !sock_queue_rcv_skb(sk, skb);
queued            319 net/x25/x25_in.c 				queued = 1;
queued            330 net/x25/x25_in.c 	return queued;
queued            389 net/x25/x25_in.c 	int queued = 0, frametype, ns, nr, q, d, m;
queued            398 net/x25/x25_in.c 		queued = x25_state1_machine(sk, skb, frametype);
queued            401 net/x25/x25_in.c 		queued = x25_state2_machine(sk, skb, frametype);
queued            404 net/x25/x25_in.c 		queued = x25_state3_machine(sk, skb, frametype, ns, nr, q, d, m);
queued            407 net/x25/x25_in.c 		queued = x25_state4_machine(sk, skb, frametype);
queued            413 net/x25/x25_in.c 	return queued;
queued            418 net/x25/x25_in.c 	int queued = x25_process_rx_frame(sk, skb);
queued            420 net/x25/x25_in.c 	if (!queued)
queued            127 sound/firewire/fireworks/fireworks_hwdep.c 	bool queued;
queued            132 sound/firewire/fireworks/fireworks_hwdep.c 	queued = efw->push_ptr != efw->pull_ptr;
queued            134 sound/firewire/fireworks/fireworks_hwdep.c 	while (!dev_lock_changed && !queued) {
queued            143 sound/firewire/fireworks/fireworks_hwdep.c 		queued = efw->push_ptr != efw->pull_ptr;
queued            150 sound/firewire/fireworks/fireworks_hwdep.c 	else if (queued)
queued            144 virt/kvm/async_pf.c 	vcpu->async_pf.queued = 0;
queued            163 virt/kvm/async_pf.c 		vcpu->async_pf.queued--;
queued            173 virt/kvm/async_pf.c 	if (vcpu->async_pf.queued >= ASYNC_PF_PER_VCPU)
queued            205 virt/kvm/async_pf.c 	vcpu->async_pf.queued++;
queued            233 virt/kvm/async_pf.c 	vcpu->async_pf.queued++;