inflight          270 block/blk-iolatency.c 	atomic_dec(&rqw->inflight);
inflight          299 block/blk-iolatency.c 		atomic_inc(&rqw->inflight);
inflight          597 block/blk-iolatency.c 	int inflight = 0;
inflight          619 block/blk-iolatency.c 		inflight = atomic_dec_return(&rqw->inflight);
inflight          620 block/blk-iolatency.c 		WARN_ON_ONCE(inflight < 0);
inflight           96 block/blk-mq.c 	unsigned int *inflight;
inflight          109 block/blk-mq.c 		mi->inflight[0]++;
inflight          116 block/blk-mq.c 	unsigned inflight[2];
inflight          117 block/blk-mq.c 	struct mq_inflight mi = { .part = part, .inflight = inflight, };
inflight          119 block/blk-mq.c 	inflight[0] = inflight[1] = 0;
inflight          122 block/blk-mq.c 	return inflight[0];
inflight          132 block/blk-mq.c 		mi->inflight[rq_data_dir(rq)]++;
inflight          138 block/blk-mq.c 			 unsigned int inflight[2])
inflight          140 block/blk-mq.c 	struct mq_inflight mi = { .part = part, .inflight = inflight, };
inflight          142 block/blk-mq.c 	inflight[0] = inflight[1] = 0;
inflight          190 block/blk-mq.h 			 unsigned int inflight[2]);
inflight           29 block/blk-rq-qos.c 	return atomic_inc_below(&rq_wait->inflight, limit);
inflight           23 block/blk-rq-qos.h 	atomic_t inflight;
inflight           96 block/blk-rq-qos.h 	atomic_set(&rq_wait->inflight, 0);
inflight          130 block/blk-wbt.c 	int inflight, limit;
inflight          132 block/blk-wbt.c 	inflight = atomic_dec_return(&rqw->inflight);
inflight          158 block/blk-wbt.c 	if (inflight && inflight >= limit)
inflight          162 block/blk-wbt.c 		int diff = limit - inflight;
inflight          164 block/blk-wbt.c 		if (!inflight || diff >= rwb->wb_background / 2)
inflight          356 block/blk-wbt.c 	unsigned int inflight = wbt_inflight(rwb);
inflight          362 block/blk-wbt.c 			inflight);
inflight          404 block/blk-wbt.c 	if (rqd->scale_step || inflight)
inflight          753 block/blk-wbt.c 			   atomic_read(&rwb->rq_wait[i].inflight));
inflight           82 block/blk-wbt.h 		ret += atomic_read(&rwb->rq_wait[i].inflight);
inflight           72 block/genhd.c  	unsigned int inflight;
inflight           78 block/genhd.c  	inflight = 0;
inflight           80 block/genhd.c  		inflight += part_stat_local_read_cpu(part, in_flight[0], cpu) +
inflight           83 block/genhd.c  	if ((int)inflight < 0)
inflight           84 block/genhd.c  		inflight = 0;
inflight           86 block/genhd.c  	return inflight;
inflight           90 block/genhd.c  		       unsigned int inflight[2])
inflight           95 block/genhd.c  		blk_mq_in_flight_rw(q, part, inflight);
inflight           99 block/genhd.c  	inflight[0] = 0;
inflight          100 block/genhd.c  	inflight[1] = 0;
inflight          102 block/genhd.c  		inflight[0] += part_stat_local_read_cpu(part, in_flight[0], cpu);
inflight          103 block/genhd.c  		inflight[1] += part_stat_local_read_cpu(part, in_flight[1], cpu);
inflight          105 block/genhd.c  	if ((int)inflight[0] < 0)
inflight          106 block/genhd.c  		inflight[0] = 0;
inflight          107 block/genhd.c  	if ((int)inflight[1] < 0)
inflight          108 block/genhd.c  		inflight[1] = 0;
inflight         1193 block/genhd.c  static DEVICE_ATTR(inflight, 0444, part_inflight_show, NULL);
inflight         1371 block/genhd.c  	unsigned int inflight;
inflight         1383 block/genhd.c  		inflight = part_in_flight(gp->queue, hd);
inflight         1399 block/genhd.c  			   inflight,
inflight          123 block/partition-generic.c 	unsigned int inflight;
inflight          125 block/partition-generic.c 	inflight = part_in_flight(q, p);
inflight          140 block/partition-generic.c 		inflight,
inflight          154 block/partition-generic.c 	unsigned int inflight[2];
inflight          156 block/partition-generic.c 	part_in_flight_rw(q, p, inflight);
inflight          157 block/partition-generic.c 	return sprintf(buf, "%8u %8u\n", inflight[0], inflight[1]);
inflight          190 block/partition-generic.c static DEVICE_ATTR(inflight, 0444, part_inflight_show, NULL);
inflight          746 drivers/block/xen-blkback/blkback.c 	if (atomic_dec_and_test(&ring->inflight) && atomic_read(&blkif->drain)) {
inflight         1067 drivers/block/xen-blkback/blkback.c 		if (atomic_read(&ring->inflight) == 0)
inflight         1360 drivers/block/xen-blkback/blkback.c 	atomic_inc(&ring->inflight);
inflight          259 drivers/block/xen-blkback/common.h 	atomic_t		inflight;
inflight          267 drivers/block/xen-blkback/xenbus.c 		if (atomic_read(&ring->inflight) > 0) {
inflight          443 drivers/crypto/cavium/cpt/cpt_hw_types.h 		u64 inflight:8;
inflight          445 drivers/crypto/cavium/cpt/cpt_hw_types.h 		u64 inflight:8;
inflight          390 drivers/crypto/cavium/cpt/cptvf_main.c 	vqx_inprg.s.inflight = val;
inflight          205 drivers/crypto/chelsio/chcr_algo.c 	atomic_inc(&dev->inflight);
inflight          211 drivers/crypto/chelsio/chcr_algo.c 	atomic_dec(&dev->inflight);
inflight           59 drivers/crypto/chelsio/chcr_core.c 	if (atomic_read(&dev->inflight)) {
inflight           63 drivers/crypto/chelsio/chcr_core.c 				atomic_read(&dev->inflight));
inflight           68 drivers/crypto/chelsio/chcr_core.c 				atomic_read(&dev->inflight));
inflight          106 drivers/crypto/chelsio/chcr_core.c 	atomic_set(&dev->inflight, 0);
inflight          125 drivers/crypto/chelsio/chcr_core.c 	atomic_set(&dev->inflight, 0);
inflight          249 drivers/crypto/chelsio/chcr_core.c 	if (atomic_read(&dev->inflight) != 0) {
inflight          146 drivers/crypto/chelsio/chcr_core.h 	atomic_t inflight;
inflight           42 drivers/gpu/drm/i915/gt/intel_context_types.h 	struct intel_engine_cs *inflight;
inflight           43 drivers/gpu/drm/i915/gt/intel_context_types.h #define intel_context_inflight(ce) ptr_mask_bits((ce)->inflight, 2)
inflight           44 drivers/gpu/drm/i915/gt/intel_context_types.h #define intel_context_inflight_count(ce) ptr_unmask_bits((ce)->inflight, 2)
inflight          134 drivers/gpu/drm/i915/gt/intel_engine.h 	GEM_BUG_ON(execlists->active - execlists->inflight >
inflight          486 drivers/gpu/drm/i915/gt/intel_engine_cs.c 		memset(execlists->inflight, 0, sizeof(execlists->inflight));
inflight          213 drivers/gpu/drm/i915/gt/intel_engine_types.h 	struct i915_request *inflight[EXECLIST_MAX_PORTS + 1 /* sentinel */];
inflight          582 drivers/gpu/drm/i915/gt/intel_lrc.c 	old = READ_ONCE(ce->inflight);
inflight          585 drivers/gpu/drm/i915/gt/intel_lrc.c 			WRITE_ONCE(ce->inflight, __execlists_schedule_in(rq));
inflight          588 drivers/gpu/drm/i915/gt/intel_lrc.c 	} while (!try_cmpxchg(&ce->inflight, &old, ptr_inc(old)));
inflight          636 drivers/gpu/drm/i915/gt/intel_lrc.c 	old = READ_ONCE(ce->inflight);
inflight          639 drivers/gpu/drm/i915/gt/intel_lrc.c 	while (!try_cmpxchg(&ce->inflight, &old, cur));
inflight          888 drivers/gpu/drm/i915/gt/intel_lrc.c 	const struct intel_engine_cs *inflight;
inflight          902 drivers/gpu/drm/i915/gt/intel_lrc.c 	inflight = intel_context_inflight(&ve->context);
inflight          903 drivers/gpu/drm/i915/gt/intel_lrc.c 	if (inflight && inflight != engine)
inflight         1220 drivers/gpu/drm/i915/gt/intel_lrc.c 				GEM_BUG_ON(READ_ONCE(ve->context.inflight));
inflight         1383 drivers/gpu/drm/i915/gt/intel_lrc.c 		memset(execlists->inflight, 0, sizeof(execlists->inflight));
inflight         1566 drivers/gpu/drm/i915/gt/intel_lrc.c 				memcpy(execlists->inflight,
inflight         1593 drivers/gpu/drm/i915/gt/intel_lrc.c 			GEM_BUG_ON(execlists->active - execlists->inflight >
inflight         3426 drivers/gpu/drm/i915/gt/intel_lrc.c 	GEM_BUG_ON(ve->context.inflight);
inflight          540 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 	struct i915_request **first = execlists->inflight;
inflight          573 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 						    port - execlists->inflight);
inflight          590 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 		*port = schedule_in(last, port - execlists->inflight);
inflight          594 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 	execlists->active = execlists->inflight;
inflight          606 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 	for (port = execlists->inflight; (rq = *port); port++) {
inflight          612 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 	if (port != execlists->inflight) {
inflight          613 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 		int idx = port - execlists->inflight;
inflight          614 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 		int rem = ARRAY_SIZE(execlists->inflight) - idx;
inflight          615 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 		memmove(execlists->inflight, port, rem * sizeof(*port));
inflight          651 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 		memset(execlists->inflight, 0, sizeof(execlists->inflight));
inflight         1135 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 	BUILD_BUG_ON(ARRAY_SIZE(engine->execlists.inflight) *
inflight          196 drivers/gpu/drm/i915/i915_scheduler.c 	const struct i915_request *inflight;
inflight          208 drivers/gpu/drm/i915/i915_scheduler.c 	inflight = execlists_active(&engine->execlists);
inflight          209 drivers/gpu/drm/i915/i915_scheduler.c 	if (!inflight)
inflight          219 drivers/gpu/drm/i915/i915_scheduler.c 	if (inflight->hw_context == rq->hw_context)
inflight          223 drivers/gpu/drm/i915/i915_scheduler.c 	if (need_preempt(prio, rq_prio(inflight)))
inflight           31 drivers/iommu/amd_iommu_v2.c 	atomic_t inflight;
inflight          418 drivers/iommu/amd_iommu_v2.c 	if (atomic_dec_and_test(&pasid_state->pri[tag].inflight) &&
inflight          561 drivers/iommu/amd_iommu_v2.c 	atomic_inc(&pasid_state->pri[tag].inflight);
inflight         3561 drivers/scsi/scsi_transport_fc.c 	int err = 0, inflight = 0;
inflight         3566 drivers/scsi/scsi_transport_fc.c 	inflight = bsg_job_get(job);
inflight         3568 drivers/scsi/scsi_transport_fc.c 	if (inflight && i->f->bsg_timeout) {
inflight         3580 drivers/scsi/scsi_transport_fc.c 	if (inflight)
inflight          114 drivers/vhost/scsi.c 	struct vhost_scsi_inflight *inflight;
inflight          233 drivers/vhost/scsi.c 	struct vhost_scsi_inflight *inflight;
inflight          235 drivers/vhost/scsi.c 	inflight = container_of(kref, struct vhost_scsi_inflight, kref);
inflight          236 drivers/vhost/scsi.c 	complete(&inflight->comp);
inflight          269 drivers/vhost/scsi.c 	struct vhost_scsi_inflight *inflight;
inflight          273 drivers/vhost/scsi.c 	inflight = &svq->inflights[svq->inflight_idx];
inflight          274 drivers/vhost/scsi.c 	kref_get(&inflight->kref);
inflight          276 drivers/vhost/scsi.c 	return inflight;
inflight          279 drivers/vhost/scsi.c static void vhost_scsi_put_inflight(struct vhost_scsi_inflight *inflight)
inflight          281 drivers/vhost/scsi.c 	kref_put(&inflight->kref, vhost_scsi_done_inflight);
inflight          339 drivers/vhost/scsi.c 	vhost_scsi_put_inflight(tv_cmd->inflight);
inflight          609 drivers/vhost/scsi.c 	cmd->inflight = vhost_scsi_get_inflight(vq);
inflight         2750 fs/io_uring.c  	unsigned inflight;
inflight         2759 fs/io_uring.c  	timeout = inflight = 0;
inflight         2764 fs/io_uring.c  		if (inflight) {
inflight         2780 fs/io_uring.c  					inflight = 0;
inflight         2787 fs/io_uring.c  				nr_events = inflight;
inflight         2790 fs/io_uring.c  			inflight -= nr_events;
inflight         2791 fs/io_uring.c  			if (!inflight)
inflight         2814 fs/io_uring.c  			if (inflight || !time_after(jiffies, timeout)) {
inflight         2856 fs/io_uring.c  		inflight += io_submit_sqes(ctx, to_submit, cur_mm != NULL,
inflight          403 include/linux/genhd.h 		       unsigned int inflight[2]);
inflight           61 include/net/af_unix.h 	atomic_long_t		inflight;
inflight           16 include/trace/events/page_pool.h 		 s32 inflight, u32 hold, u32 release),
inflight           18 include/trace/events/page_pool.h 	TP_ARGS(pool, inflight, hold, release),
inflight           22 include/trace/events/page_pool.h 		__field(s32,	inflight)
inflight           29 include/trace/events/page_pool.h 		__entry->inflight	= inflight;
inflight           35 include/trace/events/page_pool.h 	  __entry->pool, __entry->inflight, __entry->hold, __entry->release)
inflight          132 include/trace/events/wbt.h 		 int step, unsigned int inflight),
inflight          134 include/trace/events/wbt.h 	TP_ARGS(bdi, status, step, inflight),
inflight          140 include/trace/events/wbt.h 		__field(unsigned int, inflight)
inflight          148 include/trace/events/wbt.h 		__entry->inflight	= inflight;
inflight          152 include/trace/events/wbt.h 		  __entry->status, __entry->step, __entry->inflight)
inflight          747 kernel/rcu/rcutorture.c 	int inflight;
inflight          756 kernel/rcu/rcutorture.c 	smp_store_release(&rbip->inflight, 0);
inflight          802 kernel/rcu/rcutorture.c 	struct rcu_boost_inflight rbi = { .inflight = 0 };
inflight          846 kernel/rcu/rcutorture.c 			if (!smp_load_acquire(&rbi.inflight)) {
inflight          848 kernel/rcu/rcutorture.c 				smp_store_release(&rbi.inflight, 1);
inflight          866 kernel/rcu/rcutorture.c 		if (!failed && smp_load_acquire(&rbi.inflight))
inflight          892 kernel/rcu/rcutorture.c 	while (!kthread_should_stop() || smp_load_acquire(&rbi.inflight)) {
inflight           64 net/atm/pppoatm.c 	atomic_t inflight;
inflight          137 net/atm/pppoatm.c 	atomic_dec(&pvcc->inflight);
inflight          242 net/atm/pppoatm.c 	    atomic_inc_not_zero(&pvcc->inflight))
inflight          272 net/atm/pppoatm.c 	    atomic_inc_not_zero(&pvcc->inflight))
inflight          408 net/atm/pppoatm.c 	atomic_set(&pvcc->inflight, NONE_INFLIGHT);
inflight         1196 net/core/net-sysfs.c 	__ATTR(inflight, 0444, bql_show_inflight, NULL);
inflight          199 net/core/page_pool.c 	s32 inflight;
inflight          201 net/core/page_pool.c 	inflight = _distance(hold_cnt, release_cnt);
inflight          203 net/core/page_pool.c 	trace_page_pool_inflight(pool, inflight, hold_cnt, release_cnt);
inflight          204 net/core/page_pool.c 	WARN(inflight < 0, "Negative(%d) inflight packet-pages", inflight);
inflight          206 net/core/page_pool.c 	return inflight;
inflight          373 net/core/page_pool.c 	int inflight;
inflight          376 net/core/page_pool.c 	inflight = page_pool_inflight(pool);
inflight          377 net/core/page_pool.c 	if (!inflight)
inflight          380 net/core/page_pool.c 	return inflight;
inflight          387 net/core/page_pool.c 	int inflight;
inflight          389 net/core/page_pool.c 	inflight = page_pool_release(pool);
inflight          390 net/core/page_pool.c 	if (!inflight)
inflight          398 net/core/page_pool.c 			__func__, inflight, sec);
inflight          412 net/ipv4/tcp_bbr.c 	u32 inflight;
inflight          414 net/ipv4/tcp_bbr.c 	inflight = bbr_bdp(sk, bw, gain);
inflight          415 net/ipv4/tcp_bbr.c 	inflight = bbr_quantization_budget(sk, inflight);
inflight          417 net/ipv4/tcp_bbr.c 	return inflight;
inflight          559 net/ipv4/tcp_bbr.c 	u32 inflight, bw;
inflight          567 net/ipv4/tcp_bbr.c 	inflight = bbr_packets_in_net_at_edt(sk, rs->prior_in_flight);
inflight          578 net/ipv4/tcp_bbr.c 			 inflight >= bbr_inflight(sk, bw, bbr->pacing_gain));
inflight          585 net/ipv4/tcp_bbr.c 		inflight <= bbr_inflight(sk, bw, BBR_UNIT);
inflight          651 net/sctp/output.c 	size_t datasize, rwnd, inflight, flight_size;
inflight          670 net/sctp/output.c 	inflight = q->outstanding_bytes;
inflight          675 net/sctp/output.c 	if (datasize > rwnd && inflight > 0)
inflight          703 net/sctp/output.c 	if ((sctp_sk(asoc->base.sk)->nodelay || inflight == 0) &&
inflight          787 net/unix/af_unix.c 	atomic_long_set(&u->inflight, 0);
inflight          169 net/unix/garbage.c 	atomic_long_dec(&usk->inflight);
inflight          174 net/unix/garbage.c 	atomic_long_inc(&usk->inflight);
inflight          179 net/unix/garbage.c 	atomic_long_inc(&u->inflight);
inflight          237 net/unix/garbage.c 		inflight_refs = atomic_long_read(&u->inflight);
inflight          268 net/unix/garbage.c 		if (atomic_long_read(&u->inflight) > 0) {
inflight           56 net/unix/scm.c 		if (atomic_long_inc_return(&u->inflight) == 1) {
inflight           77 net/unix/scm.c 		BUG_ON(!atomic_long_read(&u->inflight));
inflight           80 net/unix/scm.c 		if (atomic_long_dec_and_test(&u->inflight))
inflight           80 tools/io_uring/io_uring-bench.c 	int inflight;
inflight          272 tools/io_uring/io_uring-bench.c 	s->inflight -= reaped;
inflight          292 tools/io_uring/io_uring-bench.c 		if (!prepped && s->inflight < DEPTH) {
inflight          293 tools/io_uring/io_uring-bench.c 			to_prep = min(DEPTH - s->inflight, BATCH_SUBMIT);
inflight          296 tools/io_uring/io_uring-bench.c 		s->inflight += prepped;
inflight          300 tools/io_uring/io_uring-bench.c 		if (to_submit && (s->inflight + to_submit <= DEPTH))
inflight          303 tools/io_uring/io_uring-bench.c 			to_wait = min(s->inflight + to_submit, BATCH_COMPLETE);
inflight          341 tools/io_uring/io_uring-bench.c 				if (s->inflight)
inflight          581 tools/io_uring/io_uring-bench.c 				this_done - done, rpc, ipc, s->inflight,