rt_nr_running     685 kernel/sched/core.c 	fifo_nr_running = rq->rt.rt_nr_running - rq->rt.rr_nr_running;
rt_nr_running     578 kernel/sched/debug.c 	PU(rt_nr_running);
rt_nr_running     496 kernel/sched/rt.c 	if (rt_rq->rt_nr_running) {
rt_nr_running     592 kernel/sched/rt.c 	if (!rt_rq->rt_nr_running)
rt_nr_running     844 kernel/sched/rt.c 		skip = !rt_rq->rt_time && !rt_rq->rt_nr_running;
rt_nr_running     871 kernel/sched/rt.c 				if (rt_rq->rt_nr_running && rq->curr == rq->idle)
rt_nr_running     874 kernel/sched/rt.c 			if (rt_rq->rt_time || rt_rq->rt_nr_running)
rt_nr_running     877 kernel/sched/rt.c 		} else if (rt_rq->rt_nr_running) {
rt_nr_running    1007 kernel/sched/rt.c 	sub_nr_running(rq, rt_rq->rt_nr_running);
rt_nr_running    1025 kernel/sched/rt.c 	if (rt_rq->rt_nr_running) {
rt_nr_running    1026 kernel/sched/rt.c 		add_nr_running(rq, rt_rq->rt_nr_running);
rt_nr_running    1094 kernel/sched/rt.c 	if (rt_rq->rt_nr_running) {
rt_nr_running    1140 kernel/sched/rt.c 	WARN_ON(!rt_rq->rt_nr_running && rt_rq->rt_nr_boosted);
rt_nr_running    1162 kernel/sched/rt.c 		return group_rq->rt_nr_running;
rt_nr_running    1187 kernel/sched/rt.c 	rt_rq->rt_nr_running += rt_se_nr_running(rt_se);
rt_nr_running    1199 kernel/sched/rt.c 	WARN_ON(!rt_rq->rt_nr_running);
rt_nr_running    1200 kernel/sched/rt.c 	rt_rq->rt_nr_running -= rt_se_nr_running(rt_se);
rt_nr_running    1244 kernel/sched/rt.c 	if (group_rq && (rt_rq_throttled(group_rq) || !group_rq->rt_nr_running)) {
rt_nr_running    1319 kernel/sched/rt.c 		if (rt_rq && rt_rq->rt_nr_running)
rt_nr_running    2197 kernel/sched/rt.c 	if (!task_on_rq_queued(p) || rq->rt.rt_nr_running)
rt_nr_running     599 kernel/sched/sched.h 	unsigned int		rt_nr_running;
rt_nr_running     634 kernel/sched/sched.h 	return rt_rq->rt_queued && rt_rq->rt_nr_running;