cfs              3531 kernel/sched/core.c 	struct sched_entity *curr = (&task_rq(p)->cfs)->curr;
cfs              3917 kernel/sched/core.c 		   rq->nr_running == rq->cfs.h_nr_running)) {
cfs              6627 kernel/sched/core.c 		init_cfs_rq(&rq->cfs);
cfs              6654 kernel/sched/core.c 		init_tg_cfs_entry(&root_task_group, &rq->cfs, NULL, i, NULL);
cfs               508 kernel/sched/debug.c 	rq0_min_vruntime = cpu_rq(0)->cfs.min_vruntime;
cfs               448 kernel/sched/fair.c 	return &task_rq(p)->cfs;
cfs               456 kernel/sched/fair.c 	return &rq->cfs;
cfs               485 kernel/sched/fair.c 		for (cfs_rq = &rq->cfs, pos = NULL; cfs_rq; cfs_rq = pos)
cfs              3113 kernel/sched/fair.c 	if (&rq->cfs == cfs_rq || (flags & SCHED_CPUFREQ_MIGRATION)) {
cfs              4184 kernel/sched/fair.c 	    rq_of(cfs_rq)->cfs.load.weight >= 2*se->load.weight) {
cfs              4616 kernel/sched/fair.c 	if (rq->curr == rq->idle && rq->cfs.nr_running)
cfs              5142 kernel/sched/fair.c 	if (rq->cfs.h_nr_running > 1) {
cfs              5219 kernel/sched/fair.c 	util_est_enqueue(&rq->cfs, p);
cfs              5367 kernel/sched/fair.c 	util_est_dequeue(&rq->cfs, p, task_sleep);
cfs              5394 kernel/sched/fair.c 	return unlikely(rq->nr_running == rq->cfs.idle_h_nr_running &&
cfs              5400 kernel/sched/fair.c 	return cfs_rq_runnable_load_avg(&rq->cfs);
cfs              5411 kernel/sched/fair.c 	unsigned long nr_running = READ_ONCE(rq->cfs.h_nr_running);
cfs              5619 kernel/sched/fair.c 			avg_load += cfs_rq_load_avg(&cpu_rq(i)->cfs);
cfs              6106 kernel/sched/fair.c 	cfs_rq = &cpu_rq(cpu)->cfs;
cfs              6137 kernel/sched/fair.c 	cfs_rq = &cpu_rq(cpu)->cfs;
cfs              6237 kernel/sched/fair.c 	struct cfs_rq *cfs_rq = &cpu_rq(cpu)->cfs;
cfs              6778 kernel/sched/fair.c 	struct cfs_rq *cfs_rq = &rq->cfs;
cfs              6821 kernel/sched/fair.c 				cfs_rq = &rq->cfs;
cfs              7603 kernel/sched/fair.c 			if (cfs_rq == &rq->cfs)
cfs              7676 kernel/sched/fair.c 	struct cfs_rq *cfs_rq = &rq->cfs;
cfs              8069 kernel/sched/fair.c 		sgs->sum_nr_running += rq->cfs.h_nr_running;
cfs              8763 kernel/sched/fair.c 	    (env->src_rq->cfs.h_nr_running == 1)) {
cfs              9445 kernel/sched/fair.c 		if (rq->cfs.h_nr_running >= 1 && check_cpu_capacity(rq, sd)) {
cfs              9894 kernel/sched/fair.c 	if (this_rq->cfs.h_nr_running && !pulled_task)
cfs              9902 kernel/sched/fair.c 	if (this_rq->nr_running != this_rq->cfs.h_nr_running)
cfs              10380 kernel/sched/fair.c 		se->cfs_rq = &rq->cfs;
cfs              10456 kernel/sched/fair.c 	if (rq->cfs.load.weight)
cfs               100 kernel/sched/pelt.h 	u32 util_sum = rq->cfs.avg.util_sum;
cfs               887 kernel/sched/sched.h 	struct cfs_rq		cfs;
cfs              1022 kernel/sched/sched.h 	return container_of(cfs_rq, struct rq, cfs);
cfs              1827 kernel/sched/sched.h 	return rq->cfs.nr_running > 0;
cfs              2407 kernel/sched/sched.h 	unsigned long util = READ_ONCE(rq->cfs.avg.util_avg);
cfs              2411 kernel/sched/sched.h 			     READ_ONCE(rq->cfs.avg.util_est.enqueued));
cfs                38 net/netfilter/xt_set.c #define ADT_OPT(n, f, d, fs, cfs, t, p, b, po, bo)	\
cfs                43 net/netfilter/xt_set.c 	.cmdflags = cfs,				\