cfs 3531 kernel/sched/core.c struct sched_entity *curr = (&task_rq(p)->cfs)->curr; cfs 3917 kernel/sched/core.c rq->nr_running == rq->cfs.h_nr_running)) { cfs 6627 kernel/sched/core.c init_cfs_rq(&rq->cfs); cfs 6654 kernel/sched/core.c init_tg_cfs_entry(&root_task_group, &rq->cfs, NULL, i, NULL); cfs 508 kernel/sched/debug.c rq0_min_vruntime = cpu_rq(0)->cfs.min_vruntime; cfs 448 kernel/sched/fair.c return &task_rq(p)->cfs; cfs 456 kernel/sched/fair.c return &rq->cfs; cfs 485 kernel/sched/fair.c for (cfs_rq = &rq->cfs, pos = NULL; cfs_rq; cfs_rq = pos) cfs 3113 kernel/sched/fair.c if (&rq->cfs == cfs_rq || (flags & SCHED_CPUFREQ_MIGRATION)) { cfs 4184 kernel/sched/fair.c rq_of(cfs_rq)->cfs.load.weight >= 2*se->load.weight) { cfs 4616 kernel/sched/fair.c if (rq->curr == rq->idle && rq->cfs.nr_running) cfs 5142 kernel/sched/fair.c if (rq->cfs.h_nr_running > 1) { cfs 5219 kernel/sched/fair.c util_est_enqueue(&rq->cfs, p); cfs 5367 kernel/sched/fair.c util_est_dequeue(&rq->cfs, p, task_sleep); cfs 5394 kernel/sched/fair.c return unlikely(rq->nr_running == rq->cfs.idle_h_nr_running && cfs 5400 kernel/sched/fair.c return cfs_rq_runnable_load_avg(&rq->cfs); cfs 5411 kernel/sched/fair.c unsigned long nr_running = READ_ONCE(rq->cfs.h_nr_running); cfs 5619 kernel/sched/fair.c avg_load += cfs_rq_load_avg(&cpu_rq(i)->cfs); cfs 6106 kernel/sched/fair.c cfs_rq = &cpu_rq(cpu)->cfs; cfs 6137 kernel/sched/fair.c cfs_rq = &cpu_rq(cpu)->cfs; cfs 6237 kernel/sched/fair.c struct cfs_rq *cfs_rq = &cpu_rq(cpu)->cfs; cfs 6778 kernel/sched/fair.c struct cfs_rq *cfs_rq = &rq->cfs; cfs 6821 kernel/sched/fair.c cfs_rq = &rq->cfs; cfs 7603 kernel/sched/fair.c if (cfs_rq == &rq->cfs) cfs 7676 kernel/sched/fair.c struct cfs_rq *cfs_rq = &rq->cfs; cfs 8069 kernel/sched/fair.c sgs->sum_nr_running += rq->cfs.h_nr_running; cfs 8763 kernel/sched/fair.c (env->src_rq->cfs.h_nr_running == 1)) { cfs 9445 kernel/sched/fair.c if (rq->cfs.h_nr_running >= 1 && check_cpu_capacity(rq, sd)) { cfs 9894 kernel/sched/fair.c if (this_rq->cfs.h_nr_running && !pulled_task) cfs 9902 kernel/sched/fair.c if (this_rq->nr_running != this_rq->cfs.h_nr_running) cfs 10380 kernel/sched/fair.c se->cfs_rq = &rq->cfs; cfs 10456 kernel/sched/fair.c if (rq->cfs.load.weight) cfs 100 kernel/sched/pelt.h u32 util_sum = rq->cfs.avg.util_sum; cfs 887 kernel/sched/sched.h struct cfs_rq cfs; cfs 1022 kernel/sched/sched.h return container_of(cfs_rq, struct rq, cfs); cfs 1827 kernel/sched/sched.h return rq->cfs.nr_running > 0; cfs 2407 kernel/sched/sched.h unsigned long util = READ_ONCE(rq->cfs.avg.util_avg); cfs 2411 kernel/sched/sched.h READ_ONCE(rq->cfs.avg.util_est.enqueued)); cfs 38 net/netfilter/xt_set.c #define ADT_OPT(n, f, d, fs, cfs, t, p, b, po, bo) \ cfs 43 net/netfilter/xt_set.c .cmdflags = cfs, \