runnable_load_sum 399 include/linux/sched.h u64 runnable_load_sum; runnable_load_sum 946 kernel/sched/debug.c P(se.avg.runnable_load_sum); runnable_load_sum 2839 kernel/sched/fair.c cfs_rq->avg.runnable_load_sum += se_runnable(se) * se->avg.runnable_load_sum; runnable_load_sum 2848 kernel/sched/fair.c sub_positive(&cfs_rq->avg.runnable_load_sum, runnable_load_sum 2849 kernel/sched/fair.c se_runnable(se) * se->avg.runnable_load_sum); runnable_load_sum 2897 kernel/sched/fair.c div_u64(se_runnable(se) * se->avg.runnable_load_sum, divider); runnable_load_sum 3314 kernel/sched/fair.c u64 runnable_load_sum, load_sum = 0; runnable_load_sum 3363 kernel/sched/fair.c runnable_load_sum = (s64)se_runnable(se) * runnable_sum; runnable_load_sum 3364 kernel/sched/fair.c runnable_load_avg = div_s64(runnable_load_sum, LOAD_AVG_MAX); runnable_load_sum 3365 kernel/sched/fair.c delta_sum = runnable_load_sum - se_weight(se) * se->avg.runnable_load_sum; runnable_load_sum 3368 kernel/sched/fair.c se->avg.runnable_load_sum = runnable_sum; runnable_load_sum 3373 kernel/sched/fair.c add_positive(&cfs_rq->avg.runnable_load_sum, delta_sum); runnable_load_sum 3547 kernel/sched/fair.c se->avg.runnable_load_sum = se->avg.load_sum; runnable_load_sum 7581 kernel/sched/fair.c if (cfs_rq->avg.runnable_load_sum) runnable_load_sum 124 kernel/sched/pelt.c sa->runnable_load_sum = runnable_load_sum 125 kernel/sched/pelt.c decay_load(sa->runnable_load_sum, periods); runnable_load_sum 140 kernel/sched/pelt.c sa->runnable_load_sum += runnable * contrib; runnable_load_sum 235 kernel/sched/pelt.c sa->runnable_load_avg = div_u64(runnable * sa->runnable_load_sum, divider);