runnable_load_sum  399 include/linux/sched.h 	u64				runnable_load_sum;
runnable_load_sum  946 kernel/sched/debug.c 	P(se.avg.runnable_load_sum);
runnable_load_sum 2839 kernel/sched/fair.c 	cfs_rq->avg.runnable_load_sum += se_runnable(se) * se->avg.runnable_load_sum;
runnable_load_sum 2848 kernel/sched/fair.c 	sub_positive(&cfs_rq->avg.runnable_load_sum,
runnable_load_sum 2849 kernel/sched/fair.c 		     se_runnable(se) * se->avg.runnable_load_sum);
runnable_load_sum 2897 kernel/sched/fair.c 			div_u64(se_runnable(se) * se->avg.runnable_load_sum, divider);
runnable_load_sum 3314 kernel/sched/fair.c 	u64 runnable_load_sum, load_sum = 0;
runnable_load_sum 3363 kernel/sched/fair.c 	runnable_load_sum = (s64)se_runnable(se) * runnable_sum;
runnable_load_sum 3364 kernel/sched/fair.c 	runnable_load_avg = div_s64(runnable_load_sum, LOAD_AVG_MAX);
runnable_load_sum 3365 kernel/sched/fair.c 	delta_sum = runnable_load_sum - se_weight(se) * se->avg.runnable_load_sum;
runnable_load_sum 3368 kernel/sched/fair.c 	se->avg.runnable_load_sum = runnable_sum;
runnable_load_sum 3373 kernel/sched/fair.c 		add_positive(&cfs_rq->avg.runnable_load_sum, delta_sum);
runnable_load_sum 3547 kernel/sched/fair.c 	se->avg.runnable_load_sum = se->avg.load_sum;
runnable_load_sum 7581 kernel/sched/fair.c 	if (cfs_rq->avg.runnable_load_sum)
runnable_load_sum  124 kernel/sched/pelt.c 		sa->runnable_load_sum =
runnable_load_sum  125 kernel/sched/pelt.c 			decay_load(sa->runnable_load_sum, periods);
runnable_load_sum  140 kernel/sched/pelt.c 		sa->runnable_load_sum += runnable * contrib;
runnable_load_sum  235 kernel/sched/pelt.c 	sa->runnable_load_avg =	div_u64(runnable * sa->runnable_load_sum, divider);