sched_avg 477 include/linux/sched.h struct sched_avg avg; sched_avg 1988 include/linux/sched.h const struct sched_avg *sched_trace_cfs_rq_avg(struct cfs_rq *cfs_rq); sched_avg 1992 include/linux/sched.h const struct sched_avg *sched_trace_rq_avg_rt(struct rq *rq); sched_avg 1993 include/linux/sched.h const struct sched_avg *sched_trace_rq_avg_dl(struct rq *rq); sched_avg 1994 include/linux/sched.h const struct sched_avg *sched_trace_rq_avg_irq(struct rq *rq); sched_avg 734 kernel/sched/fair.c struct sched_avg *sa = &se->avg; sched_avg 784 kernel/sched/fair.c struct sched_avg *sa = &se->avg; sched_avg 3473 kernel/sched/fair.c struct sched_avg *sa = &cfs_rq->avg; sched_avg 10564 kernel/sched/fair.c const struct sched_avg *sched_trace_cfs_rq_avg(struct cfs_rq *cfs_rq) sched_avg 10594 kernel/sched/fair.c const struct sched_avg *sched_trace_rq_avg_rt(struct rq *rq) sched_avg 10604 kernel/sched/fair.c const struct sched_avg *sched_trace_rq_avg_dl(struct rq *rq) sched_avg 10614 kernel/sched/fair.c const struct sched_avg *sched_trace_rq_avg_irq(struct rq *rq) sched_avg 110 kernel/sched/pelt.c accumulate_sum(u64 delta, struct sched_avg *sa, sched_avg 176 kernel/sched/pelt.c ___update_load_sum(u64 now, struct sched_avg *sa, sched_avg 227 kernel/sched/pelt.c ___update_load_avg(struct sched_avg *sa, unsigned long load, unsigned long runnable) sched_avg 29 kernel/sched/pelt.h static inline void cfs_se_util_change(struct sched_avg *avg) sched_avg 528 kernel/sched/sched.h struct sched_avg avg; sched_avg 948 kernel/sched/sched.h struct sched_avg avg_rt; sched_avg 949 kernel/sched/sched.h struct sched_avg avg_dl; sched_avg 951 kernel/sched/sched.h struct sched_avg avg_irq;