sched_avg         477 include/linux/sched.h 	struct sched_avg		avg;
sched_avg        1988 include/linux/sched.h const struct sched_avg *sched_trace_cfs_rq_avg(struct cfs_rq *cfs_rq);
sched_avg        1992 include/linux/sched.h const struct sched_avg *sched_trace_rq_avg_rt(struct rq *rq);
sched_avg        1993 include/linux/sched.h const struct sched_avg *sched_trace_rq_avg_dl(struct rq *rq);
sched_avg        1994 include/linux/sched.h const struct sched_avg *sched_trace_rq_avg_irq(struct rq *rq);
sched_avg         734 kernel/sched/fair.c 	struct sched_avg *sa = &se->avg;
sched_avg         784 kernel/sched/fair.c 	struct sched_avg *sa = &se->avg;
sched_avg        3473 kernel/sched/fair.c 	struct sched_avg *sa = &cfs_rq->avg;
sched_avg        10564 kernel/sched/fair.c const struct sched_avg *sched_trace_cfs_rq_avg(struct cfs_rq *cfs_rq)
sched_avg        10594 kernel/sched/fair.c const struct sched_avg *sched_trace_rq_avg_rt(struct rq *rq)
sched_avg        10604 kernel/sched/fair.c const struct sched_avg *sched_trace_rq_avg_dl(struct rq *rq)
sched_avg        10614 kernel/sched/fair.c const struct sched_avg *sched_trace_rq_avg_irq(struct rq *rq)
sched_avg         110 kernel/sched/pelt.c accumulate_sum(u64 delta, struct sched_avg *sa,
sched_avg         176 kernel/sched/pelt.c ___update_load_sum(u64 now, struct sched_avg *sa,
sched_avg         227 kernel/sched/pelt.c ___update_load_avg(struct sched_avg *sa, unsigned long load, unsigned long runnable)
sched_avg          29 kernel/sched/pelt.h static inline void cfs_se_util_change(struct sched_avg *avg)
sched_avg         528 kernel/sched/sched.h 	struct sched_avg	avg;
sched_avg         948 kernel/sched/sched.h 	struct sched_avg	avg_rt;
sched_avg         949 kernel/sched/sched.h 	struct sched_avg	avg_dl;
sched_avg         951 kernel/sched/sched.h 	struct sched_avg	avg_irq;