sched_entity       72 drivers/gpu/drm/etnaviv/etnaviv_drv.c 			drm_sched_entity_init(&ctx->sched_entity[i],
sched_entity       96 drivers/gpu/drm/etnaviv/etnaviv_drv.c 			drm_sched_entity_destroy(&ctx->sched_entity[i]);
sched_entity       31 drivers/gpu/drm/etnaviv/etnaviv_drv.h 	struct drm_sched_entity		sched_entity[ETNA_MAX_PIPES];
sched_entity      584 drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c 	ret = etnaviv_sched_push_job(&ctx->sched_entity[args->pipe], submit);
sched_entity      143 drivers/gpu/drm/etnaviv/etnaviv_sched.c int etnaviv_sched_push_job(struct drm_sched_entity *sched_entity,
sched_entity      155 drivers/gpu/drm/etnaviv/etnaviv_sched.c 	ret = drm_sched_job_init(&submit->sched_job, sched_entity,
sched_entity      173 drivers/gpu/drm/etnaviv/etnaviv_sched.c 	drm_sched_entity_push_job(&submit->sched_job, sched_entity);
sched_entity       21 drivers/gpu/drm/etnaviv/etnaviv_sched.h int etnaviv_sched_push_job(struct drm_sched_entity *sched_entity,
sched_entity      113 drivers/gpu/drm/panfrost/panfrost_device.h 	struct drm_sched_entity sched_entity[NUM_JOB_SLOTS];
sched_entity      214 drivers/gpu/drm/panfrost/panfrost_job.c 	struct drm_sched_entity *entity = &job->file_priv->sched_entity[slot];
sched_entity      569 drivers/gpu/drm/panfrost/panfrost_job.c 		ret = drm_sched_entity_init(&panfrost_priv->sched_entity[i], &rq, 1, NULL);
sched_entity      581 drivers/gpu/drm/panfrost/panfrost_job.c 		drm_sched_entity_destroy(&panfrost_priv->sched_entity[i]);
sched_entity      151 drivers/gpu/drm/v3d/v3d_drv.c 		drm_sched_entity_init(&v3d_priv->sched_entity[i], &rq, 1, NULL);
sched_entity      166 drivers/gpu/drm/v3d/v3d_drv.c 		drm_sched_entity_destroy(&v3d_priv->sched_entity[q]);
sched_entity      137 drivers/gpu/drm/v3d/v3d_drv.h 	struct drm_sched_entity sched_entity[V3D_MAX_QUEUES];
sched_entity      471 drivers/gpu/drm/v3d/v3d_gem.c 	ret = drm_sched_job_init(&job->base, &v3d_priv->sched_entity[queue],
sched_entity      481 drivers/gpu/drm/v3d/v3d_gem.c 	drm_sched_entity_push_job(&job->base, &v3d_priv->sched_entity[queue]);
sched_entity      463 include/linux/sched.h 	struct sched_entity		*parent;
sched_entity      676 include/linux/sched.h 	struct sched_entity		se;
sched_entity      621 include/trace/events/sched.h 	TP_PROTO(struct sched_entity *se),
sched_entity     3529 kernel/sched/core.c 	struct sched_entity *curr = (&p->se)->cfs_rq->curr;
sched_entity     3531 kernel/sched/core.c 	struct sched_entity *curr = (&task_rq(p)->cfs)->curr;
sched_entity     6573 kernel/sched/core.c 		root_task_group.se = (struct sched_entity **)ptr;
sched_entity      377 kernel/sched/debug.c 	struct sched_entity *se = tg->se[cpu];
sched_entity      488 kernel/sched/debug.c 	struct sched_entity *last;
sched_entity      251 kernel/sched/fair.c static inline struct task_struct *task_of(struct sched_entity *se)
sched_entity      267 kernel/sched/fair.c static inline struct cfs_rq *cfs_rq_of(struct sched_entity *se)
sched_entity      273 kernel/sched/fair.c static inline struct cfs_rq *group_cfs_rq(struct sched_entity *grp)
sched_entity      391 kernel/sched/fair.c is_same_group(struct sched_entity *se, struct sched_entity *pse)
sched_entity      399 kernel/sched/fair.c static inline struct sched_entity *parent_entity(struct sched_entity *se)
sched_entity      405 kernel/sched/fair.c find_matching_se(struct sched_entity **se, struct sched_entity **pse)
sched_entity      438 kernel/sched/fair.c static inline struct task_struct *task_of(struct sched_entity *se)
sched_entity      451 kernel/sched/fair.c static inline struct cfs_rq *cfs_rq_of(struct sched_entity *se)
sched_entity      460 kernel/sched/fair.c static inline struct cfs_rq *group_cfs_rq(struct sched_entity *grp)
sched_entity      487 kernel/sched/fair.c static inline struct sched_entity *parent_entity(struct sched_entity *se)
sched_entity      493 kernel/sched/fair.c find_matching_se(struct sched_entity **se, struct sched_entity **pse)
sched_entity      524 kernel/sched/fair.c static inline int entity_before(struct sched_entity *a,
sched_entity      525 kernel/sched/fair.c 				struct sched_entity *b)
sched_entity      532 kernel/sched/fair.c 	struct sched_entity *curr = cfs_rq->curr;
sched_entity      545 kernel/sched/fair.c 		struct sched_entity *se;
sched_entity      546 kernel/sched/fair.c 		se = rb_entry(leftmost, struct sched_entity, run_node);
sched_entity      565 kernel/sched/fair.c static void __enqueue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity      569 kernel/sched/fair.c 	struct sched_entity *entry;
sched_entity      577 kernel/sched/fair.c 		entry = rb_entry(parent, struct sched_entity, run_node);
sched_entity      595 kernel/sched/fair.c static void __dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity      600 kernel/sched/fair.c struct sched_entity *__pick_first_entity(struct cfs_rq *cfs_rq)
sched_entity      607 kernel/sched/fair.c 	return rb_entry(left, struct sched_entity, run_node);
sched_entity      610 kernel/sched/fair.c static struct sched_entity *__pick_next_entity(struct sched_entity *se)
sched_entity      617 kernel/sched/fair.c 	return rb_entry(next, struct sched_entity, run_node);
sched_entity      621 kernel/sched/fair.c struct sched_entity *__pick_last_entity(struct cfs_rq *cfs_rq)
sched_entity      628 kernel/sched/fair.c 	return rb_entry(last, struct sched_entity, run_node);
sched_entity      662 kernel/sched/fair.c static inline u64 calc_delta_fair(u64 delta, struct sched_entity *se)
sched_entity      692 kernel/sched/fair.c static u64 sched_slice(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity      719 kernel/sched/fair.c static u64 sched_vslice(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity      732 kernel/sched/fair.c void init_entity_runnable_average(struct sched_entity *se)
sched_entity      752 kernel/sched/fair.c static void attach_entity_cfs_rq(struct sched_entity *se);
sched_entity      782 kernel/sched/fair.c 	struct sched_entity *se = &p->se;
sched_entity      819 kernel/sched/fair.c void init_entity_runnable_average(struct sched_entity *se)
sched_entity      835 kernel/sched/fair.c 	struct sched_entity *curr = cfs_rq->curr;
sched_entity      874 kernel/sched/fair.c update_stats_wait_start(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity      892 kernel/sched/fair.c update_stats_wait_end(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity      924 kernel/sched/fair.c update_stats_enqueue_sleeper(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity      995 kernel/sched/fair.c update_stats_enqueue(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
sched_entity     1012 kernel/sched/fair.c update_stats_dequeue(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
sched_entity     1041 kernel/sched/fair.c update_stats_curr_start(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity     2757 kernel/sched/fair.c account_entity_enqueue(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity     2772 kernel/sched/fair.c account_entity_dequeue(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity     2834 kernel/sched/fair.c enqueue_runnable_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity     2843 kernel/sched/fair.c dequeue_runnable_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity     2853 kernel/sched/fair.c enqueue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity     2860 kernel/sched/fair.c dequeue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity     2867 kernel/sched/fair.c enqueue_runnable_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) { }
sched_entity     2869 kernel/sched/fair.c dequeue_runnable_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) { }
sched_entity     2871 kernel/sched/fair.c enqueue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) { }
sched_entity     2873 kernel/sched/fair.c dequeue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) { }
sched_entity     2876 kernel/sched/fair.c static void reweight_entity(struct cfs_rq *cfs_rq, struct sched_entity *se,
sched_entity     2910 kernel/sched/fair.c 	struct sched_entity *se = &p->se;
sched_entity     3079 kernel/sched/fair.c static void update_cfs_group(struct sched_entity *se)
sched_entity     3104 kernel/sched/fair.c static inline void update_cfs_group(struct sched_entity *se)
sched_entity     3170 kernel/sched/fair.c void set_task_rq_fair(struct sched_entity *se,
sched_entity     3284 kernel/sched/fair.c update_tg_cfs_util(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq)
sched_entity     3310 kernel/sched/fair.c update_tg_cfs_runnable(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq)
sched_entity     3384 kernel/sched/fair.c static inline int propagate_entity_load_avg(struct sched_entity *se)
sched_entity     3414 kernel/sched/fair.c static inline bool skip_blocked_update(struct sched_entity *se)
sched_entity     3444 kernel/sched/fair.c static inline int propagate_entity_load_avg(struct sched_entity *se)
sched_entity     3519 kernel/sched/fair.c static void attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
sched_entity     3568 kernel/sched/fair.c static void detach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity     3589 kernel/sched/fair.c static inline void update_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
sched_entity     3649 kernel/sched/fair.c static void sync_entity_load_avg(struct sched_entity *se)
sched_entity     3662 kernel/sched/fair.c static void remove_entity_load_avg(struct sched_entity *se)
sched_entity     3836 kernel/sched/fair.c static inline void update_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se, int not_used1)
sched_entity     3841 kernel/sched/fair.c static inline void remove_entity_load_avg(struct sched_entity *se) {}
sched_entity     3844 kernel/sched/fair.c attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) {}
sched_entity     3846 kernel/sched/fair.c detach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) {}
sched_entity     3863 kernel/sched/fair.c static void check_spread(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity     3877 kernel/sched/fair.c place_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int initial)
sched_entity     3963 kernel/sched/fair.c enqueue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
sched_entity     4021 kernel/sched/fair.c static void __clear_buddies_last(struct sched_entity *se)
sched_entity     4032 kernel/sched/fair.c static void __clear_buddies_next(struct sched_entity *se)
sched_entity     4043 kernel/sched/fair.c static void __clear_buddies_skip(struct sched_entity *se)
sched_entity     4054 kernel/sched/fair.c static void clear_buddies(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity     4069 kernel/sched/fair.c dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
sched_entity     4124 kernel/sched/fair.c check_preempt_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr)
sched_entity     4127 kernel/sched/fair.c 	struct sched_entity *se;
sched_entity     4161 kernel/sched/fair.c set_next_entity(struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity     4194 kernel/sched/fair.c wakeup_preempt_entity(struct sched_entity *curr, struct sched_entity *se);
sched_entity     4203 kernel/sched/fair.c static struct sched_entity *
sched_entity     4204 kernel/sched/fair.c pick_next_entity(struct cfs_rq *cfs_rq, struct sched_entity *curr)
sched_entity     4206 kernel/sched/fair.c 	struct sched_entity *left = __pick_first_entity(cfs_rq);
sched_entity     4207 kernel/sched/fair.c 	struct sched_entity *se;
sched_entity     4223 kernel/sched/fair.c 		struct sched_entity *second;
sched_entity     4256 kernel/sched/fair.c static void put_prev_entity(struct cfs_rq *cfs_rq, struct sched_entity *prev)
sched_entity     4281 kernel/sched/fair.c entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr, int queued)
sched_entity     4502 kernel/sched/fair.c 	struct sched_entity *se;
sched_entity     4562 kernel/sched/fair.c 	struct sched_entity *se;
sched_entity     5137 kernel/sched/fair.c 	struct sched_entity *se = &p->se;
sched_entity     5210 kernel/sched/fair.c 	struct sched_entity *se = &p->se;
sched_entity     5308 kernel/sched/fair.c static void set_next_buddy(struct sched_entity *se);
sched_entity     5318 kernel/sched/fair.c 	struct sched_entity *se = &p->se;
sched_entity     6531 kernel/sched/fair.c static void detach_entity_cfs_rq(struct sched_entity *se);
sched_entity     6547 kernel/sched/fair.c 		struct sched_entity *se = &p->se;
sched_entity     6610 kernel/sched/fair.c static unsigned long wakeup_gran(struct sched_entity *se)
sched_entity     6645 kernel/sched/fair.c wakeup_preempt_entity(struct sched_entity *curr, struct sched_entity *se)
sched_entity     6659 kernel/sched/fair.c static void set_last_buddy(struct sched_entity *se)
sched_entity     6671 kernel/sched/fair.c static void set_next_buddy(struct sched_entity *se)
sched_entity     6683 kernel/sched/fair.c static void set_skip_buddy(struct sched_entity *se)
sched_entity     6695 kernel/sched/fair.c 	struct sched_entity *se = &curr->se, *pse = &p->se;
sched_entity     6779 kernel/sched/fair.c 	struct sched_entity *se;
sched_entity     6800 kernel/sched/fair.c 		struct sched_entity *curr = cfs_rq->curr;
sched_entity     6842 kernel/sched/fair.c 		struct sched_entity *pse = &prev->se;
sched_entity     6924 kernel/sched/fair.c 	struct sched_entity *se = &prev->se;
sched_entity     6942 kernel/sched/fair.c 	struct sched_entity *se = &curr->se;
sched_entity     6971 kernel/sched/fair.c 	struct sched_entity *se = &p->se;
sched_entity     7598 kernel/sched/fair.c 		struct sched_entity *se;
sched_entity     7635 kernel/sched/fair.c 	struct sched_entity *se = cfs_rq->tg->se[cpu_of(rq)];
sched_entity     9982 kernel/sched/fair.c 	struct sched_entity *se = &curr->se;
sched_entity     10004 kernel/sched/fair.c 	struct sched_entity *se = &p->se, *curr;
sched_entity     10056 kernel/sched/fair.c 	struct sched_entity *se = &p->se;
sched_entity     10087 kernel/sched/fair.c static void propagate_entity_cfs_rq(struct sched_entity *se)
sched_entity     10104 kernel/sched/fair.c static void propagate_entity_cfs_rq(struct sched_entity *se) { }
sched_entity     10107 kernel/sched/fair.c static void detach_entity_cfs_rq(struct sched_entity *se)
sched_entity     10118 kernel/sched/fair.c static void attach_entity_cfs_rq(struct sched_entity *se)
sched_entity     10139 kernel/sched/fair.c 	struct sched_entity *se = &p->se;
sched_entity     10156 kernel/sched/fair.c 	struct sched_entity *se = &p->se;
sched_entity     10194 kernel/sched/fair.c 	struct sched_entity *se = &p->se;
sched_entity     10230 kernel/sched/fair.c 	struct sched_entity *se = &p->se;
sched_entity     10280 kernel/sched/fair.c 	struct sched_entity *se;
sched_entity     10301 kernel/sched/fair.c 		se = kzalloc_node(sizeof(struct sched_entity),
sched_entity     10321 kernel/sched/fair.c 	struct sched_entity *se;
sched_entity     10363 kernel/sched/fair.c 			struct sched_entity *se, int cpu,
sched_entity     10364 kernel/sched/fair.c 			struct sched_entity *parent)
sched_entity     10414 kernel/sched/fair.c 		struct sched_entity *se = tg->se[i];
sched_entity     10449 kernel/sched/fair.c 	struct sched_entity *se = &task->se;
sched_entity      266 kernel/sched/pelt.c int __update_load_avg_blocked_se(u64 now, struct sched_entity *se)
sched_entity      277 kernel/sched/pelt.c int __update_load_avg_se(u64 now, struct cfs_rq *cfs_rq, struct sched_entity *se)
sched_entity        4 kernel/sched/pelt.h int __update_load_avg_blocked_se(u64 now, struct sched_entity *se);
sched_entity        5 kernel/sched/pelt.h int __update_load_avg_se(u64 now, struct cfs_rq *cfs_rq, struct sched_entity *se);
sched_entity      366 kernel/sched/sched.h 	struct sched_entity	**se;
sched_entity      450 kernel/sched/sched.h 			struct sched_entity *se, int cpu,
sched_entity      451 kernel/sched/sched.h 			struct sched_entity *parent);
sched_entity      481 kernel/sched/sched.h extern void set_task_rq_fair(struct sched_entity *se,
sched_entity      484 kernel/sched/sched.h static inline void set_task_rq_fair(struct sched_entity *se,
sched_entity      515 kernel/sched/sched.h 	struct sched_entity	*curr;
sched_entity      516 kernel/sched/sched.h 	struct sched_entity	*next;
sched_entity      517 kernel/sched/sched.h 	struct sched_entity	*last;
sched_entity      518 kernel/sched/sched.h 	struct sched_entity	*skip;
sched_entity      553 kernel/sched/sched.h 	struct sched_entity	*h_load_next;
sched_entity      705 kernel/sched/sched.h static inline long se_weight(struct sched_entity *se)
sched_entity      710 kernel/sched/sched.h static inline long se_runnable(struct sched_entity *se)
sched_entity     1894 kernel/sched/sched.h extern void init_entity_runnable_average(struct sched_entity *se);
sched_entity     2178 kernel/sched/sched.h extern struct sched_entity *__pick_first_entity(struct cfs_rq *cfs_rq);
sched_entity     2179 kernel/sched/sched.h extern struct sched_entity *__pick_last_entity(struct cfs_rq *cfs_rq);