Lines Matching refs:rt_se

113 #define rt_entity_is_task(rt_se) (!(rt_se)->my_q)  argument
115 static inline struct task_struct *rt_task_of(struct sched_rt_entity *rt_se) in rt_task_of() argument
118 WARN_ON_ONCE(!rt_entity_is_task(rt_se)); in rt_task_of()
120 return container_of(rt_se, struct task_struct, rt); in rt_task_of()
128 static inline struct rt_rq *rt_rq_of_se(struct sched_rt_entity *rt_se) in rt_rq_of_se() argument
130 return rt_se->rt_rq; in rt_rq_of_se()
133 static inline struct rq *rq_of_rt_se(struct sched_rt_entity *rt_se) in rq_of_rt_se() argument
135 struct rt_rq *rt_rq = rt_se->rt_rq; in rq_of_rt_se()
144 if (tg->rt_se) in free_rt_sched_group()
150 if (tg->rt_se) in free_rt_sched_group()
151 kfree(tg->rt_se[i]); in free_rt_sched_group()
155 kfree(tg->rt_se); in free_rt_sched_group()
159 struct sched_rt_entity *rt_se, int cpu, in init_tg_rt_entry() argument
170 tg->rt_se[cpu] = rt_se; in init_tg_rt_entry()
172 if (!rt_se) in init_tg_rt_entry()
176 rt_se->rt_rq = &rq->rt; in init_tg_rt_entry()
178 rt_se->rt_rq = parent->my_q; in init_tg_rt_entry()
180 rt_se->my_q = rt_rq; in init_tg_rt_entry()
181 rt_se->parent = parent; in init_tg_rt_entry()
182 INIT_LIST_HEAD(&rt_se->run_list); in init_tg_rt_entry()
188 struct sched_rt_entity *rt_se; in alloc_rt_sched_group() local
194 tg->rt_se = kzalloc(sizeof(rt_se) * nr_cpu_ids, GFP_KERNEL); in alloc_rt_sched_group()
195 if (!tg->rt_se) in alloc_rt_sched_group()
207 rt_se = kzalloc_node(sizeof(struct sched_rt_entity), in alloc_rt_sched_group()
209 if (!rt_se) in alloc_rt_sched_group()
214 init_tg_rt_entry(tg, rt_rq, rt_se, i, parent->rt_se[i]); in alloc_rt_sched_group()
227 #define rt_entity_is_task(rt_se) (1) argument
229 static inline struct task_struct *rt_task_of(struct sched_rt_entity *rt_se) in rt_task_of() argument
231 return container_of(rt_se, struct task_struct, rt); in rt_task_of()
239 static inline struct rq *rq_of_rt_se(struct sched_rt_entity *rt_se) in rq_of_rt_se() argument
241 struct task_struct *p = rt_task_of(rt_se); in rq_of_rt_se()
246 static inline struct rt_rq *rt_rq_of_se(struct sched_rt_entity *rt_se) in rt_rq_of_se() argument
248 struct rq *rq = rq_of_rt_se(rt_se); in rt_rq_of_se()
318 static void inc_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in inc_rt_migration() argument
322 if (!rt_entity_is_task(rt_se)) in inc_rt_migration()
325 p = rt_task_of(rt_se); in inc_rt_migration()
335 static void dec_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in dec_rt_migration() argument
339 if (!rt_entity_is_task(rt_se)) in dec_rt_migration()
342 p = rt_task_of(rt_se); in dec_rt_migration()
411 void inc_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in inc_rt_migration() argument
416 void dec_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in dec_rt_migration() argument
437 static inline int on_rt_rq(struct sched_rt_entity *rt_se) in on_rt_rq() argument
439 return !list_empty(&rt_se->run_list); in on_rt_rq()
477 #define for_each_sched_rt_entity(rt_se) \ argument
478 for (; rt_se; rt_se = rt_se->parent)
480 static inline struct rt_rq *group_rt_rq(struct sched_rt_entity *rt_se) in group_rt_rq() argument
482 return rt_se->my_q; in group_rt_rq()
485 static void enqueue_rt_entity(struct sched_rt_entity *rt_se, bool head);
486 static void dequeue_rt_entity(struct sched_rt_entity *rt_se);
492 struct sched_rt_entity *rt_se; in sched_rt_rq_enqueue() local
496 rt_se = rt_rq->tg->rt_se[cpu]; in sched_rt_rq_enqueue()
499 if (!rt_se) in sched_rt_rq_enqueue()
501 else if (!on_rt_rq(rt_se)) in sched_rt_rq_enqueue()
502 enqueue_rt_entity(rt_se, false); in sched_rt_rq_enqueue()
511 struct sched_rt_entity *rt_se; in sched_rt_rq_dequeue() local
514 rt_se = rt_rq->tg->rt_se[cpu]; in sched_rt_rq_dequeue()
516 if (!rt_se) in sched_rt_rq_dequeue()
518 else if (on_rt_rq(rt_se)) in sched_rt_rq_dequeue()
519 dequeue_rt_entity(rt_se); in sched_rt_rq_dequeue()
527 static int rt_se_boosted(struct sched_rt_entity *rt_se) in rt_se_boosted() argument
529 struct rt_rq *rt_rq = group_rt_rq(rt_se); in rt_se_boosted()
535 p = rt_task_of(rt_se); in rt_se_boosted()
579 #define for_each_sched_rt_entity(rt_se) \ argument
580 for (; rt_se; rt_se = NULL)
582 static inline struct rt_rq *group_rt_rq(struct sched_rt_entity *rt_se) in group_rt_rq() argument
880 static inline int rt_se_prio(struct sched_rt_entity *rt_se) in rt_se_prio() argument
883 struct rt_rq *rt_rq = group_rt_rq(rt_se); in rt_se_prio()
889 return rt_task_of(rt_se)->prio; in rt_se_prio()
942 struct sched_rt_entity *rt_se = &curr->rt; in update_curr_rt() local
966 for_each_sched_rt_entity(rt_se) { in update_curr_rt()
967 struct rt_rq *rt_rq = rt_rq_of_se(rt_se); in update_curr_rt()
1102 inc_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in inc_rt_group() argument
1104 if (rt_se_boosted(rt_se)) in inc_rt_group()
1112 dec_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in dec_rt_group() argument
1114 if (rt_se_boosted(rt_se)) in dec_rt_group()
1123 inc_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in inc_rt_group() argument
1129 void dec_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) {} in dec_rt_group() argument
1134 unsigned int rt_se_nr_running(struct sched_rt_entity *rt_se) in rt_se_nr_running() argument
1136 struct rt_rq *group_rq = group_rt_rq(rt_se); in rt_se_nr_running()
1145 void inc_rt_tasks(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in inc_rt_tasks() argument
1147 int prio = rt_se_prio(rt_se); in inc_rt_tasks()
1150 rt_rq->rt_nr_running += rt_se_nr_running(rt_se); in inc_rt_tasks()
1153 inc_rt_migration(rt_se, rt_rq); in inc_rt_tasks()
1154 inc_rt_group(rt_se, rt_rq); in inc_rt_tasks()
1158 void dec_rt_tasks(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) in dec_rt_tasks() argument
1160 WARN_ON(!rt_prio(rt_se_prio(rt_se))); in dec_rt_tasks()
1162 rt_rq->rt_nr_running -= rt_se_nr_running(rt_se); in dec_rt_tasks()
1164 dec_rt_prio(rt_rq, rt_se_prio(rt_se)); in dec_rt_tasks()
1165 dec_rt_migration(rt_se, rt_rq); in dec_rt_tasks()
1166 dec_rt_group(rt_se, rt_rq); in dec_rt_tasks()
1169 static void __enqueue_rt_entity(struct sched_rt_entity *rt_se, bool head) in __enqueue_rt_entity() argument
1171 struct rt_rq *rt_rq = rt_rq_of_se(rt_se); in __enqueue_rt_entity()
1173 struct rt_rq *group_rq = group_rt_rq(rt_se); in __enqueue_rt_entity()
1174 struct list_head *queue = array->queue + rt_se_prio(rt_se); in __enqueue_rt_entity()
1186 list_add(&rt_se->run_list, queue); in __enqueue_rt_entity()
1188 list_add_tail(&rt_se->run_list, queue); in __enqueue_rt_entity()
1189 __set_bit(rt_se_prio(rt_se), array->bitmap); in __enqueue_rt_entity()
1191 inc_rt_tasks(rt_se, rt_rq); in __enqueue_rt_entity()
1194 static void __dequeue_rt_entity(struct sched_rt_entity *rt_se) in __dequeue_rt_entity() argument
1196 struct rt_rq *rt_rq = rt_rq_of_se(rt_se); in __dequeue_rt_entity()
1199 list_del_init(&rt_se->run_list); in __dequeue_rt_entity()
1200 if (list_empty(array->queue + rt_se_prio(rt_se))) in __dequeue_rt_entity()
1201 __clear_bit(rt_se_prio(rt_se), array->bitmap); in __dequeue_rt_entity()
1203 dec_rt_tasks(rt_se, rt_rq); in __dequeue_rt_entity()
1210 static void dequeue_rt_stack(struct sched_rt_entity *rt_se) in dequeue_rt_stack() argument
1214 for_each_sched_rt_entity(rt_se) { in dequeue_rt_stack()
1215 rt_se->back = back; in dequeue_rt_stack()
1216 back = rt_se; in dequeue_rt_stack()
1221 for (rt_se = back; rt_se; rt_se = rt_se->back) { in dequeue_rt_stack()
1222 if (on_rt_rq(rt_se)) in dequeue_rt_stack()
1223 __dequeue_rt_entity(rt_se); in dequeue_rt_stack()
1227 static void enqueue_rt_entity(struct sched_rt_entity *rt_se, bool head) in enqueue_rt_entity() argument
1229 struct rq *rq = rq_of_rt_se(rt_se); in enqueue_rt_entity()
1231 dequeue_rt_stack(rt_se); in enqueue_rt_entity()
1232 for_each_sched_rt_entity(rt_se) in enqueue_rt_entity()
1233 __enqueue_rt_entity(rt_se, head); in enqueue_rt_entity()
1237 static void dequeue_rt_entity(struct sched_rt_entity *rt_se) in dequeue_rt_entity() argument
1239 struct rq *rq = rq_of_rt_se(rt_se); in dequeue_rt_entity()
1241 dequeue_rt_stack(rt_se); in dequeue_rt_entity()
1243 for_each_sched_rt_entity(rt_se) { in dequeue_rt_entity()
1244 struct rt_rq *rt_rq = group_rt_rq(rt_se); in dequeue_rt_entity()
1247 __enqueue_rt_entity(rt_se, false); in dequeue_rt_entity()
1258 struct sched_rt_entity *rt_se = &p->rt; in enqueue_task_rt() local
1261 rt_se->timeout = 0; in enqueue_task_rt()
1263 enqueue_rt_entity(rt_se, flags & ENQUEUE_HEAD); in enqueue_task_rt()
1271 struct sched_rt_entity *rt_se = &p->rt; in dequeue_task_rt() local
1274 dequeue_rt_entity(rt_se); in dequeue_task_rt()
1284 requeue_rt_entity(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se, int head) in requeue_rt_entity() argument
1286 if (on_rt_rq(rt_se)) { in requeue_rt_entity()
1288 struct list_head *queue = array->queue + rt_se_prio(rt_se); in requeue_rt_entity()
1291 list_move(&rt_se->run_list, queue); in requeue_rt_entity()
1293 list_move_tail(&rt_se->run_list, queue); in requeue_rt_entity()
1299 struct sched_rt_entity *rt_se = &p->rt; in requeue_task_rt() local
1302 for_each_sched_rt_entity(rt_se) { in requeue_task_rt()
1303 rt_rq = rt_rq_of_se(rt_se); in requeue_task_rt()
1304 requeue_rt_entity(rt_rq, rt_se, head); in requeue_task_rt()
1448 struct sched_rt_entity *rt_se; in _pick_next_task_rt() local
1453 rt_se = pick_next_rt_entity(rq, rt_rq); in _pick_next_task_rt()
1454 BUG_ON(!rt_se); in _pick_next_task_rt()
1455 rt_rq = group_rt_rq(rt_se); in _pick_next_task_rt()
1458 p = rt_task_of(rt_se); in _pick_next_task_rt()
2211 struct sched_rt_entity *rt_se = &p->rt; in task_tick_rt() local
2233 for_each_sched_rt_entity(rt_se) { in task_tick_rt()
2234 if (rt_se->run_list.prev != rt_se->run_list.next) { in task_tick_rt()