Lines Matching refs:idle

3493 			cfs_b->idle = 0;  in assign_cfs_rq_runtime()
3719 if (rq->curr == rq->idle && rq->cfs.nr_running) in unthrottle_cfs_rq()
3784 if (cfs_b->idle && !throttled) in do_sched_cfs_period_timer()
3798 cfs_b->idle = 1; in do_sched_cfs_period_timer()
3833 cfs_b->idle = 0; in do_sched_cfs_period_timer()
4013 int idle = 0; in sched_cfs_period_timer() local
4023 idle = do_sched_cfs_period_timer(cfs_b, overrun); in sched_cfs_period_timer()
4027 return idle ? HRTIMER_NORESTART : HRTIMER_RESTART; in sched_cfs_period_timer()
4712 struct cpuidle_state *idle = idle_get_state(rq); in find_idlest_cpu() local
4713 if (idle && idle->exit_latency < min_exit_latency) { in find_idlest_cpu()
4719 min_exit_latency = idle->exit_latency; in find_idlest_cpu()
4722 } else if ((!idle || idle->exit_latency == min_exit_latency) && in find_idlest_cpu()
5089 if (unlikely(!se->on_rq || curr == rq->idle)) in check_preempt_wakeup()
5107 goto idle; in pick_next_task_fair()
5186 goto idle; in pick_next_task_fair()
5203 idle: in pick_next_task_fair()
5424 enum cpu_idle_type idle; member
5631 schedstat_inc(env->sd, lb_hot_gained[env->idle]); in can_migrate_task()
5677 schedstat_inc(env->sd, lb_gained[env->idle]); in detach_one_task()
5741 if (env->idle == CPU_NEWLY_IDLE) in detach_tasks()
5762 schedstat_add(env->sd, lb_gained[env->idle], detached); in detach_tasks()
6000 enum cpu_idle_type idle) in get_sd_load_idx() argument
6004 switch (idle) { in get_sd_load_idx()
6404 load_idx = get_sd_load_idx(env->sd, env->idle); in update_sd_lb_stats()
6415 if (env->idle != CPU_NEWLY_IDLE || in update_sd_lb_stats()
6689 if ((env->idle == CPU_IDLE || env->idle == CPU_NEWLY_IDLE) && in find_busiest_group()
6709 if (env->idle == CPU_NEWLY_IDLE && group_has_capacity(env, local) && in find_busiest_group()
6727 if (env->idle == CPU_IDLE) { in find_busiest_group()
6844 if (env->idle == CPU_NEWLY_IDLE) { in need_active_balance()
6861 if ((env->idle != CPU_NOT_IDLE) && in need_active_balance()
6883 if (env->idle == CPU_NEWLY_IDLE) in should_we_balance()
6912 struct sched_domain *sd, enum cpu_idle_type idle, in load_balance() argument
6927 .idle = idle, in load_balance()
6938 if (idle == CPU_NEWLY_IDLE) in load_balance()
6943 schedstat_inc(sd, lb_count[idle]); in load_balance()
6953 schedstat_inc(sd, lb_nobusyg[idle]); in load_balance()
6959 schedstat_inc(sd, lb_nobusyq[idle]); in load_balance()
6965 schedstat_add(sd, lb_imbalance[idle], env.imbalance); in load_balance()
7072 schedstat_inc(sd, lb_failed[idle]); in load_balance()
7079 if (idle != CPU_NEWLY_IDLE) in load_balance()
7158 schedstat_inc(sd, lb_balanced[idle]); in load_balance()
7354 .idle = CPU_IDLE, in active_load_balance_cpu_stop()
7538 static void rebalance_domains(struct rq *rq, enum cpu_idle_type idle) in rebalance_domains() argument
7580 interval = get_sd_balance_interval(sd, idle != CPU_IDLE); in rebalance_domains()
7589 if (load_balance(cpu, rq, sd, idle, &continue_balancing)) { in rebalance_domains()
7595 idle = idle_cpu(cpu) ? CPU_IDLE : CPU_NOT_IDLE; in rebalance_domains()
7598 interval = get_sd_balance_interval(sd, idle != CPU_IDLE); in rebalance_domains()
7632 static void nohz_idle_balance(struct rq *this_rq, enum cpu_idle_type idle) in nohz_idle_balance() argument
7638 if (idle != CPU_IDLE || in nohz_idle_balance()
7752 static void nohz_idle_balance(struct rq *this_rq, enum cpu_idle_type idle) { } in nohz_idle_balance() argument
7762 enum cpu_idle_type idle = this_rq->idle_balance ? in run_rebalance_domains()
7773 nohz_idle_balance(this_rq, idle); in run_rebalance_domains()
7774 rebalance_domains(this_rq, idle); in run_rebalance_domains()