Lines Matching refs:idle

3417 			cfs_b->idle = 0;  in assign_cfs_rq_runtime()
3652 if (rq->curr == rq->idle && rq->cfs.nr_running) in unthrottle_cfs_rq()
3717 if (cfs_b->idle && !throttled) in do_sched_cfs_period_timer()
3724 cfs_b->idle = 1; in do_sched_cfs_period_timer()
3759 cfs_b->idle = 0; in do_sched_cfs_period_timer()
3939 int idle = 0; in sched_cfs_period_timer() local
3947 idle = do_sched_cfs_period_timer(cfs_b, overrun); in sched_cfs_period_timer()
3949 if (idle) in sched_cfs_period_timer()
3953 return idle ? HRTIMER_NORESTART : HRTIMER_RESTART; in sched_cfs_period_timer()
4802 struct cpuidle_state *idle = idle_get_state(rq); in find_idlest_cpu() local
4803 if (idle && idle->exit_latency < min_exit_latency) { in find_idlest_cpu()
4809 min_exit_latency = idle->exit_latency; in find_idlest_cpu()
4812 } else if ((!idle || idle->exit_latency == min_exit_latency) && in find_idlest_cpu()
5190 if (unlikely(!se->on_rq || curr == rq->idle)) in check_preempt_wakeup()
5208 goto idle; in pick_next_task_fair()
5287 goto idle; in pick_next_task_fair()
5304 idle: in pick_next_task_fair()
5533 enum cpu_idle_type idle; member
5707 schedstat_inc(env->sd, lb_hot_gained[env->idle]); in can_migrate_task()
5753 schedstat_inc(env->sd, lb_gained[env->idle]); in detach_one_task()
5784 if (env->idle != CPU_NOT_IDLE && env->src_rq->nr_running <= 1) in detach_tasks()
5824 if (env->idle == CPU_NEWLY_IDLE) in detach_tasks()
5845 schedstat_add(env->sd, lb_gained[env->idle], detached); in detach_tasks()
6058 enum cpu_idle_type idle) in get_sd_load_idx() argument
6062 switch (idle) { in get_sd_load_idx()
6442 load_idx = get_sd_load_idx(env->sd, env->idle); in update_sd_lb_stats()
6453 if (env->idle != CPU_NEWLY_IDLE || in update_sd_lb_stats()
6727 if ((env->idle == CPU_IDLE || env->idle == CPU_NEWLY_IDLE) && in find_busiest_group()
6747 if (env->idle == CPU_NEWLY_IDLE && group_has_capacity(env, local) && in find_busiest_group()
6765 if (env->idle == CPU_IDLE) { in find_busiest_group()
6882 if (env->idle == CPU_NEWLY_IDLE) { in need_active_balance()
6899 if ((env->idle != CPU_NOT_IDLE) && in need_active_balance()
6921 if (env->idle == CPU_NEWLY_IDLE) in should_we_balance()
6950 struct sched_domain *sd, enum cpu_idle_type idle, in load_balance() argument
6965 .idle = idle, in load_balance()
6976 if (idle == CPU_NEWLY_IDLE) in load_balance()
6981 schedstat_inc(sd, lb_count[idle]); in load_balance()
6991 schedstat_inc(sd, lb_nobusyg[idle]); in load_balance()
6997 schedstat_inc(sd, lb_nobusyq[idle]); in load_balance()
7003 schedstat_add(sd, lb_imbalance[idle], env.imbalance); in load_balance()
7110 schedstat_inc(sd, lb_failed[idle]); in load_balance()
7117 if (idle != CPU_NEWLY_IDLE) in load_balance()
7196 schedstat_inc(sd, lb_balanced[idle]); in load_balance()
7389 .idle = CPU_IDLE, in active_load_balance_cpu_stop()
7573 static void rebalance_domains(struct rq *rq, enum cpu_idle_type idle) in rebalance_domains() argument
7615 interval = get_sd_balance_interval(sd, idle != CPU_IDLE); in rebalance_domains()
7624 if (load_balance(cpu, rq, sd, idle, &continue_balancing)) { in rebalance_domains()
7630 idle = idle_cpu(cpu) ? CPU_IDLE : CPU_NOT_IDLE; in rebalance_domains()
7633 interval = get_sd_balance_interval(sd, idle != CPU_IDLE); in rebalance_domains()
7670 if ((idle == CPU_IDLE) && time_after(nohz.next_balance, rq->next_balance)) in rebalance_domains()
7681 static void nohz_idle_balance(struct rq *this_rq, enum cpu_idle_type idle) in nohz_idle_balance() argument
7690 if (idle != CPU_IDLE || in nohz_idle_balance()
7813 static void nohz_idle_balance(struct rq *this_rq, enum cpu_idle_type idle) { } in nohz_idle_balance() argument
7823 enum cpu_idle_type idle = this_rq->idle_balance ? in run_rebalance_domains()
7834 nohz_idle_balance(this_rq, idle); in run_rebalance_domains()
7835 rebalance_domains(this_rq, idle); in run_rebalance_domains()