dst_rq 1758 kernel/sched/core.c struct rq *src_rq, *dst_rq; dst_rq 1762 kernel/sched/core.c dst_rq = cpu_rq(cpu); dst_rq 1765 kernel/sched/core.c rq_pin_lock(dst_rq, &drf); dst_rq 1769 kernel/sched/core.c activate_task(dst_rq, p, 0); dst_rq 1770 kernel/sched/core.c check_preempt_curr(dst_rq, p, 0); dst_rq 1772 kernel/sched/core.c rq_unpin_lock(dst_rq, &drf); dst_rq 1793 kernel/sched/core.c struct rq *src_rq, *dst_rq; dst_rq 1800 kernel/sched/core.c dst_rq = cpu_rq(arg->dst_cpu); dst_rq 1804 kernel/sched/core.c double_rq_lock(src_rq, dst_rq); dst_rq 1824 kernel/sched/core.c double_rq_unlock(src_rq, dst_rq); dst_rq 1593 kernel/sched/fair.c struct rq *dst_rq = cpu_rq(env->dst_cpu); dst_rq 1601 kernel/sched/fair.c if (READ_ONCE(dst_rq->numa_migrate_on)) dst_rq 1605 kernel/sched/fair.c cur = rcu_dereference(dst_rq->curr); dst_rq 7129 kernel/sched/fair.c struct rq *dst_rq; dst_rq 7167 kernel/sched/fair.c if (sched_feat(CACHE_HOT_BUDDY) && env->dst_rq->nr_running && dst_rq 7492 kernel/sched/fair.c rq_lock(env->dst_rq, &rf); dst_rq 7493 kernel/sched/fair.c update_rq_clock(env->dst_rq); dst_rq 7499 kernel/sched/fair.c attach_task(env->dst_rq, p); dst_rq 7502 kernel/sched/fair.c rq_unlock(env->dst_rq, &rf); dst_rq 8307 kernel/sched/fair.c struct root_domain *rd = env->dst_rq->rd; dst_rq 8316 kernel/sched/fair.c struct root_domain *rd = env->dst_rq->rd; dst_rq 8546 kernel/sched/fair.c struct root_domain *rd = env->dst_rq->rd; dst_rq 8843 kernel/sched/fair.c .dst_rq = this_rq, dst_rq 8874 kernel/sched/fair.c BUG_ON(busiest == env.dst_rq); dst_rq 8948 kernel/sched/fair.c env.dst_rq = cpu_rq(env.new_dst_cpu); dst_rq 9184 kernel/sched/fair.c .dst_rq = target_rq,