dst_rq           1758 kernel/sched/core.c 		struct rq *src_rq, *dst_rq;
dst_rq           1762 kernel/sched/core.c 		dst_rq = cpu_rq(cpu);
dst_rq           1765 kernel/sched/core.c 		rq_pin_lock(dst_rq, &drf);
dst_rq           1769 kernel/sched/core.c 		activate_task(dst_rq, p, 0);
dst_rq           1770 kernel/sched/core.c 		check_preempt_curr(dst_rq, p, 0);
dst_rq           1772 kernel/sched/core.c 		rq_unpin_lock(dst_rq, &drf);
dst_rq           1793 kernel/sched/core.c 	struct rq *src_rq, *dst_rq;
dst_rq           1800 kernel/sched/core.c 	dst_rq = cpu_rq(arg->dst_cpu);
dst_rq           1804 kernel/sched/core.c 	double_rq_lock(src_rq, dst_rq);
dst_rq           1824 kernel/sched/core.c 	double_rq_unlock(src_rq, dst_rq);
dst_rq           1593 kernel/sched/fair.c 	struct rq *dst_rq = cpu_rq(env->dst_cpu);
dst_rq           1601 kernel/sched/fair.c 	if (READ_ONCE(dst_rq->numa_migrate_on))
dst_rq           1605 kernel/sched/fair.c 	cur = rcu_dereference(dst_rq->curr);
dst_rq           7129 kernel/sched/fair.c 	struct rq		*dst_rq;
dst_rq           7167 kernel/sched/fair.c 	if (sched_feat(CACHE_HOT_BUDDY) && env->dst_rq->nr_running &&
dst_rq           7492 kernel/sched/fair.c 	rq_lock(env->dst_rq, &rf);
dst_rq           7493 kernel/sched/fair.c 	update_rq_clock(env->dst_rq);
dst_rq           7499 kernel/sched/fair.c 		attach_task(env->dst_rq, p);
dst_rq           7502 kernel/sched/fair.c 	rq_unlock(env->dst_rq, &rf);
dst_rq           8307 kernel/sched/fair.c 		struct root_domain *rd = env->dst_rq->rd;
dst_rq           8316 kernel/sched/fair.c 		struct root_domain *rd = env->dst_rq->rd;
dst_rq           8546 kernel/sched/fair.c 		struct root_domain *rd = env->dst_rq->rd;
dst_rq           8843 kernel/sched/fair.c 		.dst_rq		= this_rq,
dst_rq           8874 kernel/sched/fair.c 	BUG_ON(busiest == env.dst_rq);
dst_rq           8948 kernel/sched/fair.c 			env.dst_rq	 = cpu_rq(env.new_dst_cpu);
dst_rq           9184 kernel/sched/fair.c 			.dst_rq		= target_rq,