src_rq           1758 kernel/sched/core.c 		struct rq *src_rq, *dst_rq;
src_rq           1761 kernel/sched/core.c 		src_rq = task_rq(p);
src_rq           1764 kernel/sched/core.c 		rq_pin_lock(src_rq, &srf);
src_rq           1767 kernel/sched/core.c 		deactivate_task(src_rq, p, 0);
src_rq           1773 kernel/sched/core.c 		rq_unpin_lock(src_rq, &srf);
src_rq           1793 kernel/sched/core.c 	struct rq *src_rq, *dst_rq;
src_rq           1799 kernel/sched/core.c 	src_rq = cpu_rq(arg->src_cpu);
src_rq           1804 kernel/sched/core.c 	double_rq_lock(src_rq, dst_rq);
src_rq           1824 kernel/sched/core.c 	double_rq_unlock(src_rq, dst_rq);
src_rq           2139 kernel/sched/deadline.c 	struct rq *src_rq;
src_rq           2155 kernel/sched/deadline.c 		src_rq = cpu_rq(cpu);
src_rq           2163 kernel/sched/deadline.c 				   src_rq->dl.earliest_dl.next))
src_rq           2167 kernel/sched/deadline.c 		double_lock_balance(this_rq, src_rq);
src_rq           2173 kernel/sched/deadline.c 		if (src_rq->dl.dl_nr_running <= 1)
src_rq           2176 kernel/sched/deadline.c 		p = pick_earliest_pushable_dl_task(src_rq, this_cpu);
src_rq           2187 kernel/sched/deadline.c 			WARN_ON(p == src_rq->curr);
src_rq           2195 kernel/sched/deadline.c 					   src_rq->curr->dl.deadline))
src_rq           2200 kernel/sched/deadline.c 			deactivate_task(src_rq, p, 0);
src_rq           2208 kernel/sched/deadline.c 		double_unlock_balance(this_rq, src_rq);
src_rq           7125 kernel/sched/fair.c 	struct rq		*src_rq;
src_rq           7156 kernel/sched/fair.c 	lockdep_assert_held(&env->src_rq->lock);
src_rq           7177 kernel/sched/fair.c 	delta = rq_clock_task(env->src_rq) - p->se.exec_start;
src_rq           7208 kernel/sched/fair.c 		if (env->src_rq->nr_running > env->src_rq->nr_preferred_running)
src_rq           7250 kernel/sched/fair.c 	lockdep_assert_held(&env->src_rq->lock);
src_rq           7295 kernel/sched/fair.c 	if (task_running(env->src_rq, p)) {
src_rq           7328 kernel/sched/fair.c 	lockdep_assert_held(&env->src_rq->lock);
src_rq           7330 kernel/sched/fair.c 	deactivate_task(env->src_rq, p, DEQUEUE_NOCLOCK);
src_rq           7344 kernel/sched/fair.c 	lockdep_assert_held(&env->src_rq->lock);
src_rq           7347 kernel/sched/fair.c 			&env->src_rq->cfs_tasks, se.group_node) {
src_rq           7375 kernel/sched/fair.c 	struct list_head *tasks = &env->src_rq->cfs_tasks;
src_rq           7380 kernel/sched/fair.c 	lockdep_assert_held(&env->src_rq->lock);
src_rq           7390 kernel/sched/fair.c 		if (env->idle != CPU_NOT_IDLE && env->src_rq->nr_running <= 1)
src_rq           8763 kernel/sched/fair.c 	    (env->src_rq->cfs.h_nr_running == 1)) {
src_rq           8764 kernel/sched/fair.c 		if ((check_cpu_capacity(env->src_rq, sd)) &&
src_rq           8879 kernel/sched/fair.c 	env.src_rq = busiest;
src_rq           9186 kernel/sched/fair.c 			.src_rq		= busiest_rq,
src_rq           2054 kernel/sched/rt.c 	struct rq *src_rq;
src_rq           2082 kernel/sched/rt.c 		src_rq = cpu_rq(cpu);
src_rq           2091 kernel/sched/rt.c 		if (src_rq->rt.highest_prio.next >=
src_rq           2100 kernel/sched/rt.c 		double_lock_balance(this_rq, src_rq);
src_rq           2106 kernel/sched/rt.c 		p = pick_highest_pushable_task(src_rq, this_cpu);
src_rq           2113 kernel/sched/rt.c 			WARN_ON(p == src_rq->curr);
src_rq           2124 kernel/sched/rt.c 			if (p->prio < src_rq->curr->prio)
src_rq           2129 kernel/sched/rt.c 			deactivate_task(src_rq, p, 0);
src_rq           2140 kernel/sched/rt.c 		double_unlock_balance(this_rq, src_rq);