dst_cpu 100 drivers/irqchip/irq-ompic.c unsigned int dst_cpu; dst_cpu 103 drivers/irqchip/irq-ompic.c for_each_cpu(dst_cpu, mask) { dst_cpu 104 drivers/irqchip/irq-ompic.c set_bit(ipi_msg, &per_cpu(ops, dst_cpu)); dst_cpu 114 drivers/irqchip/irq-ompic.c OMPIC_CTRL_DST(dst_cpu) | dst_cpu 925 include/linux/perf_event.h int src_cpu, int dst_cpu); dst_cpu 24 include/linux/sched/numa_balancing.h int src_nid, int dst_cpu); dst_cpu 41 include/linux/sched/numa_balancing.h struct page *page, int src_nid, int dst_cpu) dst_cpu 492 include/trace/events/sched.h TP_PROTO(struct task_struct *tsk, int src_cpu, int dst_cpu), dst_cpu 494 include/trace/events/sched.h TP_ARGS(tsk, src_cpu, dst_cpu), dst_cpu 502 include/trace/events/sched.h __field( int, dst_cpu ) dst_cpu 512 include/trace/events/sched.h __entry->dst_cpu = dst_cpu; dst_cpu 513 include/trace/events/sched.h __entry->dst_nid = cpu_to_node(dst_cpu); dst_cpu 519 include/trace/events/sched.h __entry->dst_cpu, __entry->dst_nid) dst_cpu 527 include/trace/events/sched.h TP_PROTO(struct task_struct *tsk, int src_cpu, int dst_cpu), dst_cpu 529 include/trace/events/sched.h TP_ARGS(tsk, src_cpu, dst_cpu) dst_cpu 533 include/trace/events/sched.h TP_PROTO(struct task_struct *tsk, int src_cpu, int dst_cpu), dst_cpu 535 include/trace/events/sched.h TP_ARGS(tsk, src_cpu, dst_cpu) dst_cpu 541 include/trace/events/sched.h struct task_struct *dst_tsk, int dst_cpu), dst_cpu 543 include/trace/events/sched.h TP_ARGS(src_tsk, src_cpu, dst_tsk, dst_cpu), dst_cpu 554 include/trace/events/sched.h __field( int, dst_cpu ) dst_cpu 567 include/trace/events/sched.h __entry->dst_cpu = dst_cpu; dst_cpu 568 include/trace/events/sched.h __entry->dst_nid = cpu_to_node(dst_cpu); dst_cpu 575 include/trace/events/sched.h __entry->dst_cpu, __entry->dst_nid) dst_cpu 11427 kernel/events/core.c void perf_pmu_migrate_context(struct pmu *pmu, int src_cpu, int dst_cpu) dst_cpu 11435 kernel/events/core.c dst_ctx = &per_cpu_ptr(pmu->pmu_cpu_context, dst_cpu)->ctx; dst_cpu 11470 kernel/events/core.c account_event_cpu(event, dst_cpu); dst_cpu 11471 kernel/events/core.c perf_install_in_context(dst_ctx, event, dst_cpu); dst_cpu 11483 kernel/events/core.c account_event_cpu(event, dst_cpu); dst_cpu 11484 kernel/events/core.c perf_install_in_context(dst_ctx, event, dst_cpu); dst_cpu 1787 kernel/sched/core.c int src_cpu, dst_cpu; dst_cpu 1796 kernel/sched/core.c if (!cpu_active(arg->src_cpu) || !cpu_active(arg->dst_cpu)) dst_cpu 1800 kernel/sched/core.c dst_rq = cpu_rq(arg->dst_cpu); dst_cpu 1806 kernel/sched/core.c if (task_cpu(arg->dst_task) != arg->dst_cpu) dst_cpu 1812 kernel/sched/core.c if (!cpumask_test_cpu(arg->dst_cpu, arg->src_task->cpus_ptr)) dst_cpu 1818 kernel/sched/core.c __migrate_swap_task(arg->src_task, arg->dst_cpu); dst_cpu 1844 kernel/sched/core.c .dst_cpu = target_cpu, dst_cpu 1847 kernel/sched/core.c if (arg.src_cpu == arg.dst_cpu) dst_cpu 1854 kernel/sched/core.c if (!cpu_active(arg.src_cpu) || !cpu_active(arg.dst_cpu)) dst_cpu 1857 kernel/sched/core.c if (!cpumask_test_cpu(arg.dst_cpu, arg.src_task->cpus_ptr)) dst_cpu 1863 kernel/sched/core.c trace_sched_swap_numa(cur, arg.src_cpu, p, arg.dst_cpu); dst_cpu 1864 kernel/sched/core.c ret = stop_two_cpus(arg.dst_cpu, arg.src_cpu, migrate_swap_stop, &arg); dst_cpu 1409 kernel/sched/fair.c int src_nid, int dst_cpu) dst_cpu 1412 kernel/sched/fair.c int dst_nid = cpu_to_node(dst_cpu); dst_cpu 1415 kernel/sched/fair.c this_cpupid = cpu_pid_to_cpupid(dst_cpu, current->pid); dst_cpu 1508 kernel/sched/fair.c int dst_cpu, dst_nid; dst_cpu 1523 kernel/sched/fair.c struct rq *rq = cpu_rq(env->dst_cpu); dst_cpu 1545 kernel/sched/fair.c env->best_cpu = env->dst_cpu; dst_cpu 1593 kernel/sched/fair.c struct rq *dst_rq = cpu_rq(env->dst_cpu); dst_cpu 1700 kernel/sched/fair.c env->dst_cpu = select_idle_sibling(env->p, env->src_cpu, dst_cpu 1701 kernel/sched/fair.c env->dst_cpu); dst_cpu 1732 kernel/sched/fair.c env->dst_cpu = cpu; dst_cpu 6235 kernel/sched/fair.c static unsigned long cpu_util_next(int cpu, struct task_struct *p, int dst_cpu) dst_cpu 6246 kernel/sched/fair.c if (task_cpu(p) == cpu && dst_cpu != cpu) dst_cpu 6248 kernel/sched/fair.c else if (task_cpu(p) != cpu && dst_cpu == cpu) dst_cpu 6260 kernel/sched/fair.c if (dst_cpu == cpu) dst_cpu 6277 kernel/sched/fair.c compute_energy(struct task_struct *p, int dst_cpu, struct perf_domain *pd) dst_cpu 6294 kernel/sched/fair.c unsigned long cpu_util, util_cfs = cpu_util_next(cpu, p, dst_cpu); dst_cpu 6295 kernel/sched/fair.c struct task_struct *tsk = cpu == dst_cpu ? p : NULL; dst_cpu 7128 kernel/sched/fair.c int dst_cpu; dst_cpu 7201 kernel/sched/fair.c dst_nid = cpu_to_node(env->dst_cpu); dst_cpu 7259 kernel/sched/fair.c if (throttled_lb_pair(task_group(p), env->src_cpu, env->dst_cpu)) dst_cpu 7262 kernel/sched/fair.c if (!cpumask_test_cpu(env->dst_cpu, p->cpus_ptr)) { dst_cpu 7331 kernel/sched/fair.c set_task_cpu(p, env->dst_cpu); dst_cpu 8182 kernel/sched/fair.c sched_asym_prefer(env->dst_cpu, sg->asym_prefer_cpu)) { dst_cpu 8248 kernel/sched/fair.c local_group = cpumask_test_cpu(env->dst_cpu, sched_group_span(sg)); dst_cpu 8255 kernel/sched/fair.c update_group_capacity(env->sd, env->dst_cpu); dst_cpu 8360 kernel/sched/fair.c if (sched_asym_prefer(busiest_cpu, env->dst_cpu)) dst_cpu 8387 kernel/sched/fair.c local->load_per_task = cpu_avg_load_per_task(env->dst_cpu); dst_cpu 8694 kernel/sched/fair.c capacity_of(env->dst_cpu) < capacity && dst_cpu 8745 kernel/sched/fair.c sched_asym_prefer(env->dst_cpu, env->src_cpu); dst_cpu 8765 kernel/sched/fair.c (capacity_of(env->src_cpu)*sd->imbalance_pct < capacity_of(env->dst_cpu)*100)) dst_cpu 8796 kernel/sched/fair.c if (!cpumask_test_cpu(env->dst_cpu, env->cpus)) dst_cpu 8822 kernel/sched/fair.c return balance_cpu == env->dst_cpu; dst_cpu 8842 kernel/sched/fair.c .dst_cpu = this_cpu, dst_cpu 8946 kernel/sched/fair.c __cpumask_clear_cpu(env.dst_cpu, env.cpus); dst_cpu 8949 kernel/sched/fair.c env.dst_cpu = env.new_dst_cpu; dst_cpu 9183 kernel/sched/fair.c .dst_cpu = target_cpu,