/linux-4.1.27/include/linux/ |
D | cpumask.h | 94 #define num_online_cpus() cpumask_weight(cpu_online_mask) 95 #define num_possible_cpus() cpumask_weight(cpu_possible_mask) 96 #define num_present_cpus() cpumask_weight(cpu_present_mask) 97 #define num_active_cpus() cpumask_weight(cpu_active_mask) 474 static inline unsigned int cpumask_weight(const struct cpumask *srcp) in cpumask_weight() function
|
D | topology.h | 42 #define nr_cpus_node(node) cpumask_weight(cpumask_of_node(node))
|
D | cpufreq.h | 140 return cpumask_weight(policy->cpus) > 1; in policy_is_shared()
|
/linux-4.1.27/drivers/staging/lustre/lustre/libcfs/linux/ |
D | linux-cpu.c | 243 cpumask_weight(cptab->ctb_cpumask) : in cfs_cpt_weight() 244 cpumask_weight(cptab->ctb_parts[cpt].cpt_cpumask); in cfs_cpt_weight() 388 if (cpumask_weight(mask) == 0 || in cfs_cpt_set_cpumask() 622 if (number >= cpumask_weight(node)) { in cfs_cpt_choose_ncpus() 782 n = num - cpumask_weight(part->cpt_cpumask); in cfs_cpt_table_create() 789 LASSERT(num >= cpumask_weight(part->cpt_cpumask)); in cfs_cpt_table_create() 790 if (num == cpumask_weight(part->cpt_cpumask)) in cfs_cpt_table_create() 796 num != cpumask_weight(cptab->ctb_parts[ncpt - 1].cpt_cpumask)) { in cfs_cpt_table_create() 799 cpumask_weight(cptab->ctb_parts[ncpt - 1].cpt_cpumask)); in cfs_cpt_table_create()
|
/linux-4.1.27/arch/x86/platform/uv/ |
D | uv_nmi.c | 309 k = n - cpumask_weight(uv_nmi_cpu_mask); in uv_nmi_wait_cpus() 361 cpumask_weight(uv_nmi_cpu_mask), in uv_nmi_wait() 371 cpumask_weight(uv_nmi_cpu_mask), in uv_nmi_wait() 638 if (cpumask_weight(uv_nmi_cpu_mask)) in uv_handle_nmi()
|
/linux-4.1.27/arch/mips/kernel/ |
D | crash.c | 53 while ((cpumask_weight(&cpus_in_crash) < ncpus) && (--msecs > 0)) { in crash_kexec_prepare_cpus()
|
D | pm-cps.c | 141 online = cpumask_weight(coupled_mask); in cps_pm_enter_state()
|
/linux-4.1.27/mm/ |
D | quicklist.c | 44 num_cpus_on_node = cpumask_weight(cpumask_of_node(node)); in max_pages()
|
/linux-4.1.27/arch/ia64/include/asm/ |
D | acpi.h | 128 low_cpu = cpumask_weight(&early_cpu_possible_map); in per_cpu_scan_finalize()
|
/linux-4.1.27/arch/x86/kernel/ |
D | tsc_sync.c | 116 return (cpumask_weight(cpu_core_mask(cpu)) > 1) ? 2 : 20; in loop_timeout()
|
D | smpboot.c | 432 if (cpumask_weight(cpu_sibling_mask(cpu)) == 1) { in set_cpu_sibling_map() 1301 if (cpumask_weight(cpu_sibling_mask(cpu)) == 1) in remove_siblinginfo()
|
/linux-4.1.27/arch/ia64/kernel/ |
D | irq.c | 80 if (cpumask_weight(cpumask) != 1) in is_affinity_mask_valid()
|
D | setup.c | 565 per_cpu_scan_finalize((cpumask_weight(&early_cpu_possible_map) == 0 ? in setup_arch() 566 32 : cpumask_weight(&early_cpu_possible_map)), in setup_arch() 706 cpumask_weight(&cpu_core_map[cpunum])); in show_cpuinfo()
|
D | irq_ia64.c | 314 cfg->move_cleanup_count = cpumask_weight(&cleanup_mask); in irq_complete_move()
|
D | smpboot.c | 601 last = (cpumask_weight(&cpu_core_map[cpu]) == 1 ? 1 : 0); in remove_siblinginfo()
|
/linux-4.1.27/crypto/ |
D | pcrypt.c | 85 if (!cpumask_weight(cpumask->mask)) in pcrypt_do_parallel() 88 cpu_index = cpu % cpumask_weight(cpumask->mask); in pcrypt_do_parallel() 283 cpu_index = ictx->tfm_count % cpumask_weight(cpu_online_mask); in pcrypt_aead_init_tfm()
|
/linux-4.1.27/arch/powerpc/mm/ |
D | numa.c | 1257 return cpumask_weight(changes); in update_cpu_associativity_changes_mask() 1373 weight = cpumask_weight(&cpu_associativity_changes_mask); in arch_update_cpu_topology() 1427 if (cpumask_weight(&updated_cpus)) { in arch_update_cpu_topology() 1446 if (!cpumask_weight(&updated_cpus)) in arch_update_cpu_topology() 1488 if (prrn_enabled && cpumask_weight(&cpu_associativity_changes_mask)) in topology_timer_fn()
|
/linux-4.1.27/arch/tile/kernel/ |
D | smpboot.c | 118 if (cpumask_weight(cpu_present_mask) > 1) { in smp_prepare_cpus()
|
D | hardwall.c | 593 if (cpumask_weight(&p->cpus_allowed) != 1) in hardwall_activate() 631 if (cpumask_weight(&task->cpus_allowed) != 1) { in _hardwall_deactivate() 634 cpumask_weight(&task->cpus_allowed)); in _hardwall_deactivate() 887 cpumask_weight(&info->cpumask)); in hardwall_destroy()
|
D | pci_gx.c | 132 count = cpumask_weight(&intr_cpus_map); in tile_irq_cpu()
|
/linux-4.1.27/drivers/idle/ |
D | i7300_idle.c | 465 if (cpumask_weight(idle_cpumask) != num_online_cpus()) in i7300_idle_notifier() 483 if (cpumask_weight(idle_cpumask) == (num_online_cpus() - 1)) { in i7300_idle_notifier()
|
/linux-4.1.27/arch/x86/kernel/cpu/ |
D | proc.c | 15 seq_printf(m, "siblings\t: %d\n", cpumask_weight(cpu_core_mask(cpu))); in show_cpuinfo_core()
|
D | perf_event_intel.c | 3394 w = cpumask_weight(topology_thread_cpumask(cpu)); in fixup_ht_bug()
|
/linux-4.1.27/arch/alpha/kernel/ |
D | process.c | 125 while (cpumask_weight(cpu_present_mask)) in common_shutdown_1()
|
/linux-4.1.27/arch/tile/mm/ |
D | init.c | 359 if (cpumask_weight(&ktext_mask) > 1) { in setup_ktext() 559 if (cpumask_weight(&ktext_mask) == 1) { in kernel_physical_mapping_init()
|
/linux-4.1.27/lib/ |
D | percpu_ida.c | 66 for (cpus_have_tags = cpumask_weight(&pool->cpus_have_tags); in steal_tags()
|
/linux-4.1.27/arch/x86/mm/ |
D | mmio-mod.c | 418 if (downed_cpus == NULL || cpumask_weight(downed_cpus) == 0) in leave_uniprocessor()
|
/linux-4.1.27/drivers/cpuidle/ |
D | coupled.c | 618 coupled->online_count = cpumask_weight(&cpus); in cpuidle_coupled_update_online_cpus()
|
/linux-4.1.27/kernel/ |
D | padata.c | 58 cpu_index = seq_nr % cpumask_weight(pd->cpumask.pcpu); in padata_cpu_hash() 177 num_cpus = cpumask_weight(pd->cpumask.pcpu); in padata_get_next()
|
D | smp.c | 445 if (unlikely(!cpumask_weight(cfd->cpumask))) in smp_call_function_many()
|
D | stop_machine.c | 361 cpu_stop_init_done(&done, cpumask_weight(cpumask)); in __stop_cpus()
|
D | workqueue.c | 4554 if (cpumask_weight(&cpumask) != 1) in restore_unbound_workers_cpumask()
|
/linux-4.1.27/arch/ia64/mm/ |
D | tlb.c | 325 if (mm != current->active_mm || cpumask_weight(mm_cpumask(mm)) != 1) { in flush_tlb_range()
|
/linux-4.1.27/kernel/sched/ |
D | core.c | 4657 if (!cpumask_weight(cur)) in cpuset_cpumask_can_shrink() 4662 trial_cpus = cpumask_weight(trial); in cpuset_cpumask_can_shrink() 4762 p->nr_cpus_allowed = cpumask_weight(new_mask); in do_set_cpus_allowed() 5444 if (!cpumask_weight(sched_group_cpus(group))) { in sched_domain_debug_one() 5513 if (cpumask_weight(sched_domain_span(sd)) == 1) in sd_degenerate() 5750 size = cpumask_weight(sched_domain_span(sd)); in update_top_cache_domain() 5923 sg->sgc->capacity = SCHED_CAPACITY_SCALE * cpumask_weight(sg_span); in build_overlap_sched_groups() 6041 sg->group_weight = cpumask_weight(sched_group_cpus(sg)); in init_sched_groups_capacity() 6184 sd_weight = cpumask_weight(tl->mask(cpu)); in sd_init() 6741 sd->span_weight = cpumask_weight(sched_domain_span(sd)); in build_sched_domains()
|
D | rt.c | 632 weight = cpumask_weight(rd->span); in do_balance_runtime() 2073 weight = cpumask_weight(new_mask); in set_cpus_allowed_rt()
|
D | deadline.c | 1615 weight = cpumask_weight(new_mask); in set_cpus_allowed_dl()
|
/linux-4.1.27/arch/sparc/kernel/ |
D | ds.c | 538 ncpus = cpumask_weight(mask); in dr_cpu_configure() 597 ncpus = cpumask_weight(mask); in dr_cpu_unconfigure()
|
/linux-4.1.27/net/core/ |
D | flow.c | 358 atomic_set(&info.cpuleft, cpumask_weight(mask)); in flow_cache_flush()
|
D | net-sysfs.c | 689 RPS_MAP_SIZE(cpumask_weight(mask)), L1_CACHE_BYTES), in store_rps_map()
|
/linux-4.1.27/arch/mips/cavium-octeon/ |
D | octeon-irq.c | 228 int weight = cpumask_weight(data->affinity); in next_cpu_for_irq() 717 if (cpumask_weight(data->affinity) > 1) { in octeon_irq_cpu_offline_ciu() 749 if (cpumask_weight(dest) != 1) in octeon_irq_ciu_set_affinity()
|
/linux-4.1.27/kernel/time/ |
D | clockevents.c | 666 cpumask_weight(dev->cpumask) == 1 && in tick_cleanup_dead_cpu()
|
/linux-4.1.27/drivers/staging/lustre/lustre/ptlrpc/ |
D | ptlrpcd.c | 517 pc->pc_npartners = cpumask_weight(&mask) - 1; in ptlrpcd_bind()
|
D | service.c | 560 if (cpumask_weight(topology_thread_cpumask(0)) > 1) { in ptlrpc_server_nthreads_check() 2771 weight = cpumask_weight(topology_thread_cpumask(0)); in ptlrpc_hr_init()
|
/linux-4.1.27/include/trace/events/ |
D | xen.h | 416 TP_fast_assign(__entry->ncpus = cpumask_weight(cpus);
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
D | qib_file_ops.c | 1169 const unsigned int weight = cpumask_weight(¤t->cpus_allowed); in assign_ctxt_affinity() 1178 (cpumask_weight(local_mask) <= qib_cpulist_count)) { in assign_ctxt_affinity() 1652 cpumask_weight(¤t->cpus_allowed); in qib_assign_ctxt()
|
D | qib_iba7322.c | 3461 cpumask_weight(local_mask) == num_online_cpus()) { in qib_setup_7322_interrupt()
|
/linux-4.1.27/drivers/cpufreq/ |
D | cpufreq.c | 1363 cpus = cpumask_weight(policy->cpus); in __cpufreq_remove_dev_prepare() 1412 cpus = cpumask_weight(policy->cpus); in __cpufreq_remove_dev_finish()
|
/linux-4.1.27/kernel/rcu/ |
D | tree_plugin.h | 1283 if (cpumask_weight(cm) == 0) in rcu_boost_kthread_setaffinity() 2354 if (tick_nohz_full_running && cpumask_weight(tick_nohz_full_mask)) in rcu_init_nohz()
|
D | tree.c | 3320 if (cpumask_weight(cm) == 0) in synchronize_sched_expedited()
|
/linux-4.1.27/drivers/net/ethernet/tile/ |
D | tilepro.c | 1420 count = cpumask_weight(&priv->network_cpus_map); in tile_net_open()
|
D | tilegx.c | 1302 int network_cpus_count = cpumask_weight(&network_cpus_map); in tile_net_init_mpipe()
|
/linux-4.1.27/arch/powerpc/xmon/ |
D | xmon.c | 496 if (cpumask_weight(&cpus_in_xmon) >= ncpus) in xmon_core()
|
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/o2iblnd/ |
D | o2iblnd.c | 625 off = do_div(nid, cpumask_weight(mask)); in kiblnd_get_completion_vector()
|
/linux-4.1.27/drivers/target/iscsi/ |
D | iscsi_target.c | 3724 ord = conn->bitmap_id % cpumask_weight(cpu_online_mask); in iscsit_thread_get_cpumask()
|