used_mask 49 arch/arc/kernel/perf_event.c unsigned long used_mask[BITS_TO_LONGS(ARC_PERF_MAX_COUNTERS)]; used_mask 352 arch/arc/kernel/perf_event.c __clear_bit(event->hw.idx, pmu_cpu->used_mask); used_mask 366 arch/arc/kernel/perf_event.c idx = ffz(pmu_cpu->used_mask[0]); used_mask 370 arch/arc/kernel/perf_event.c __set_bit(idx, pmu_cpu->used_mask); used_mask 394 arch/arm/kernel/perf_event_v6.c if (test_and_set_bit(ARMV6_CYCLE_COUNTER, cpuc->used_mask)) used_mask 403 arch/arm/kernel/perf_event_v6.c if (!test_and_set_bit(ARMV6_COUNTER1, cpuc->used_mask)) used_mask 406 arch/arm/kernel/perf_event_v6.c if (!test_and_set_bit(ARMV6_COUNTER0, cpuc->used_mask)) used_mask 417 arch/arm/kernel/perf_event_v6.c clear_bit(event->hw.idx, cpuc->used_mask); used_mask 1042 arch/arm/kernel/perf_event_v7.c if (test_and_set_bit(ARMV7_IDX_CYCLE_COUNTER, cpuc->used_mask)) used_mask 1053 arch/arm/kernel/perf_event_v7.c if (!test_and_set_bit(idx, cpuc->used_mask)) used_mask 1064 arch/arm/kernel/perf_event_v7.c clear_bit(event->hw.idx, cpuc->used_mask); used_mask 1626 arch/arm/kernel/perf_event_v7.c if (test_and_set_bit(bit, cpuc->used_mask)) used_mask 1632 arch/arm/kernel/perf_event_v7.c clear_bit(bit, cpuc->used_mask); used_mask 1650 arch/arm/kernel/perf_event_v7.c clear_bit(bit, cpuc->used_mask); used_mask 1956 arch/arm/kernel/perf_event_v7.c if (test_and_set_bit(bit, cpuc->used_mask)) used_mask 1962 arch/arm/kernel/perf_event_v7.c clear_bit(bit, cpuc->used_mask); used_mask 1980 arch/arm/kernel/perf_event_v7.c clear_bit(bit, cpuc->used_mask); used_mask 280 arch/arm/kernel/perf_event_xscale.c if (test_and_set_bit(XSCALE_CYCLE_COUNTER, cpuc->used_mask)) used_mask 285 arch/arm/kernel/perf_event_xscale.c if (!test_and_set_bit(XSCALE_COUNTER1, cpuc->used_mask)) used_mask 288 arch/arm/kernel/perf_event_xscale.c if (!test_and_set_bit(XSCALE_COUNTER0, cpuc->used_mask)) used_mask 298 arch/arm/kernel/perf_event_xscale.c clear_bit(event->hw.idx, cpuc->used_mask); used_mask 656 arch/arm/kernel/perf_event_xscale.c if (!test_and_set_bit(XSCALE_COUNTER3, cpuc->used_mask)) used_mask 658 arch/arm/kernel/perf_event_xscale.c else if (!test_and_set_bit(XSCALE_COUNTER2, cpuc->used_mask)) used_mask 762 arch/arm64/kernel/perf_event.c if (!test_and_set_bit(idx, cpuc->used_mask)) used_mask 778 arch/arm64/kernel/perf_event.c if (!test_and_set_bit(idx, cpuc->used_mask)) { used_mask 780 arch/arm64/kernel/perf_event.c if (!test_and_set_bit(idx - 1, cpuc->used_mask)) used_mask 783 arch/arm64/kernel/perf_event.c clear_bit(idx, cpuc->used_mask); used_mask 798 arch/arm64/kernel/perf_event.c if (!test_and_set_bit(ARMV8_IDX_CYCLE_COUNTER, cpuc->used_mask)) used_mask 816 arch/arm64/kernel/perf_event.c clear_bit(idx, cpuc->used_mask); used_mask 818 arch/arm64/kernel/perf_event.c clear_bit(idx - 1, cpuc->used_mask); used_mask 33 arch/csky/kernel/perf_event.c unsigned long used_mask[BITS_TO_LONGS(CSKY_PMU_MAX_EVENTS)]; used_mask 39 arch/mips/kernel/perf_event_mipsxx.c unsigned long used_mask[BITS_TO_LONGS(MIPS_MAX_HWEVENTS)]; used_mask 308 arch/mips/kernel/perf_event_mipsxx.c !test_and_set_bit(i, cpuc->used_mask)) used_mask 502 arch/mips/kernel/perf_event_mipsxx.c clear_bit(idx, cpuc->used_mask); used_mask 1409 arch/mips/kernel/perf_event_mipsxx.c if (!test_bit(n, cpuc->used_mask)) used_mask 54 arch/nds32/include/asm/pmu.h unsigned long used_mask[BITS_TO_LONGS(MAX_COUNTERS)]; used_mask 574 arch/nds32/kernel/perf_event_cpu.c if (!test_and_set_bit(idx, cpuc->used_mask)) used_mask 576 arch/nds32/kernel/perf_event_cpu.c if (!test_and_set_bit(NDS32_IDX_COUNTER0, cpuc->used_mask)) used_mask 578 arch/nds32/kernel/perf_event_cpu.c if (!test_and_set_bit(NDS32_IDX_COUNTER1, cpuc->used_mask)) used_mask 581 arch/nds32/kernel/perf_event_cpu.c if (!test_and_set_bit(idx, cpuc->used_mask)) used_mask 583 arch/nds32/kernel/perf_event_cpu.c else if (!test_and_set_bit(NDS32_IDX_COUNTER1, cpuc->used_mask)) used_mask 586 arch/nds32/kernel/perf_event_cpu.c (NDS32_IDX_CYCLE_COUNTER, cpuc->used_mask)) used_mask 589 arch/nds32/kernel/perf_event_cpu.c if (!test_and_set_bit(idx, cpuc->used_mask)) used_mask 698 arch/nds32/kernel/perf_event_cpu.c int enabled = bitmap_weight(hw_events->used_mask, used_mask 993 arch/nds32/kernel/perf_event_cpu.c clear_bit(idx, hw_events->used_mask); used_mask 30 arch/sh/kernel/perf_event.c unsigned long used_mask[BITS_TO_LONGS(MAX_HWEVENTS)]; used_mask 257 arch/sh/kernel/perf_event.c __clear_bit(event->hw.idx, cpuc->used_mask); used_mask 271 arch/sh/kernel/perf_event.c if (__test_and_set_bit(idx, cpuc->used_mask)) { used_mask 272 arch/sh/kernel/perf_event.c idx = find_first_zero_bit(cpuc->used_mask, sh_pmu->num_events); used_mask 276 arch/sh/kernel/perf_event.c __set_bit(idx, cpuc->used_mask); used_mask 874 arch/x86/events/core.c unsigned long used_mask[BITS_TO_LONGS(X86_PMC_IDX_MAX)]; used_mask 879 arch/x86/events/core.c bitmap_zero(used_mask, X86_PMC_IDX_MAX); used_mask 934 arch/x86/events/core.c if (test_bit(hwc->idx, used_mask)) used_mask 937 arch/x86/events/core.c __set_bit(hwc->idx, used_mask); used_mask 1192 arch/x86/events/intel/p4.c static int p4_next_cntr(int thread, unsigned long *used_mask, used_mask 1199 arch/x86/events/intel/p4.c if (j != -1 && !test_bit(j, used_mask)) used_mask 1208 arch/x86/events/intel/p4.c unsigned long used_mask[BITS_TO_LONGS(X86_PMC_IDX_MAX)]; used_mask 1218 arch/x86/events/intel/p4.c bitmap_zero(used_mask, X86_PMC_IDX_MAX); used_mask 1248 arch/x86/events/intel/p4.c cntr_idx = p4_next_cntr(thread, used_mask, bind); used_mask 1282 arch/x86/events/intel/p4.c set_bit(cntr_idx, used_mask); used_mask 424 arch/x86/events/intel/uncore.c unsigned long used_mask[BITS_TO_LONGS(UNCORE_PMC_IDX_MAX)]; used_mask 429 arch/x86/events/intel/uncore.c bitmap_zero(used_mask, UNCORE_PMC_IDX_MAX); used_mask 452 arch/x86/events/intel/uncore.c if (test_bit(hwc->idx, used_mask)) used_mask 455 arch/x86/events/intel/uncore.c __set_bit(hwc->idx, used_mask); used_mask 54 arch/xtensa/kernel/perf_event.c unsigned long used_mask[BITS_TO_LONGS(XCHAL_NUM_PERF_COUNTERS)]; used_mask 287 arch/xtensa/kernel/perf_event.c if (__test_and_set_bit(idx, ev->used_mask)) { used_mask 288 arch/xtensa/kernel/perf_event.c idx = find_first_zero_bit(ev->used_mask, used_mask 293 arch/xtensa/kernel/perf_event.c __set_bit(idx, ev->used_mask); used_mask 312 arch/xtensa/kernel/perf_event.c __clear_bit(event->hw.idx, ev->used_mask); used_mask 365 arch/xtensa/kernel/perf_event.c for (i = find_first_bit(ev->used_mask, XCHAL_NUM_PERF_COUNTERS); used_mask 367 arch/xtensa/kernel/perf_event.c i = find_next_bit(ev->used_mask, XCHAL_NUM_PERF_COUNTERS, i + 1)) { used_mask 73 drivers/perf/arm-cci.c unsigned long *used_mask; used_mask 320 drivers/perf/arm-cci.c if (test_and_set_bit(CCI400_PMU_CYCLE_CNTR_IDX, hw->used_mask)) used_mask 327 drivers/perf/arm-cci.c if (!test_and_set_bit(idx, hw->used_mask)) used_mask 649 drivers/perf/arm-cci.c for_each_set_bit(i, cci_pmu->hw_events.used_mask, cci_pmu->num_cntrs) { used_mask 810 drivers/perf/arm-cci.c if (!test_and_set_bit(idx, hw->used_mask)) used_mask 1100 drivers/perf/arm-cci.c int enabled = bitmap_weight(hw_events->used_mask, cci_pmu->num_cntrs); used_mask 1224 drivers/perf/arm-cci.c clear_bit(idx, hw_events->used_mask); used_mask 1263 drivers/perf/arm-cci.c .used_mask = mask, used_mask 1633 drivers/perf/arm-cci.c cci_pmu->hw_events.used_mask = devm_kcalloc(dev, used_mask 1635 drivers/perf/arm-cci.c sizeof(*cci_pmu->hw_events.used_mask), used_mask 1637 drivers/perf/arm-cci.c if (!cci_pmu->hw_events.used_mask) used_mask 93 drivers/perf/arm_dsu_pmu.c DECLARE_BITMAP(used_mask, DSU_PMU_MAX_HW_CNTRS); used_mask 305 drivers/perf/arm_dsu_pmu.c unsigned long *used_mask = hw_events->used_mask; used_mask 308 drivers/perf/arm_dsu_pmu.c if (test_and_set_bit(DSU_PMU_IDX_CYCLE_COUNTER, used_mask)) used_mask 313 drivers/perf/arm_dsu_pmu.c idx = find_first_zero_bit(used_mask, dsu_pmu->num_counters); used_mask 316 drivers/perf/arm_dsu_pmu.c set_bit(idx, hw_events->used_mask); used_mask 475 drivers/perf/arm_dsu_pmu.c clear_bit(idx, hw_events->used_mask); used_mask 486 drivers/perf/arm_dsu_pmu.c if (bitmap_empty(dsu_pmu->hw_events.used_mask, DSU_PMU_MAX_HW_CNTRS)) used_mask 533 drivers/perf/arm_dsu_pmu.c memset(fake_hw.used_mask, 0, sizeof(fake_hw.used_mask)); used_mask 320 drivers/perf/arm_pmu.c memset(&fake_pmu.used_mask, 0, sizeof(fake_pmu.used_mask)); used_mask 451 drivers/perf/arm_pmu.c int enabled = bitmap_weight(hw_events->used_mask, armpmu->num_events); used_mask 705 drivers/perf/arm_pmu.c int enabled = bitmap_weight(hw_events->used_mask, armpmu->num_events); used_mask 146 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c unsigned long *used_mask = ddrc_pmu->pmu_events.used_mask; used_mask 151 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c if (test_bit(idx, used_mask)) used_mask 154 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c set_bit(idx, used_mask); used_mask 103 drivers/perf/hisilicon/hisi_uncore_pmu.c unsigned long *used_mask = hisi_pmu->pmu_events.used_mask; used_mask 107 drivers/perf/hisilicon/hisi_uncore_pmu.c idx = find_first_zero_bit(used_mask, num_counters); used_mask 111 drivers/perf/hisilicon/hisi_uncore_pmu.c set_bit(idx, used_mask); used_mask 123 drivers/perf/hisilicon/hisi_uncore_pmu.c clear_bit(idx, hisi_pmu->pmu_events.used_mask); used_mask 323 drivers/perf/hisilicon/hisi_uncore_pmu.c int enabled = bitmap_weight(hisi_pmu->pmu_events.used_mask, used_mask 53 drivers/perf/hisilicon/hisi_uncore_pmu.h DECLARE_BITMAP(used_mask, HISI_MAX_COUNTERS); used_mask 159 drivers/perf/qcom_l3_pmu.c unsigned long used_mask[BITS_TO_LONGS(L3_NUM_COUNTERS)]; used_mask 557 drivers/perf/qcom_l3_pmu.c idx = bitmap_find_free_region(l3pmu->used_mask, L3_NUM_COUNTERS, order); used_mask 584 drivers/perf/qcom_l3_pmu.c bitmap_release_region(l3pmu->used_mask, hwc->idx, order); used_mask 55 include/linux/perf/arm_pmu.h DECLARE_BITMAP(used_mask, ARMPMU_MAX_HWEVENTS); used_mask 5633 mm/page_alloc.c nodemask_t used_mask; used_mask 5640 mm/page_alloc.c nodes_clear(used_mask); used_mask 5643 mm/page_alloc.c while ((node = find_next_best_node(local_node, &used_mask)) >= 0) {