Lines Matching refs:left
1077 s64 left = local64_read(&hwc->period_left); in x86_perf_event_set_period() local
1087 if (unlikely(left <= -period)) { in x86_perf_event_set_period()
1088 left = period; in x86_perf_event_set_period()
1089 local64_set(&hwc->period_left, left); in x86_perf_event_set_period()
1094 if (unlikely(left <= 0)) { in x86_perf_event_set_period()
1095 left += period; in x86_perf_event_set_period()
1096 local64_set(&hwc->period_left, left); in x86_perf_event_set_period()
1103 if (unlikely(left < 2)) in x86_perf_event_set_period()
1104 left = 2; in x86_perf_event_set_period()
1106 if (left > x86_pmu.max_period) in x86_perf_event_set_period()
1107 left = x86_pmu.max_period; in x86_perf_event_set_period()
1110 left = x86_pmu.limit_period(event, left); in x86_perf_event_set_period()
1112 per_cpu(pmc_prev_left[idx], smp_processor_id()) = left; in x86_perf_event_set_period()
1118 local64_set(&hwc->prev_count, (u64)-left); in x86_perf_event_set_period()
1120 wrmsrl(hwc->event_base, (u64)(-left) & x86_pmu.cntval_mask); in x86_perf_event_set_period()
1129 (u64)(-left) & x86_pmu.cntval_mask); in x86_perf_event_set_period()