Lines Matching refs:perf_event
151 if (pmc->perf_event) in kvm_pmu_get_pair_counter_value()
152 counter += perf_event_read_value(pmc->perf_event, &enabled, in kvm_pmu_get_pair_counter_value()
205 if (pmc->perf_event) { in kvm_pmu_release_perf_event()
206 perf_event_disable(pmc->perf_event); in kvm_pmu_release_perf_event()
207 perf_event_release_kernel(pmc->perf_event); in kvm_pmu_release_perf_event()
208 pmc->perf_event = NULL; in kvm_pmu_release_perf_event()
223 if (!pmc->perf_event) in kvm_pmu_stop_counter()
328 if (pmc->perf_event) { in kvm_pmu_enable_counter_mask()
329 perf_event_enable(pmc->perf_event); in kvm_pmu_enable_counter_mask()
330 if (pmc->perf_event->state != PERF_EVENT_STATE_ACTIVE) in kvm_pmu_enable_counter_mask()
363 if (pmc->perf_event) in kvm_pmu_disable_counter_mask()
364 perf_event_disable(pmc->perf_event); in kvm_pmu_disable_counter_mask()
471 static void kvm_pmu_perf_overflow(struct perf_event *perf_event, in kvm_pmu_perf_overflow() argument
475 struct kvm_pmc *pmc = perf_event->overflow_handler_context; in kvm_pmu_perf_overflow()
476 struct arm_pmu *cpu_pmu = to_arm_pmu(perf_event->pmu); in kvm_pmu_perf_overflow()
481 cpu_pmu->pmu.stop(perf_event, PERF_EF_UPDATE); in kvm_pmu_perf_overflow()
487 period = -(local64_read(&perf_event->count)); in kvm_pmu_perf_overflow()
492 local64_set(&perf_event->hw.period_left, 0); in kvm_pmu_perf_overflow()
493 perf_event->attr.sample_period = period; in kvm_pmu_perf_overflow()
494 perf_event->hw.sample_period = period; in kvm_pmu_perf_overflow()
507 cpu_pmu->pmu.start(perf_event, PERF_EF_RELOAD); in kvm_pmu_perf_overflow()
602 struct perf_event *event; in kvm_pmu_create_perf_event()
677 pmc->perf_event = event; in kvm_pmu_create_perf_event()
746 struct perf_event *event; in kvm_pmu_probe_pmuver()