On Mon, Mar 24, 2025, Mingwei Zhang wrote:
static void kvm_pmu_incr_counter(struct kvm_pmc *pmc) {
- pmc->emulated_counter++;
- kvm_pmu_request_counter_reprogram(pmc);
- struct kvm_vcpu *vcpu = pmc->vcpu;
- /*
* For perf-based PMUs, accumulate software-emulated events separately
* from pmc->counter, as pmc->counter is offset by the count of the
* associated perf event. Request reprogramming, which will consult
* both emulated and hardware-generated events to detect overflow.
*/
- if (!kvm_mediated_pmu_enabled(vcpu)) {
pmc->emulated_counter++;
kvm_pmu_request_counter_reprogram(pmc);
return;
- }
- /*
* For mediated PMUs, pmc->counter is updated when the vCPU's PMU is
* put, and will be loaded into hardware when the PMU is loaded. Simply
* increment the counter and signal overflow if it wraps to zero.
*/
- pmc->counter = (pmc->counter + 1) & pmc_bitmask(pmc);
- if (!pmc->counter) {
Ugh, this is broken for the fastpath. If kvm_skip_emulated_instruction() is invoked by handle_fastpath_set_msr_irqoff() or handle_fastpath_hlt(), KVM may consume stale information (GLOBAL_CTRL, GLOBAL_STATUS and PMCs), and even if KVM gets lucky and those are all fresh, the PMC and GLOBAL_STATUS changes won't be propagated back to hardware.
The best idea I have is to track whether or not the guest may be counting branches and/or instruction based on eventsels, and then bail from fastpaths that need to skip instructions. That flag would also be useful to further optimize kvm_pmu_trigger_event().
pmc_to_pmu(pmc)->global_status |= BIT_ULL(pmc->idx);
if (pmc_pmi_enabled(pmc))
kvm_make_request(KVM_REQ_PMI, vcpu);
- }
} static inline bool cpl_is_matched(struct kvm_pmc *pmc) -- 2.49.0.395.g12beb8f557-goog