mirror of
https://github.com/tbsdtv/linux_media.git
synced 2025-07-23 12:43:29 +02:00
KVM: x86: Use standard mmu_notifier invalidate hooks for APIC access page
Now that KVM honors past and in-progress mmu_notifier invalidations when reloading the APIC-access page, use KVM's "standard" invalidation hooks to trigger a reload and delete the one-off usage of invalidate_range(). Aside from eliminating one-off code in KVM, dropping KVM's use of invalidate_range() will allow common mmu_notifier to redefine the API to be more strictly focused on invalidating secondary TLBs that share the primary MMU's page tables. Suggested-by: Jason Gunthorpe <jgg@nvidia.com> Cc: Alistair Popple <apopple@nvidia.com> Cc: Robin Murphy <robin.murphy@arm.com> Reviewed-by: Alistair Popple <apopple@nvidia.com> Reviewed-by: Paolo Bonzini <pbonzini@redhat.com> Link: https://lore.kernel.org/r/20230602011518.787006-3-seanjc@google.com Signed-off-by: Sean Christopherson <seanjc@google.com>
This commit is contained in:
@@ -1600,6 +1600,9 @@ bool kvm_unmap_gfn_range(struct kvm *kvm, struct kvm_gfn_range *range)
|
|||||||
if (tdp_mmu_enabled)
|
if (tdp_mmu_enabled)
|
||||||
flush = kvm_tdp_mmu_unmap_gfn_range(kvm, range, flush);
|
flush = kvm_tdp_mmu_unmap_gfn_range(kvm, range, flush);
|
||||||
|
|
||||||
|
if (range->slot->id == APIC_ACCESS_PAGE_PRIVATE_MEMSLOT)
|
||||||
|
kvm_make_all_cpus_request(kvm, KVM_REQ_APIC_PAGE_RELOAD);
|
||||||
|
|
||||||
return flush;
|
return flush;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -10435,20 +10435,6 @@ static void vcpu_load_eoi_exitmap(struct kvm_vcpu *vcpu)
|
|||||||
vcpu, (u64 *)vcpu->arch.ioapic_handled_vectors);
|
vcpu, (u64 *)vcpu->arch.ioapic_handled_vectors);
|
||||||
}
|
}
|
||||||
|
|
||||||
void kvm_arch_mmu_notifier_invalidate_range(struct kvm *kvm,
|
|
||||||
unsigned long start, unsigned long end)
|
|
||||||
{
|
|
||||||
unsigned long apic_address;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* The physical address of apic access page is stored in the VMCS.
|
|
||||||
* Update it when it becomes invalid.
|
|
||||||
*/
|
|
||||||
apic_address = gfn_to_hva(kvm, APIC_DEFAULT_PHYS_BASE >> PAGE_SHIFT);
|
|
||||||
if (start <= apic_address && apic_address < end)
|
|
||||||
kvm_make_all_cpus_request(kvm, KVM_REQ_APIC_PAGE_RELOAD);
|
|
||||||
}
|
|
||||||
|
|
||||||
void kvm_arch_guest_memory_reclaimed(struct kvm *kvm)
|
void kvm_arch_guest_memory_reclaimed(struct kvm *kvm)
|
||||||
{
|
{
|
||||||
static_call_cond(kvm_x86_guest_memory_reclaimed)(kvm);
|
static_call_cond(kvm_x86_guest_memory_reclaimed)(kvm);
|
||||||
|
@@ -2237,9 +2237,6 @@ static inline long kvm_arch_vcpu_async_ioctl(struct file *filp,
|
|||||||
}
|
}
|
||||||
#endif /* CONFIG_HAVE_KVM_VCPU_ASYNC_IOCTL */
|
#endif /* CONFIG_HAVE_KVM_VCPU_ASYNC_IOCTL */
|
||||||
|
|
||||||
void kvm_arch_mmu_notifier_invalidate_range(struct kvm *kvm,
|
|
||||||
unsigned long start, unsigned long end);
|
|
||||||
|
|
||||||
void kvm_arch_guest_memory_reclaimed(struct kvm *kvm);
|
void kvm_arch_guest_memory_reclaimed(struct kvm *kvm);
|
||||||
|
|
||||||
#ifdef CONFIG_HAVE_KVM_VCPU_RUN_PID_CHANGE
|
#ifdef CONFIG_HAVE_KVM_VCPU_RUN_PID_CHANGE
|
||||||
|
@@ -154,11 +154,6 @@ static unsigned long long kvm_active_vms;
|
|||||||
|
|
||||||
static DEFINE_PER_CPU(cpumask_var_t, cpu_kick_mask);
|
static DEFINE_PER_CPU(cpumask_var_t, cpu_kick_mask);
|
||||||
|
|
||||||
__weak void kvm_arch_mmu_notifier_invalidate_range(struct kvm *kvm,
|
|
||||||
unsigned long start, unsigned long end)
|
|
||||||
{
|
|
||||||
}
|
|
||||||
|
|
||||||
__weak void kvm_arch_guest_memory_reclaimed(struct kvm *kvm)
|
__weak void kvm_arch_guest_memory_reclaimed(struct kvm *kvm)
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
@@ -521,18 +516,6 @@ static inline struct kvm *mmu_notifier_to_kvm(struct mmu_notifier *mn)
|
|||||||
return container_of(mn, struct kvm, mmu_notifier);
|
return container_of(mn, struct kvm, mmu_notifier);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void kvm_mmu_notifier_invalidate_range(struct mmu_notifier *mn,
|
|
||||||
struct mm_struct *mm,
|
|
||||||
unsigned long start, unsigned long end)
|
|
||||||
{
|
|
||||||
struct kvm *kvm = mmu_notifier_to_kvm(mn);
|
|
||||||
int idx;
|
|
||||||
|
|
||||||
idx = srcu_read_lock(&kvm->srcu);
|
|
||||||
kvm_arch_mmu_notifier_invalidate_range(kvm, start, end);
|
|
||||||
srcu_read_unlock(&kvm->srcu, idx);
|
|
||||||
}
|
|
||||||
|
|
||||||
typedef bool (*hva_handler_t)(struct kvm *kvm, struct kvm_gfn_range *range);
|
typedef bool (*hva_handler_t)(struct kvm *kvm, struct kvm_gfn_range *range);
|
||||||
|
|
||||||
typedef void (*on_lock_fn_t)(struct kvm *kvm, unsigned long start,
|
typedef void (*on_lock_fn_t)(struct kvm *kvm, unsigned long start,
|
||||||
@@ -892,7 +875,6 @@ static void kvm_mmu_notifier_release(struct mmu_notifier *mn,
|
|||||||
}
|
}
|
||||||
|
|
||||||
static const struct mmu_notifier_ops kvm_mmu_notifier_ops = {
|
static const struct mmu_notifier_ops kvm_mmu_notifier_ops = {
|
||||||
.invalidate_range = kvm_mmu_notifier_invalidate_range,
|
|
||||||
.invalidate_range_start = kvm_mmu_notifier_invalidate_range_start,
|
.invalidate_range_start = kvm_mmu_notifier_invalidate_range_start,
|
||||||
.invalidate_range_end = kvm_mmu_notifier_invalidate_range_end,
|
.invalidate_range_end = kvm_mmu_notifier_invalidate_range_end,
|
||||||
.clear_flush_young = kvm_mmu_notifier_clear_flush_young,
|
.clear_flush_young = kvm_mmu_notifier_clear_flush_young,
|
||||||
|
Reference in New Issue
Block a user