|
@@ -2705,7 +2705,8 @@ EXPORT_SYMBOL_GPL(kvm_mmu_unprotect_page_virt);
|
|
|
|
|
|
void __kvm_mmu_free_some_pages(struct kvm_vcpu *vcpu)
|
|
void __kvm_mmu_free_some_pages(struct kvm_vcpu *vcpu)
|
|
{
|
|
{
|
|
- while (vcpu->kvm->arch.n_free_mmu_pages < KVM_REFILL_PAGES) {
|
|
|
|
|
|
+ while (vcpu->kvm->arch.n_free_mmu_pages < KVM_REFILL_PAGES &&
|
|
|
|
+ !list_empty(&vcpu->kvm->arch.active_mmu_pages)) {
|
|
struct kvm_mmu_page *sp;
|
|
struct kvm_mmu_page *sp;
|
|
|
|
|
|
sp = container_of(vcpu->kvm->arch.active_mmu_pages.prev,
|
|
sp = container_of(vcpu->kvm->arch.active_mmu_pages.prev,
|