|
@@ -2094,6 +2094,7 @@ static int mmu_alloc_roots(struct kvm_vcpu *vcpu)
|
|
|
root_gfn = 0;
|
|
|
}
|
|
|
spin_lock(&vcpu->kvm->mmu_lock);
|
|
|
+ kvm_mmu_free_some_pages(vcpu->kvm);
|
|
|
sp = kvm_mmu_get_page(vcpu, root_gfn, 0,
|
|
|
PT64_ROOT_LEVEL, direct,
|
|
|
ACC_ALL, NULL);
|
|
@@ -2124,6 +2125,7 @@ static int mmu_alloc_roots(struct kvm_vcpu *vcpu)
|
|
|
root_gfn = i << 30;
|
|
|
}
|
|
|
spin_lock(&vcpu->kvm->mmu_lock);
|
|
|
+ kvm_mmu_free_some_pages(vcpu->kvm);
|
|
|
sp = kvm_mmu_get_page(vcpu, root_gfn, i << 30,
|
|
|
PT32_ROOT_LEVEL, direct,
|
|
|
ACC_ALL, NULL);
|
|
@@ -2496,9 +2498,6 @@ int kvm_mmu_load(struct kvm_vcpu *vcpu)
|
|
|
r = mmu_topup_memory_caches(vcpu);
|
|
|
if (r)
|
|
|
goto out;
|
|
|
- spin_lock(&vcpu->kvm->mmu_lock);
|
|
|
- kvm_mmu_free_some_pages(vcpu);
|
|
|
- spin_unlock(&vcpu->kvm->mmu_lock);
|
|
|
r = mmu_alloc_roots(vcpu);
|
|
|
spin_lock(&vcpu->kvm->mmu_lock);
|
|
|
mmu_sync_roots(vcpu);
|