|
@@ -681,8 +681,7 @@ static struct kvm_mmu_page *kvm_mmu_get_page(struct kvm_vcpu *vcpu,
|
|
|
unsigned level,
|
|
|
int metaphysical,
|
|
|
unsigned access,
|
|
|
- u64 *parent_pte,
|
|
|
- bool *new_page)
|
|
|
+ u64 *parent_pte)
|
|
|
{
|
|
|
union kvm_mmu_page_role role;
|
|
|
unsigned index;
|
|
@@ -722,8 +721,6 @@ static struct kvm_mmu_page *kvm_mmu_get_page(struct kvm_vcpu *vcpu,
|
|
|
vcpu->arch.mmu.prefetch_page(vcpu, sp);
|
|
|
if (!metaphysical)
|
|
|
rmap_write_protect(vcpu->kvm, gfn);
|
|
|
- if (new_page)
|
|
|
- *new_page = 1;
|
|
|
return sp;
|
|
|
}
|
|
|
|
|
@@ -1006,8 +1003,7 @@ static int __nonpaging_map(struct kvm_vcpu *vcpu, gva_t v, int write,
|
|
|
>> PAGE_SHIFT;
|
|
|
new_table = kvm_mmu_get_page(vcpu, pseudo_gfn,
|
|
|
v, level - 1,
|
|
|
- 1, ACC_ALL, &table[index],
|
|
|
- NULL);
|
|
|
+ 1, ACC_ALL, &table[index]);
|
|
|
if (!new_table) {
|
|
|
pgprintk("nonpaging_map: ENOMEM\n");
|
|
|
kvm_release_page_clean(page);
|
|
@@ -1100,7 +1096,7 @@ static void mmu_alloc_roots(struct kvm_vcpu *vcpu)
|
|
|
|
|
|
ASSERT(!VALID_PAGE(root));
|
|
|
sp = kvm_mmu_get_page(vcpu, root_gfn, 0,
|
|
|
- PT64_ROOT_LEVEL, 0, ACC_ALL, NULL, NULL);
|
|
|
+ PT64_ROOT_LEVEL, 0, ACC_ALL, NULL);
|
|
|
root = __pa(sp->spt);
|
|
|
++sp->root_count;
|
|
|
vcpu->arch.mmu.root_hpa = root;
|
|
@@ -1121,7 +1117,7 @@ static void mmu_alloc_roots(struct kvm_vcpu *vcpu)
|
|
|
root_gfn = 0;
|
|
|
sp = kvm_mmu_get_page(vcpu, root_gfn, i << 30,
|
|
|
PT32_ROOT_LEVEL, !is_paging(vcpu),
|
|
|
- ACC_ALL, NULL, NULL);
|
|
|
+ ACC_ALL, NULL);
|
|
|
root = __pa(sp->spt);
|
|
|
++sp->root_count;
|
|
|
vcpu->arch.mmu.pae_root[i] = root | PT_PRESENT_MASK;
|