|
@@ -168,7 +168,7 @@ walk:
|
|
|
break;
|
|
|
}
|
|
|
|
|
|
- if (is_rsvd_bits_set(vcpu, pte, walker->level)) {
|
|
|
+ if (is_rsvd_bits_set(&vcpu->arch.mmu, pte, walker->level)) {
|
|
|
rsvd_fault = true;
|
|
|
break;
|
|
|
}
|
|
@@ -327,6 +327,7 @@ static void FNAME(pte_prefetch)(struct kvm_vcpu *vcpu, struct guest_walker *gw,
|
|
|
u64 *sptep)
|
|
|
{
|
|
|
struct kvm_mmu_page *sp;
|
|
|
+ struct kvm_mmu *mmu = &vcpu->arch.mmu;
|
|
|
pt_element_t *gptep = gw->prefetch_ptes;
|
|
|
u64 *spte;
|
|
|
int i;
|
|
@@ -358,7 +359,7 @@ static void FNAME(pte_prefetch)(struct kvm_vcpu *vcpu, struct guest_walker *gw,
|
|
|
gpte = gptep[i];
|
|
|
|
|
|
if (!is_present_gpte(gpte) ||
|
|
|
- is_rsvd_bits_set(vcpu, gpte, PT_PAGE_TABLE_LEVEL)) {
|
|
|
+ is_rsvd_bits_set(mmu, gpte, PT_PAGE_TABLE_LEVEL)) {
|
|
|
if (!sp->unsync)
|
|
|
__set_spte(spte, shadow_notrap_nonpresent_pte);
|
|
|
continue;
|
|
@@ -713,7 +714,7 @@ static int FNAME(sync_page)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
|
|
|
return -EINVAL;
|
|
|
|
|
|
gfn = gpte_to_gfn(gpte);
|
|
|
- if (is_rsvd_bits_set(vcpu, gpte, PT_PAGE_TABLE_LEVEL)
|
|
|
+ if (is_rsvd_bits_set(&vcpu->arch.mmu, gpte, PT_PAGE_TABLE_LEVEL)
|
|
|
|| gfn != sp->gfns[i] || !is_present_gpte(gpte)
|
|
|
|| !(gpte & PT_ACCESSED_MASK)) {
|
|
|
u64 nonpresent;
|