|
@@ -467,7 +467,7 @@ static void FNAME(pte_prefetch)(struct kvm_vcpu *vcpu, struct guest_walker *gw,
|
|
|
static u64 *FNAME(fetch)(struct kvm_vcpu *vcpu, gva_t addr,
|
|
|
struct guest_walker *gw,
|
|
|
int user_fault, int write_fault, int hlevel,
|
|
|
- int *ptwrite, pfn_t pfn, bool map_writable,
|
|
|
+ int *emulate, pfn_t pfn, bool map_writable,
|
|
|
bool prefault)
|
|
|
{
|
|
|
unsigned access = gw->pt_access;
|
|
@@ -538,7 +538,7 @@ static u64 *FNAME(fetch)(struct kvm_vcpu *vcpu, gva_t addr,
|
|
|
}
|
|
|
|
|
|
mmu_set_spte(vcpu, it.sptep, access, gw->pte_access,
|
|
|
- user_fault, write_fault, ptwrite, it.level,
|
|
|
+ user_fault, write_fault, emulate, it.level,
|
|
|
gw->gfn, pfn, prefault, map_writable);
|
|
|
FNAME(pte_prefetch)(vcpu, gw, it.sptep);
|
|
|
|
|
@@ -572,7 +572,7 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr, u32 error_code,
|
|
|
int user_fault = error_code & PFERR_USER_MASK;
|
|
|
struct guest_walker walker;
|
|
|
u64 *sptep;
|
|
|
- int write_pt = 0;
|
|
|
+ int emulate = 0;
|
|
|
int r;
|
|
|
pfn_t pfn;
|
|
|
int level = PT_PAGE_TABLE_LEVEL;
|
|
@@ -633,19 +633,19 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr, u32 error_code,
|
|
|
if (!force_pt_level)
|
|
|
transparent_hugepage_adjust(vcpu, &walker.gfn, &pfn, &level);
|
|
|
sptep = FNAME(fetch)(vcpu, addr, &walker, user_fault, write_fault,
|
|
|
- level, &write_pt, pfn, map_writable, prefault);
|
|
|
+ level, &emulate, pfn, map_writable, prefault);
|
|
|
(void)sptep;
|
|
|
- pgprintk("%s: shadow pte %p %llx ptwrite %d\n", __func__,
|
|
|
- sptep, *sptep, write_pt);
|
|
|
+ pgprintk("%s: shadow pte %p %llx emulate %d\n", __func__,
|
|
|
+ sptep, *sptep, emulate);
|
|
|
|
|
|
- if (!write_pt)
|
|
|
+ if (!emulate)
|
|
|
vcpu->arch.last_pt_write_count = 0; /* reset fork detector */
|
|
|
|
|
|
++vcpu->stat.pf_fixed;
|
|
|
trace_kvm_mmu_audit(vcpu, AUDIT_POST_PAGE_FAULT);
|
|
|
spin_unlock(&vcpu->kvm->mmu_lock);
|
|
|
|
|
|
- return write_pt;
|
|
|
+ return emulate;
|
|
|
|
|
|
out_unlock:
|
|
|
spin_unlock(&vcpu->kvm->mmu_lock);
|