|
@@ -75,7 +75,6 @@ static void async_pf_execute(struct work_struct *work)
|
|
|
spin_lock(&vcpu->async_pf.lock);
|
|
|
list_add_tail(&apf->link, &vcpu->async_pf.done);
|
|
|
apf->page = page;
|
|
|
- apf->done = true;
|
|
|
spin_unlock(&vcpu->async_pf.lock);
|
|
|
|
|
|
/*
|
|
@@ -99,9 +98,8 @@ void kvm_clear_async_pf_completion_queue(struct kvm_vcpu *vcpu)
|
|
|
struct kvm_async_pf *work =
|
|
|
list_entry(vcpu->async_pf.queue.next,
|
|
|
typeof(*work), queue);
|
|
|
- cancel_work_sync(&work->work);
|
|
|
list_del(&work->queue);
|
|
|
- if (!work->done) { /* work was canceled */
|
|
|
+ if (cancel_work_sync(&work->work)) {
|
|
|
mmdrop(work->mm);
|
|
|
kvm_put_kvm(vcpu->kvm); /* == work->vcpu->kvm */
|
|
|
kmem_cache_free(async_pf_cache, work);
|
|
@@ -166,7 +164,6 @@ int kvm_setup_async_pf(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn,
|
|
|
return 0;
|
|
|
|
|
|
work->page = NULL;
|
|
|
- work->done = false;
|
|
|
work->vcpu = vcpu;
|
|
|
work->gva = gva;
|
|
|
work->addr = gfn_to_hva(vcpu->kvm, gfn);
|