|
@@ -97,13 +97,13 @@ static struct kmmio_probe *get_kmmio_probe(unsigned long addr)
|
|
|
static struct kmmio_fault_page *get_kmmio_fault_page(unsigned long page)
|
|
|
{
|
|
|
struct list_head *head;
|
|
|
- struct kmmio_fault_page *p;
|
|
|
+ struct kmmio_fault_page *f;
|
|
|
|
|
|
page &= PAGE_MASK;
|
|
|
head = kmmio_page_list(page);
|
|
|
- list_for_each_entry_rcu(p, head, list) {
|
|
|
- if (p->page == page)
|
|
|
- return p;
|
|
|
+ list_for_each_entry_rcu(f, head, list) {
|
|
|
+ if (f->page == page)
|
|
|
+ return f;
|
|
|
}
|
|
|
return NULL;
|
|
|
}
|
|
@@ -439,12 +439,12 @@ static void rcu_free_kmmio_fault_pages(struct rcu_head *head)
|
|
|
head,
|
|
|
struct kmmio_delayed_release,
|
|
|
rcu);
|
|
|
- struct kmmio_fault_page *p = dr->release_list;
|
|
|
- while (p) {
|
|
|
- struct kmmio_fault_page *next = p->release_next;
|
|
|
- BUG_ON(p->count);
|
|
|
- kfree(p);
|
|
|
- p = next;
|
|
|
+ struct kmmio_fault_page *f = dr->release_list;
|
|
|
+ while (f) {
|
|
|
+ struct kmmio_fault_page *next = f->release_next;
|
|
|
+ BUG_ON(f->count);
|
|
|
+ kfree(f);
|
|
|
+ f = next;
|
|
|
}
|
|
|
kfree(dr);
|
|
|
}
|
|
@@ -453,19 +453,19 @@ static void remove_kmmio_fault_pages(struct rcu_head *head)
|
|
|
{
|
|
|
struct kmmio_delayed_release *dr =
|
|
|
container_of(head, struct kmmio_delayed_release, rcu);
|
|
|
- struct kmmio_fault_page *p = dr->release_list;
|
|
|
+ struct kmmio_fault_page *f = dr->release_list;
|
|
|
struct kmmio_fault_page **prevp = &dr->release_list;
|
|
|
unsigned long flags;
|
|
|
|
|
|
spin_lock_irqsave(&kmmio_lock, flags);
|
|
|
- while (p) {
|
|
|
- if (!p->count) {
|
|
|
- list_del_rcu(&p->list);
|
|
|
- prevp = &p->release_next;
|
|
|
+ while (f) {
|
|
|
+ if (!f->count) {
|
|
|
+ list_del_rcu(&f->list);
|
|
|
+ prevp = &f->release_next;
|
|
|
} else {
|
|
|
- *prevp = p->release_next;
|
|
|
+ *prevp = f->release_next;
|
|
|
}
|
|
|
- p = p->release_next;
|
|
|
+ f = f->release_next;
|
|
|
}
|
|
|
spin_unlock_irqrestore(&kmmio_lock, flags);
|
|
|
|