|
@@ -71,7 +71,7 @@ mc_copy_user_page(void *from, void *to)
|
|
|
void v4_mc_copy_user_highpage(struct page *to, struct page *from,
|
|
|
unsigned long vaddr, struct vm_area_struct *vma)
|
|
|
{
|
|
|
- void *kto = kmap_atomic(to, KM_USER1);
|
|
|
+ void *kto = kmap_atomic(to);
|
|
|
|
|
|
if (!test_and_set_bit(PG_dcache_clean, &from->flags))
|
|
|
__flush_dcache_page(page_mapping(from), from);
|
|
@@ -85,7 +85,7 @@ void v4_mc_copy_user_highpage(struct page *to, struct page *from,
|
|
|
|
|
|
raw_spin_unlock(&minicache_lock);
|
|
|
|
|
|
- kunmap_atomic(kto, KM_USER1);
|
|
|
+ kunmap_atomic(kto);
|
|
|
}
|
|
|
|
|
|
/*
|
|
@@ -93,7 +93,7 @@ void v4_mc_copy_user_highpage(struct page *to, struct page *from,
|
|
|
*/
|
|
|
void v4_mc_clear_user_highpage(struct page *page, unsigned long vaddr)
|
|
|
{
|
|
|
- void *ptr, *kaddr = kmap_atomic(page, KM_USER0);
|
|
|
+ void *ptr, *kaddr = kmap_atomic(page);
|
|
|
asm volatile("\
|
|
|
mov r1, %2 @ 1\n\
|
|
|
mov r2, #0 @ 1\n\
|
|
@@ -111,7 +111,7 @@ void v4_mc_clear_user_highpage(struct page *page, unsigned long vaddr)
|
|
|
: "=r" (ptr)
|
|
|
: "0" (kaddr), "I" (PAGE_SIZE / 64)
|
|
|
: "r1", "r2", "r3", "ip", "lr");
|
|
|
- kunmap_atomic(kaddr, KM_USER0);
|
|
|
+ kunmap_atomic(kaddr);
|
|
|
}
|
|
|
|
|
|
struct cpu_user_fns v4_mc_user_fns __initdata = {
|