|
@@ -69,7 +69,7 @@ int sysctl_max_map_count = DEFAULT_MAX_MAP_COUNT;
|
|
int sysctl_nr_trim_pages = 1; /* page trimming behaviour */
|
|
int sysctl_nr_trim_pages = 1; /* page trimming behaviour */
|
|
int heap_stack_gap = 0;
|
|
int heap_stack_gap = 0;
|
|
|
|
|
|
-atomic_t mmap_pages_allocated;
|
|
|
|
|
|
+atomic_long_t mmap_pages_allocated;
|
|
|
|
|
|
EXPORT_SYMBOL(mem_map);
|
|
EXPORT_SYMBOL(mem_map);
|
|
EXPORT_SYMBOL(num_physpages);
|
|
EXPORT_SYMBOL(num_physpages);
|
|
@@ -463,12 +463,7 @@ SYSCALL_DEFINE1(brk, unsigned long, brk)
|
|
*/
|
|
*/
|
|
void __init mmap_init(void)
|
|
void __init mmap_init(void)
|
|
{
|
|
{
|
|
- vm_region_jar = kmem_cache_create("vm_region_jar",
|
|
|
|
- sizeof(struct vm_region), 0,
|
|
|
|
- SLAB_PANIC, NULL);
|
|
|
|
- vm_area_cachep = kmem_cache_create("vm_area_struct",
|
|
|
|
- sizeof(struct vm_area_struct), 0,
|
|
|
|
- SLAB_PANIC, NULL);
|
|
|
|
|
|
+ vm_region_jar = KMEM_CACHE(vm_region, SLAB_PANIC);
|
|
}
|
|
}
|
|
|
|
|
|
/*
|
|
/*
|
|
@@ -486,27 +481,24 @@ static noinline void validate_nommu_regions(void)
|
|
return;
|
|
return;
|
|
|
|
|
|
last = rb_entry(lastp, struct vm_region, vm_rb);
|
|
last = rb_entry(lastp, struct vm_region, vm_rb);
|
|
- if (unlikely(last->vm_end <= last->vm_start))
|
|
|
|
- BUG();
|
|
|
|
- if (unlikely(last->vm_top < last->vm_end))
|
|
|
|
- BUG();
|
|
|
|
|
|
+ BUG_ON(unlikely(last->vm_end <= last->vm_start));
|
|
|
|
+ BUG_ON(unlikely(last->vm_top < last->vm_end));
|
|
|
|
|
|
while ((p = rb_next(lastp))) {
|
|
while ((p = rb_next(lastp))) {
|
|
region = rb_entry(p, struct vm_region, vm_rb);
|
|
region = rb_entry(p, struct vm_region, vm_rb);
|
|
last = rb_entry(lastp, struct vm_region, vm_rb);
|
|
last = rb_entry(lastp, struct vm_region, vm_rb);
|
|
|
|
|
|
- if (unlikely(region->vm_end <= region->vm_start))
|
|
|
|
- BUG();
|
|
|
|
- if (unlikely(region->vm_top < region->vm_end))
|
|
|
|
- BUG();
|
|
|
|
- if (unlikely(region->vm_start < last->vm_top))
|
|
|
|
- BUG();
|
|
|
|
|
|
+ BUG_ON(unlikely(region->vm_end <= region->vm_start));
|
|
|
|
+ BUG_ON(unlikely(region->vm_top < region->vm_end));
|
|
|
|
+ BUG_ON(unlikely(region->vm_start < last->vm_top));
|
|
|
|
|
|
lastp = p;
|
|
lastp = p;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
#else
|
|
#else
|
|
-#define validate_nommu_regions() do {} while(0)
|
|
|
|
|
|
+static void validate_nommu_regions(void)
|
|
|
|
+{
|
|
|
|
+}
|
|
#endif
|
|
#endif
|
|
|
|
|
|
/*
|
|
/*
|
|
@@ -563,16 +555,17 @@ static void free_page_series(unsigned long from, unsigned long to)
|
|
struct page *page = virt_to_page(from);
|
|
struct page *page = virt_to_page(from);
|
|
|
|
|
|
kdebug("- free %lx", from);
|
|
kdebug("- free %lx", from);
|
|
- atomic_dec(&mmap_pages_allocated);
|
|
|
|
|
|
+ atomic_long_dec(&mmap_pages_allocated);
|
|
if (page_count(page) != 1)
|
|
if (page_count(page) != 1)
|
|
- kdebug("free page %p [%d]", page, page_count(page));
|
|
|
|
|
|
+ kdebug("free page %p: refcount not one: %d",
|
|
|
|
+ page, page_count(page));
|
|
put_page(page);
|
|
put_page(page);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
/*
|
|
/*
|
|
* release a reference to a region
|
|
* release a reference to a region
|
|
- * - the caller must hold the region semaphore, which this releases
|
|
|
|
|
|
+ * - the caller must hold the region semaphore for writing, which this releases
|
|
* - the region may not have been added to the tree yet, in which case vm_top
|
|
* - the region may not have been added to the tree yet, in which case vm_top
|
|
* will equal vm_start
|
|
* will equal vm_start
|
|
*/
|
|
*/
|
|
@@ -1096,7 +1089,7 @@ static int do_mmap_private(struct vm_area_struct *vma,
|
|
goto enomem;
|
|
goto enomem;
|
|
|
|
|
|
total = 1 << order;
|
|
total = 1 << order;
|
|
- atomic_add(total, &mmap_pages_allocated);
|
|
|
|
|
|
+ atomic_long_add(total, &mmap_pages_allocated);
|
|
|
|
|
|
point = rlen >> PAGE_SHIFT;
|
|
point = rlen >> PAGE_SHIFT;
|
|
|
|
|
|
@@ -1107,7 +1100,7 @@ static int do_mmap_private(struct vm_area_struct *vma,
|
|
order = ilog2(total - point);
|
|
order = ilog2(total - point);
|
|
n = 1 << order;
|
|
n = 1 << order;
|
|
kdebug("shave %lu/%lu @%lu", n, total - point, total);
|
|
kdebug("shave %lu/%lu @%lu", n, total - point, total);
|
|
- atomic_sub(n, &mmap_pages_allocated);
|
|
|
|
|
|
+ atomic_long_sub(n, &mmap_pages_allocated);
|
|
total -= n;
|
|
total -= n;
|
|
set_page_refcounted(pages + total);
|
|
set_page_refcounted(pages + total);
|
|
__free_pages(pages + total, order);
|
|
__free_pages(pages + total, order);
|
|
@@ -1536,10 +1529,15 @@ int do_munmap(struct mm_struct *mm, unsigned long start, size_t len)
|
|
/* find the first potentially overlapping VMA */
|
|
/* find the first potentially overlapping VMA */
|
|
vma = find_vma(mm, start);
|
|
vma = find_vma(mm, start);
|
|
if (!vma) {
|
|
if (!vma) {
|
|
- printk(KERN_WARNING
|
|
|
|
- "munmap of memory not mmapped by process %d (%s):"
|
|
|
|
- " 0x%lx-0x%lx\n",
|
|
|
|
- current->pid, current->comm, start, start + len - 1);
|
|
|
|
|
|
+ static int limit = 0;
|
|
|
|
+ if (limit < 5) {
|
|
|
|
+ printk(KERN_WARNING
|
|
|
|
+ "munmap of memory not mmapped by process %d"
|
|
|
|
+ " (%s): 0x%lx-0x%lx\n",
|
|
|
|
+ current->pid, current->comm,
|
|
|
|
+ start, start + len - 1);
|
|
|
|
+ limit++;
|
|
|
|
+ }
|
|
return -EINVAL;
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
|