|
@@ -840,6 +840,11 @@ static inline unsigned long slabs_node(struct kmem_cache *s, int node)
|
|
|
return atomic_long_read(&n->nr_slabs);
|
|
|
}
|
|
|
|
|
|
+static inline unsigned long node_nr_slabs(struct kmem_cache_node *n)
|
|
|
+{
|
|
|
+ return atomic_long_read(&n->nr_slabs);
|
|
|
+}
|
|
|
+
|
|
|
static inline void inc_slabs_node(struct kmem_cache *s, int node, int objects)
|
|
|
{
|
|
|
struct kmem_cache_node *n = get_node(s, node);
|
|
@@ -1058,6 +1063,8 @@ static inline unsigned long kmem_cache_flags(unsigned long objsize,
|
|
|
|
|
|
static inline unsigned long slabs_node(struct kmem_cache *s, int node)
|
|
|
{ return 0; }
|
|
|
+static inline unsigned long node_nr_slabs(struct kmem_cache_node *n)
|
|
|
+ { return 0; }
|
|
|
static inline void inc_slabs_node(struct kmem_cache *s, int node,
|
|
|
int objects) {}
|
|
|
static inline void dec_slabs_node(struct kmem_cache *s, int node,
|
|
@@ -1514,6 +1521,65 @@ static inline int node_match(struct kmem_cache_cpu *c, int node)
|
|
|
return 1;
|
|
|
}
|
|
|
|
|
|
+static int count_free(struct page *page)
|
|
|
+{
|
|
|
+ return page->objects - page->inuse;
|
|
|
+}
|
|
|
+
|
|
|
+static unsigned long count_partial(struct kmem_cache_node *n,
|
|
|
+ int (*get_count)(struct page *))
|
|
|
+{
|
|
|
+ unsigned long flags;
|
|
|
+ unsigned long x = 0;
|
|
|
+ struct page *page;
|
|
|
+
|
|
|
+ spin_lock_irqsave(&n->list_lock, flags);
|
|
|
+ list_for_each_entry(page, &n->partial, lru)
|
|
|
+ x += get_count(page);
|
|
|
+ spin_unlock_irqrestore(&n->list_lock, flags);
|
|
|
+ return x;
|
|
|
+}
|
|
|
+
|
|
|
+static inline unsigned long node_nr_objs(struct kmem_cache_node *n)
|
|
|
+{
|
|
|
+#ifdef CONFIG_SLUB_DEBUG
|
|
|
+ return atomic_long_read(&n->total_objects);
|
|
|
+#else
|
|
|
+ return 0;
|
|
|
+#endif
|
|
|
+}
|
|
|
+
|
|
|
+static noinline void
|
|
|
+slab_out_of_memory(struct kmem_cache *s, gfp_t gfpflags, int nid)
|
|
|
+{
|
|
|
+ int node;
|
|
|
+
|
|
|
+ printk(KERN_WARNING
|
|
|
+ "SLUB: Unable to allocate memory on node %d (gfp=0x%x)\n",
|
|
|
+ nid, gfpflags);
|
|
|
+ printk(KERN_WARNING " cache: %s, object size: %d, buffer size: %d, "
|
|
|
+ "default order: %d, min order: %d\n", s->name, s->objsize,
|
|
|
+ s->size, oo_order(s->oo), oo_order(s->min));
|
|
|
+
|
|
|
+ for_each_online_node(node) {
|
|
|
+ struct kmem_cache_node *n = get_node(s, node);
|
|
|
+ unsigned long nr_slabs;
|
|
|
+ unsigned long nr_objs;
|
|
|
+ unsigned long nr_free;
|
|
|
+
|
|
|
+ if (!n)
|
|
|
+ continue;
|
|
|
+
|
|
|
+ nr_free = count_partial(n, count_free);
|
|
|
+ nr_slabs = node_nr_slabs(n);
|
|
|
+ nr_objs = node_nr_objs(n);
|
|
|
+
|
|
|
+ printk(KERN_WARNING
|
|
|
+ " node %d: slabs: %ld, objs: %ld, free: %ld\n",
|
|
|
+ node, nr_slabs, nr_objs, nr_free);
|
|
|
+ }
|
|
|
+}
|
|
|
+
|
|
|
/*
|
|
|
* Slow path. The lockless freelist is empty or we need to perform
|
|
|
* debugging duties.
|
|
@@ -1595,6 +1661,8 @@ new_slab:
|
|
|
c->page = new;
|
|
|
goto load_freelist;
|
|
|
}
|
|
|
+ if (!(gfpflags & __GFP_NOWARN) && printk_ratelimit())
|
|
|
+ slab_out_of_memory(s, gfpflags, node);
|
|
|
return NULL;
|
|
|
debug:
|
|
|
if (!alloc_debug_processing(s, c->page, object, addr))
|
|
@@ -3368,20 +3436,6 @@ void *__kmalloc_node_track_caller(size_t size, gfp_t gfpflags,
|
|
|
}
|
|
|
|
|
|
#ifdef CONFIG_SLUB_DEBUG
|
|
|
-static unsigned long count_partial(struct kmem_cache_node *n,
|
|
|
- int (*get_count)(struct page *))
|
|
|
-{
|
|
|
- unsigned long flags;
|
|
|
- unsigned long x = 0;
|
|
|
- struct page *page;
|
|
|
-
|
|
|
- spin_lock_irqsave(&n->list_lock, flags);
|
|
|
- list_for_each_entry(page, &n->partial, lru)
|
|
|
- x += get_count(page);
|
|
|
- spin_unlock_irqrestore(&n->list_lock, flags);
|
|
|
- return x;
|
|
|
-}
|
|
|
-
|
|
|
static int count_inuse(struct page *page)
|
|
|
{
|
|
|
return page->inuse;
|
|
@@ -3392,11 +3446,6 @@ static int count_total(struct page *page)
|
|
|
return page->objects;
|
|
|
}
|
|
|
|
|
|
-static int count_free(struct page *page)
|
|
|
-{
|
|
|
- return page->objects - page->inuse;
|
|
|
-}
|
|
|
-
|
|
|
static int validate_slab(struct kmem_cache *s, struct page *page,
|
|
|
unsigned long *map)
|
|
|
{
|