|
@@ -2459,7 +2459,6 @@ static void __slab_free(struct kmem_cache *s, struct page *page,
|
|
|
void *prior;
|
|
|
void **object = (void *)x;
|
|
|
int was_frozen;
|
|
|
- int inuse;
|
|
|
struct page new;
|
|
|
unsigned long counters;
|
|
|
struct kmem_cache_node *n = NULL;
|
|
@@ -2472,13 +2471,17 @@ static void __slab_free(struct kmem_cache *s, struct page *page,
|
|
|
return;
|
|
|
|
|
|
do {
|
|
|
+ if (unlikely(n)) {
|
|
|
+ spin_unlock_irqrestore(&n->list_lock, flags);
|
|
|
+ n = NULL;
|
|
|
+ }
|
|
|
prior = page->freelist;
|
|
|
counters = page->counters;
|
|
|
set_freepointer(s, object, prior);
|
|
|
new.counters = counters;
|
|
|
was_frozen = new.frozen;
|
|
|
new.inuse--;
|
|
|
- if ((!new.inuse || !prior) && !was_frozen && !n) {
|
|
|
+ if ((!new.inuse || !prior) && !was_frozen) {
|
|
|
|
|
|
if (!kmem_cache_debug(s) && !prior)
|
|
|
|
|
@@ -2503,7 +2506,6 @@ static void __slab_free(struct kmem_cache *s, struct page *page,
|
|
|
|
|
|
}
|
|
|
}
|
|
|
- inuse = new.inuse;
|
|
|
|
|
|
} while (!cmpxchg_double_slab(s, page,
|
|
|
prior, counters,
|
|
@@ -2529,25 +2531,17 @@ static void __slab_free(struct kmem_cache *s, struct page *page,
|
|
|
return;
|
|
|
}
|
|
|
|
|
|
+ if (unlikely(!new.inuse && n->nr_partial > s->min_partial))
|
|
|
+ goto slab_empty;
|
|
|
+
|
|
|
/*
|
|
|
- * was_frozen may have been set after we acquired the list_lock in
|
|
|
- * an earlier loop. So we need to check it here again.
|
|
|
+ * Objects left in the slab. If it was not on the partial list before
|
|
|
+ * then add it.
|
|
|
*/
|
|
|
- if (was_frozen)
|
|
|
- stat(s, FREE_FROZEN);
|
|
|
- else {
|
|
|
- if (unlikely(!inuse && n->nr_partial > s->min_partial))
|
|
|
- goto slab_empty;
|
|
|
-
|
|
|
- /*
|
|
|
- * Objects left in the slab. If it was not on the partial list before
|
|
|
- * then add it.
|
|
|
- */
|
|
|
- if (unlikely(!prior)) {
|
|
|
- remove_full(s, page);
|
|
|
- add_partial(n, page, DEACTIVATE_TO_TAIL);
|
|
|
- stat(s, FREE_ADD_PARTIAL);
|
|
|
- }
|
|
|
+ if (kmem_cache_debug(s) && unlikely(!prior)) {
|
|
|
+ remove_full(s, page);
|
|
|
+ add_partial(n, page, DEACTIVATE_TO_TAIL);
|
|
|
+ stat(s, FREE_ADD_PARTIAL);
|
|
|
}
|
|
|
spin_unlock_irqrestore(&n->list_lock, flags);
|
|
|
return;
|