|
@@ -114,8 +114,7 @@ radix_tree_node_alloc(struct radix_tree_root *root)
|
|
|
}
|
|
|
}
|
|
|
if (ret == NULL)
|
|
|
- ret = kmem_cache_alloc(radix_tree_node_cachep,
|
|
|
- set_migrateflags(gfp_mask, __GFP_RECLAIMABLE));
|
|
|
+ ret = kmem_cache_alloc(radix_tree_node_cachep, gfp_mask);
|
|
|
|
|
|
BUG_ON(radix_tree_is_indirect_ptr(ret));
|
|
|
return ret;
|
|
@@ -150,8 +149,7 @@ int radix_tree_preload(gfp_t gfp_mask)
|
|
|
rtp = &__get_cpu_var(radix_tree_preloads);
|
|
|
while (rtp->nr < ARRAY_SIZE(rtp->nodes)) {
|
|
|
preempt_enable();
|
|
|
- node = kmem_cache_alloc(radix_tree_node_cachep,
|
|
|
- set_migrateflags(gfp_mask, __GFP_RECLAIMABLE));
|
|
|
+ node = kmem_cache_alloc(radix_tree_node_cachep, gfp_mask);
|
|
|
if (node == NULL)
|
|
|
goto out;
|
|
|
preempt_disable();
|
|
@@ -1098,7 +1096,8 @@ void __init radix_tree_init(void)
|
|
|
{
|
|
|
radix_tree_node_cachep = kmem_cache_create("radix_tree_node",
|
|
|
sizeof(struct radix_tree_node), 0,
|
|
|
- SLAB_PANIC, radix_tree_node_ctor);
|
|
|
+ SLAB_PANIC | SLAB_RECLAIM_ACCOUNT,
|
|
|
+ radix_tree_node_ctor);
|
|
|
radix_tree_init_maxindex();
|
|
|
hotcpu_notifier(radix_tree_callback, 0);
|
|
|
}
|