|
@@ -47,8 +47,6 @@ static int __init alloc_node_page_cgroup(int nid)
|
|
struct page_cgroup *base, *pc;
|
|
struct page_cgroup *base, *pc;
|
|
unsigned long table_size;
|
|
unsigned long table_size;
|
|
unsigned long start_pfn, nr_pages, index;
|
|
unsigned long start_pfn, nr_pages, index;
|
|
- struct page *page;
|
|
|
|
- unsigned int order;
|
|
|
|
|
|
|
|
start_pfn = NODE_DATA(nid)->node_start_pfn;
|
|
start_pfn = NODE_DATA(nid)->node_start_pfn;
|
|
nr_pages = NODE_DATA(nid)->node_spanned_pages;
|
|
nr_pages = NODE_DATA(nid)->node_spanned_pages;
|
|
@@ -57,13 +55,11 @@ static int __init alloc_node_page_cgroup(int nid)
|
|
return 0;
|
|
return 0;
|
|
|
|
|
|
table_size = sizeof(struct page_cgroup) * nr_pages;
|
|
table_size = sizeof(struct page_cgroup) * nr_pages;
|
|
- order = get_order(table_size);
|
|
|
|
- page = alloc_pages_node(nid, GFP_NOWAIT | __GFP_ZERO, order);
|
|
|
|
- if (!page)
|
|
|
|
- page = alloc_pages_node(-1, GFP_NOWAIT | __GFP_ZERO, order);
|
|
|
|
- if (!page)
|
|
|
|
|
|
+
|
|
|
|
+ base = __alloc_bootmem_node_nopanic(NODE_DATA(nid),
|
|
|
|
+ table_size, PAGE_SIZE, __pa(MAX_DMA_ADDRESS));
|
|
|
|
+ if (!base)
|
|
return -ENOMEM;
|
|
return -ENOMEM;
|
|
- base = page_address(page);
|
|
|
|
for (index = 0; index < nr_pages; index++) {
|
|
for (index = 0; index < nr_pages; index++) {
|
|
pc = base + index;
|
|
pc = base + index;
|
|
__init_page_cgroup(pc, start_pfn + index);
|
|
__init_page_cgroup(pc, start_pfn + index);
|
|
@@ -73,7 +69,7 @@ static int __init alloc_node_page_cgroup(int nid)
|
|
return 0;
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
|
|
-void __init page_cgroup_init(void)
|
|
|
|
|
|
+void __init page_cgroup_init_flatmem(void)
|
|
{
|
|
{
|
|
|
|
|
|
int nid, fail;
|
|
int nid, fail;
|
|
@@ -117,16 +113,11 @@ static int __init_refok init_section_page_cgroup(unsigned long pfn)
|
|
if (!section->page_cgroup) {
|
|
if (!section->page_cgroup) {
|
|
nid = page_to_nid(pfn_to_page(pfn));
|
|
nid = page_to_nid(pfn_to_page(pfn));
|
|
table_size = sizeof(struct page_cgroup) * PAGES_PER_SECTION;
|
|
table_size = sizeof(struct page_cgroup) * PAGES_PER_SECTION;
|
|
- if (slab_is_available()) {
|
|
|
|
- base = kmalloc_node(table_size,
|
|
|
|
- GFP_KERNEL | __GFP_NOWARN, nid);
|
|
|
|
- if (!base)
|
|
|
|
- base = vmalloc_node(table_size, nid);
|
|
|
|
- } else {
|
|
|
|
- base = __alloc_bootmem_node_nopanic(NODE_DATA(nid),
|
|
|
|
- table_size,
|
|
|
|
- PAGE_SIZE, __pa(MAX_DMA_ADDRESS));
|
|
|
|
- }
|
|
|
|
|
|
+ VM_BUG_ON(!slab_is_available());
|
|
|
|
+ base = kmalloc_node(table_size,
|
|
|
|
+ GFP_KERNEL | __GFP_NOWARN, nid);
|
|
|
|
+ if (!base)
|
|
|
|
+ base = vmalloc_node(table_size, nid);
|
|
} else {
|
|
} else {
|
|
/*
|
|
/*
|
|
* We don't have to allocate page_cgroup again, but
|
|
* We don't have to allocate page_cgroup again, but
|