|
@@ -1376,8 +1376,8 @@ static struct vm_struct *__get_vm_area_node(unsigned long size,
|
|
|
struct vm_struct *__get_vm_area(unsigned long size, unsigned long flags,
|
|
|
unsigned long start, unsigned long end)
|
|
|
{
|
|
|
- return __get_vm_area_node(size, 1, flags, start, end, -1, GFP_KERNEL,
|
|
|
- __builtin_return_address(0));
|
|
|
+ return __get_vm_area_node(size, 1, flags, start, end, NUMA_NO_NODE,
|
|
|
+ GFP_KERNEL, __builtin_return_address(0));
|
|
|
}
|
|
|
EXPORT_SYMBOL_GPL(__get_vm_area);
|
|
|
|
|
@@ -1385,8 +1385,8 @@ struct vm_struct *__get_vm_area_caller(unsigned long size, unsigned long flags,
|
|
|
unsigned long start, unsigned long end,
|
|
|
const void *caller)
|
|
|
{
|
|
|
- return __get_vm_area_node(size, 1, flags, start, end, -1, GFP_KERNEL,
|
|
|
- caller);
|
|
|
+ return __get_vm_area_node(size, 1, flags, start, end, NUMA_NO_NODE,
|
|
|
+ GFP_KERNEL, caller);
|
|
|
}
|
|
|
|
|
|
/**
|
|
@@ -1401,14 +1401,15 @@ struct vm_struct *__get_vm_area_caller(unsigned long size, unsigned long flags,
|
|
|
struct vm_struct *get_vm_area(unsigned long size, unsigned long flags)
|
|
|
{
|
|
|
return __get_vm_area_node(size, 1, flags, VMALLOC_START, VMALLOC_END,
|
|
|
- -1, GFP_KERNEL, __builtin_return_address(0));
|
|
|
+ NUMA_NO_NODE, GFP_KERNEL,
|
|
|
+ __builtin_return_address(0));
|
|
|
}
|
|
|
|
|
|
struct vm_struct *get_vm_area_caller(unsigned long size, unsigned long flags,
|
|
|
const void *caller)
|
|
|
{
|
|
|
return __get_vm_area_node(size, 1, flags, VMALLOC_START, VMALLOC_END,
|
|
|
- -1, GFP_KERNEL, caller);
|
|
|
+ NUMA_NO_NODE, GFP_KERNEL, caller);
|
|
|
}
|
|
|
|
|
|
/**
|
|
@@ -1650,7 +1651,7 @@ fail:
|
|
|
* @end: vm area range end
|
|
|
* @gfp_mask: flags for the page level allocator
|
|
|
* @prot: protection mask for the allocated pages
|
|
|
- * @node: node to use for allocation or -1
|
|
|
+ * @node: node to use for allocation or NUMA_NO_NODE
|
|
|
* @caller: caller's return address
|
|
|
*
|
|
|
* Allocate enough pages to cover @size from the page level
|
|
@@ -1706,7 +1707,7 @@ fail:
|
|
|
* @align: desired alignment
|
|
|
* @gfp_mask: flags for the page level allocator
|
|
|
* @prot: protection mask for the allocated pages
|
|
|
- * @node: node to use for allocation or -1
|
|
|
+ * @node: node to use for allocation or NUMA_NO_NODE
|
|
|
* @caller: caller's return address
|
|
|
*
|
|
|
* Allocate enough pages to cover @size from the page level
|
|
@@ -1723,7 +1724,7 @@ static void *__vmalloc_node(unsigned long size, unsigned long align,
|
|
|
|
|
|
void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot)
|
|
|
{
|
|
|
- return __vmalloc_node(size, 1, gfp_mask, prot, -1,
|
|
|
+ return __vmalloc_node(size, 1, gfp_mask, prot, NUMA_NO_NODE,
|
|
|
__builtin_return_address(0));
|
|
|
}
|
|
|
EXPORT_SYMBOL(__vmalloc);
|
|
@@ -1746,7 +1747,8 @@ static inline void *__vmalloc_node_flags(unsigned long size,
|
|
|
*/
|
|
|
void *vmalloc(unsigned long size)
|
|
|
{
|
|
|
- return __vmalloc_node_flags(size, -1, GFP_KERNEL | __GFP_HIGHMEM);
|
|
|
+ return __vmalloc_node_flags(size, NUMA_NO_NODE,
|
|
|
+ GFP_KERNEL | __GFP_HIGHMEM);
|
|
|
}
|
|
|
EXPORT_SYMBOL(vmalloc);
|
|
|
|
|
@@ -1762,7 +1764,7 @@ EXPORT_SYMBOL(vmalloc);
|
|
|
*/
|
|
|
void *vzalloc(unsigned long size)
|
|
|
{
|
|
|
- return __vmalloc_node_flags(size, -1,
|
|
|
+ return __vmalloc_node_flags(size, NUMA_NO_NODE,
|
|
|
GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO);
|
|
|
}
|
|
|
EXPORT_SYMBOL(vzalloc);
|
|
@@ -1781,7 +1783,8 @@ void *vmalloc_user(unsigned long size)
|
|
|
|
|
|
ret = __vmalloc_node(size, SHMLBA,
|
|
|
GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO,
|
|
|
- PAGE_KERNEL, -1, __builtin_return_address(0));
|
|
|
+ PAGE_KERNEL, NUMA_NO_NODE,
|
|
|
+ __builtin_return_address(0));
|
|
|
if (ret) {
|
|
|
area = find_vm_area(ret);
|
|
|
area->flags |= VM_USERMAP;
|
|
@@ -1846,7 +1849,7 @@ EXPORT_SYMBOL(vzalloc_node);
|
|
|
void *vmalloc_exec(unsigned long size)
|
|
|
{
|
|
|
return __vmalloc_node(size, 1, GFP_KERNEL | __GFP_HIGHMEM, PAGE_KERNEL_EXEC,
|
|
|
- -1, __builtin_return_address(0));
|
|
|
+ NUMA_NO_NODE, __builtin_return_address(0));
|
|
|
}
|
|
|
|
|
|
#if defined(CONFIG_64BIT) && defined(CONFIG_ZONE_DMA32)
|
|
@@ -1867,7 +1870,7 @@ void *vmalloc_exec(unsigned long size)
|
|
|
void *vmalloc_32(unsigned long size)
|
|
|
{
|
|
|
return __vmalloc_node(size, 1, GFP_VMALLOC32, PAGE_KERNEL,
|
|
|
- -1, __builtin_return_address(0));
|
|
|
+ NUMA_NO_NODE, __builtin_return_address(0));
|
|
|
}
|
|
|
EXPORT_SYMBOL(vmalloc_32);
|
|
|
|
|
@@ -1884,7 +1887,7 @@ void *vmalloc_32_user(unsigned long size)
|
|
|
void *ret;
|
|
|
|
|
|
ret = __vmalloc_node(size, 1, GFP_VMALLOC32 | __GFP_ZERO, PAGE_KERNEL,
|
|
|
- -1, __builtin_return_address(0));
|
|
|
+ NUMA_NO_NODE, __builtin_return_address(0));
|
|
|
if (ret) {
|
|
|
area = find_vm_area(ret);
|
|
|
area->flags |= VM_USERMAP;
|