linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
* [patch] mm: use NUMA_NO_NODE
@ 2013-02-04 21:57 David Rientjes
  2013-02-05  0:21 ` Yasuaki Ishimatsu
  2013-02-05  0:24 ` Yasuaki Ishimatsu
  0 siblings, 2 replies; 4+ messages in thread
From: David Rientjes @ 2013-02-04 21:57 UTC (permalink / raw)
  To: Andrew Morton; +Cc: linux-mm

Make a sweep through mm/ and convert code that uses -1 directly to using 
the more appropriate NUMA_NO_NODE.

Signed-off-by: David Rientjes <rientjes@google.com>
---
 mm/dmapool.c     |  2 +-
 mm/huge_memory.c |  4 ++--
 mm/mempolicy.c   | 10 +++++-----
 mm/page_alloc.c  |  2 +-
 mm/vmalloc.c     | 33 ++++++++++++++++++---------------
 5 files changed, 27 insertions(+), 24 deletions(-)

diff --git a/mm/dmapool.c b/mm/dmapool.c
index 668f263..6a402c8 100644
--- a/mm/dmapool.c
+++ b/mm/dmapool.c
@@ -157,7 +157,7 @@ struct dma_pool *dma_pool_create(const char *name, struct device *dev,
 		return NULL;
 	}
 
-	node = WARN_ON(!dev) ? -1 : dev_to_node(dev);
+	node = WARN_ON(!dev) ? NUMA_NO_NODE : dev_to_node(dev);
 
 	retval = kmalloc_node(sizeof(*retval), GFP_KERNEL, node);
 	if (!retval)
diff --git a/mm/huge_memory.c b/mm/huge_memory.c
index c63a21d..d41fa11 100644
--- a/mm/huge_memory.c
+++ b/mm/huge_memory.c
@@ -2376,7 +2376,7 @@ static int khugepaged_scan_pmd(struct mm_struct *mm,
 	struct page *page;
 	unsigned long _address;
 	spinlock_t *ptl;
-	int node = -1;
+	int node = NUMA_NO_NODE;
 
 	VM_BUG_ON(address & ~HPAGE_PMD_MASK);
 
@@ -2406,7 +2406,7 @@ static int khugepaged_scan_pmd(struct mm_struct *mm,
 		 * be more sophisticated and look at more pages,
 		 * but isn't for now.
 		 */
-		if (node == -1)
+		if (node == NUMA_NO_NODE)
 			node = page_to_nid(page);
 		VM_BUG_ON(PageCompound(page));
 		if (!PageLRU(page) || PageLocked(page) || !PageAnon(page))
diff --git a/mm/mempolicy.c b/mm/mempolicy.c
index 6f7979c..0a10d40 100644
--- a/mm/mempolicy.c
+++ b/mm/mempolicy.c
@@ -26,7 +26,7 @@
  *                the allocation to memory nodes instead
  *
  * preferred       Try a specific node first before normal fallback.
- *                As a special case node -1 here means do the allocation
+ *                As a special case NUMA_NO_NODE here means do the allocation
  *                on the local CPU. This is normally identical to default,
  *                but useful to set in a VMA when you have a non default
  *                process policy.
@@ -127,7 +127,7 @@ static struct mempolicy *get_task_policy(struct task_struct *p)
 
 	if (!pol) {
 		node = numa_node_id();
-		if (node != -1)
+		if (node != NUMA_NO_NODE)
 			pol = &preferred_node_policy[node];
 
 		/* preferred_node_policy is not initialised early in boot */
@@ -258,7 +258,7 @@ static struct mempolicy *mpol_new(unsigned short mode, unsigned short flags,
 	struct mempolicy *policy;
 
 	pr_debug("setting mode %d flags %d nodes[0] %lx\n",
-		 mode, flags, nodes ? nodes_addr(*nodes)[0] : -1);
+		 mode, flags, nodes ? nodes_addr(*nodes)[0] : NUMA_NO_NODE);
 
 	if (mode == MPOL_DEFAULT) {
 		if (nodes && !nodes_empty(*nodes))
@@ -1223,7 +1223,7 @@ static long do_mbind(unsigned long start, unsigned long len,
 
 	pr_debug("mbind %lx-%lx mode:%d flags:%d nodes:%lx\n",
 		 start, start + len, mode, mode_flags,
-		 nmask ? nodes_addr(*nmask)[0] : -1);
+		 nmask ? nodes_addr(*nmask)[0] : NUMA_NO_NODE);
 
 	if (flags & (MPOL_MF_MOVE | MPOL_MF_MOVE_ALL)) {
 
@@ -2491,7 +2491,7 @@ int mpol_set_shared_policy(struct shared_policy *info,
 		 vma->vm_pgoff,
 		 sz, npol ? npol->mode : -1,
 		 npol ? npol->flags : -1,
-		 npol ? nodes_addr(npol->v.nodes)[0] : -1);
+		 npol ? nodes_addr(npol->v.nodes)[0] : NUMA_NO_NODE);
 
 	if (npol) {
 		new = sp_alloc(vma->vm_pgoff, vma->vm_pgoff + sz, npol);
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index 087845c..35d4714 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -3246,7 +3246,7 @@ static int find_next_best_node(int node, nodemask_t *used_node_mask)
 {
 	int n, val;
 	int min_val = INT_MAX;
-	int best_node = -1;
+	int best_node = NUMA_NO_NODE;
 	const struct cpumask *tmp = cpumask_of_node(0);
 
 	/* Use the local node if we haven't already */
diff --git a/mm/vmalloc.c b/mm/vmalloc.c
index 5123a16..0f751f2 100644
--- a/mm/vmalloc.c
+++ b/mm/vmalloc.c
@@ -1376,8 +1376,8 @@ static struct vm_struct *__get_vm_area_node(unsigned long size,
 struct vm_struct *__get_vm_area(unsigned long size, unsigned long flags,
 				unsigned long start, unsigned long end)
 {
-	return __get_vm_area_node(size, 1, flags, start, end, -1, GFP_KERNEL,
-						__builtin_return_address(0));
+	return __get_vm_area_node(size, 1, flags, start, end, NUMA_NO_NODE,
+				  GFP_KERNEL, __builtin_return_address(0));
 }
 EXPORT_SYMBOL_GPL(__get_vm_area);
 
@@ -1385,8 +1385,8 @@ struct vm_struct *__get_vm_area_caller(unsigned long size, unsigned long flags,
 				       unsigned long start, unsigned long end,
 				       const void *caller)
 {
-	return __get_vm_area_node(size, 1, flags, start, end, -1, GFP_KERNEL,
-				  caller);
+	return __get_vm_area_node(size, 1, flags, start, end, NUMA_NO_NODE,
+				  GFP_KERNEL, caller);
 }
 
 /**
@@ -1401,14 +1401,15 @@ struct vm_struct *__get_vm_area_caller(unsigned long size, unsigned long flags,
 struct vm_struct *get_vm_area(unsigned long size, unsigned long flags)
 {
 	return __get_vm_area_node(size, 1, flags, VMALLOC_START, VMALLOC_END,
-				-1, GFP_KERNEL, __builtin_return_address(0));
+				  NUMA_NO_NODE, GFP_KERNEL,
+				  __builtin_return_address(0));
 }
 
 struct vm_struct *get_vm_area_caller(unsigned long size, unsigned long flags,
 				const void *caller)
 {
 	return __get_vm_area_node(size, 1, flags, VMALLOC_START, VMALLOC_END,
-						-1, GFP_KERNEL, caller);
+				  NUMA_NO_NODE, GFP_KERNEL, caller);
 }
 
 /**
@@ -1650,7 +1651,7 @@ fail:
  *	@end:		vm area range end
  *	@gfp_mask:	flags for the page level allocator
  *	@prot:		protection mask for the allocated pages
- *	@node:		node to use for allocation or -1
+ *	@node:		node to use for allocation or NUMA_NO_NODE
  *	@caller:	caller's return address
  *
  *	Allocate enough pages to cover @size from the page level
@@ -1706,7 +1707,7 @@ fail:
  *	@align:		desired alignment
  *	@gfp_mask:	flags for the page level allocator
  *	@prot:		protection mask for the allocated pages
- *	@node:		node to use for allocation or -1
+ *	@node:		node to use for allocation or NUMA_NO_NODE
  *	@caller:	caller's return address
  *
  *	Allocate enough pages to cover @size from the page level
@@ -1723,7 +1724,7 @@ static void *__vmalloc_node(unsigned long size, unsigned long align,
 
 void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot)
 {
-	return __vmalloc_node(size, 1, gfp_mask, prot, -1,
+	return __vmalloc_node(size, 1, gfp_mask, prot, NUMA_NO_NODE,
 				__builtin_return_address(0));
 }
 EXPORT_SYMBOL(__vmalloc);
@@ -1746,7 +1747,8 @@ static inline void *__vmalloc_node_flags(unsigned long size,
  */
 void *vmalloc(unsigned long size)
 {
-	return __vmalloc_node_flags(size, -1, GFP_KERNEL | __GFP_HIGHMEM);
+	return __vmalloc_node_flags(size, NUMA_NO_NODE,
+				    GFP_KERNEL | __GFP_HIGHMEM);
 }
 EXPORT_SYMBOL(vmalloc);
 
@@ -1762,7 +1764,7 @@ EXPORT_SYMBOL(vmalloc);
  */
 void *vzalloc(unsigned long size)
 {
-	return __vmalloc_node_flags(size, -1,
+	return __vmalloc_node_flags(size, NUMA_NO_NODE,
 				GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO);
 }
 EXPORT_SYMBOL(vzalloc);
@@ -1781,7 +1783,8 @@ void *vmalloc_user(unsigned long size)
 
 	ret = __vmalloc_node(size, SHMLBA,
 			     GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO,
-			     PAGE_KERNEL, -1, __builtin_return_address(0));
+			     PAGE_KERNEL, NUMA_NO_NODE,
+			     __builtin_return_address(0));
 	if (ret) {
 		area = find_vm_area(ret);
 		area->flags |= VM_USERMAP;
@@ -1846,7 +1849,7 @@ EXPORT_SYMBOL(vzalloc_node);
 void *vmalloc_exec(unsigned long size)
 {
 	return __vmalloc_node(size, 1, GFP_KERNEL | __GFP_HIGHMEM, PAGE_KERNEL_EXEC,
-			      -1, __builtin_return_address(0));
+			      NUMA_NO_NODE, __builtin_return_address(0));
 }
 
 #if defined(CONFIG_64BIT) && defined(CONFIG_ZONE_DMA32)
@@ -1867,7 +1870,7 @@ void *vmalloc_exec(unsigned long size)
 void *vmalloc_32(unsigned long size)
 {
 	return __vmalloc_node(size, 1, GFP_VMALLOC32, PAGE_KERNEL,
-			      -1, __builtin_return_address(0));
+			      NUMA_NO_NODE, __builtin_return_address(0));
 }
 EXPORT_SYMBOL(vmalloc_32);
 
@@ -1884,7 +1887,7 @@ void *vmalloc_32_user(unsigned long size)
 	void *ret;
 
 	ret = __vmalloc_node(size, 1, GFP_VMALLOC32 | __GFP_ZERO, PAGE_KERNEL,
-			     -1, __builtin_return_address(0));
+			     NUMA_NO_NODE, __builtin_return_address(0));
 	if (ret) {
 		area = find_vm_area(ret);
 		area->flags |= VM_USERMAP;

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [patch] mm: use NUMA_NO_NODE
  2013-02-04 21:57 [patch] mm: use NUMA_NO_NODE David Rientjes
@ 2013-02-05  0:21 ` Yasuaki Ishimatsu
  2013-02-05  0:24 ` Yasuaki Ishimatsu
  1 sibling, 0 replies; 4+ messages in thread
From: Yasuaki Ishimatsu @ 2013-02-05  0:21 UTC (permalink / raw)
  To: David Rientjes; +Cc: Andrew Morton, linux-mm

2013/02/05 6:57, David Rientjes wrote:
> Make a sweep through mm/ and convert code that uses -1 directly to using
> the more appropriate NUMA_NO_NODE.
>
> Signed-off-by: David Rientjes <rientjes@google.com>
> ---

Reviewed-by: Yasuaki Ishimatsu <isimatu.yasuaki@jp.fujitsu.com>

Thanks,
Yasuaki Ishimatsu

>   mm/dmapool.c     |  2 +-
>   mm/huge_memory.c |  4 ++--
>   mm/mempolicy.c   | 10 +++++-----
>   mm/page_alloc.c  |  2 +-
>   mm/vmalloc.c     | 33 ++++++++++++++++++---------------
>   5 files changed, 27 insertions(+), 24 deletions(-)
>
> diff --git a/mm/dmapool.c b/mm/dmapool.c
> index 668f263..6a402c8 100644
> --- a/mm/dmapool.c
> +++ b/mm/dmapool.c
> @@ -157,7 +157,7 @@ struct dma_pool *dma_pool_create(const char *name, struct device *dev,
>   		return NULL;
>   	}
>
> -	node = WARN_ON(!dev) ? -1 : dev_to_node(dev);
> +	node = WARN_ON(!dev) ? NUMA_NO_NODE : dev_to_node(dev);
>
>   	retval = kmalloc_node(sizeof(*retval), GFP_KERNEL, node);
>   	if (!retval)
> diff --git a/mm/huge_memory.c b/mm/huge_memory.c
> index c63a21d..d41fa11 100644
> --- a/mm/huge_memory.c
> +++ b/mm/huge_memory.c
> @@ -2376,7 +2376,7 @@ static int khugepaged_scan_pmd(struct mm_struct *mm,
>   	struct page *page;
>   	unsigned long _address;
>   	spinlock_t *ptl;
> -	int node = -1;
> +	int node = NUMA_NO_NODE;
>
>   	VM_BUG_ON(address & ~HPAGE_PMD_MASK);
>
> @@ -2406,7 +2406,7 @@ static int khugepaged_scan_pmd(struct mm_struct *mm,
>   		 * be more sophisticated and look at more pages,
>   		 * but isn't for now.
>   		 */
> -		if (node == -1)
> +		if (node == NUMA_NO_NODE)
>   			node = page_to_nid(page);
>   		VM_BUG_ON(PageCompound(page));
>   		if (!PageLRU(page) || PageLocked(page) || !PageAnon(page))
> diff --git a/mm/mempolicy.c b/mm/mempolicy.c
> index 6f7979c..0a10d40 100644
> --- a/mm/mempolicy.c
> +++ b/mm/mempolicy.c
> @@ -26,7 +26,7 @@
>    *                the allocation to memory nodes instead
>    *
>    * preferred       Try a specific node first before normal fallback.
> - *                As a special case node -1 here means do the allocation
> + *                As a special case NUMA_NO_NODE here means do the allocation
>    *                on the local CPU. This is normally identical to default,
>    *                but useful to set in a VMA when you have a non default
>    *                process policy.
> @@ -127,7 +127,7 @@ static struct mempolicy *get_task_policy(struct task_struct *p)
>
>   	if (!pol) {
>   		node = numa_node_id();
> -		if (node != -1)
> +		if (node != NUMA_NO_NODE)
>   			pol = &preferred_node_policy[node];
>
>   		/* preferred_node_policy is not initialised early in boot */
> @@ -258,7 +258,7 @@ static struct mempolicy *mpol_new(unsigned short mode, unsigned short flags,
>   	struct mempolicy *policy;
>
>   	pr_debug("setting mode %d flags %d nodes[0] %lx\n",
> -		 mode, flags, nodes ? nodes_addr(*nodes)[0] : -1);
> +		 mode, flags, nodes ? nodes_addr(*nodes)[0] : NUMA_NO_NODE);
>
>   	if (mode == MPOL_DEFAULT) {
>   		if (nodes && !nodes_empty(*nodes))
> @@ -1223,7 +1223,7 @@ static long do_mbind(unsigned long start, unsigned long len,
>
>   	pr_debug("mbind %lx-%lx mode:%d flags:%d nodes:%lx\n",
>   		 start, start + len, mode, mode_flags,
> -		 nmask ? nodes_addr(*nmask)[0] : -1);
> +		 nmask ? nodes_addr(*nmask)[0] : NUMA_NO_NODE);
>
>   	if (flags & (MPOL_MF_MOVE | MPOL_MF_MOVE_ALL)) {
>
> @@ -2491,7 +2491,7 @@ int mpol_set_shared_policy(struct shared_policy *info,
>   		 vma->vm_pgoff,
>   		 sz, npol ? npol->mode : -1,
>   		 npol ? npol->flags : -1,
> -		 npol ? nodes_addr(npol->v.nodes)[0] : -1);
> +		 npol ? nodes_addr(npol->v.nodes)[0] : NUMA_NO_NODE);
>
>   	if (npol) {
>   		new = sp_alloc(vma->vm_pgoff, vma->vm_pgoff + sz, npol);
> diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> index 087845c..35d4714 100644
> --- a/mm/page_alloc.c
> +++ b/mm/page_alloc.c
> @@ -3246,7 +3246,7 @@ static int find_next_best_node(int node, nodemask_t *used_node_mask)
>   {
>   	int n, val;
>   	int min_val = INT_MAX;
> -	int best_node = -1;
> +	int best_node = NUMA_NO_NODE;
>   	const struct cpumask *tmp = cpumask_of_node(0);
>
>   	/* Use the local node if we haven't already */
> diff --git a/mm/vmalloc.c b/mm/vmalloc.c
> index 5123a16..0f751f2 100644
> --- a/mm/vmalloc.c
> +++ b/mm/vmalloc.c
> @@ -1376,8 +1376,8 @@ static struct vm_struct *__get_vm_area_node(unsigned long size,
>   struct vm_struct *__get_vm_area(unsigned long size, unsigned long flags,
>   				unsigned long start, unsigned long end)
>   {
> -	return __get_vm_area_node(size, 1, flags, start, end, -1, GFP_KERNEL,
> -						__builtin_return_address(0));
> +	return __get_vm_area_node(size, 1, flags, start, end, NUMA_NO_NODE,
> +				  GFP_KERNEL, __builtin_return_address(0));
>   }
>   EXPORT_SYMBOL_GPL(__get_vm_area);
>
> @@ -1385,8 +1385,8 @@ struct vm_struct *__get_vm_area_caller(unsigned long size, unsigned long flags,
>   				       unsigned long start, unsigned long end,
>   				       const void *caller)
>   {
> -	return __get_vm_area_node(size, 1, flags, start, end, -1, GFP_KERNEL,
> -				  caller);
> +	return __get_vm_area_node(size, 1, flags, start, end, NUMA_NO_NODE,
> +				  GFP_KERNEL, caller);
>   }
>
>   /**
> @@ -1401,14 +1401,15 @@ struct vm_struct *__get_vm_area_caller(unsigned long size, unsigned long flags,
>   struct vm_struct *get_vm_area(unsigned long size, unsigned long flags)
>   {
>   	return __get_vm_area_node(size, 1, flags, VMALLOC_START, VMALLOC_END,
> -				-1, GFP_KERNEL, __builtin_return_address(0));
> +				  NUMA_NO_NODE, GFP_KERNEL,
> +				  __builtin_return_address(0));
>   }
>
>   struct vm_struct *get_vm_area_caller(unsigned long size, unsigned long flags,
>   				const void *caller)
>   {
>   	return __get_vm_area_node(size, 1, flags, VMALLOC_START, VMALLOC_END,
> -						-1, GFP_KERNEL, caller);
> +				  NUMA_NO_NODE, GFP_KERNEL, caller);
>   }
>
>   /**
> @@ -1650,7 +1651,7 @@ fail:
>    *	@end:		vm area range end
>    *	@gfp_mask:	flags for the page level allocator
>    *	@prot:		protection mask for the allocated pages
> - *	@node:		node to use for allocation or -1
> + *	@node:		node to use for allocation or NUMA_NO_NODE
>    *	@caller:	caller's return address
>    *
>    *	Allocate enough pages to cover @size from the page level
> @@ -1706,7 +1707,7 @@ fail:
>    *	@align:		desired alignment
>    *	@gfp_mask:	flags for the page level allocator
>    *	@prot:		protection mask for the allocated pages
> - *	@node:		node to use for allocation or -1
> + *	@node:		node to use for allocation or NUMA_NO_NODE
>    *	@caller:	caller's return address
>    *
>    *	Allocate enough pages to cover @size from the page level
> @@ -1723,7 +1724,7 @@ static void *__vmalloc_node(unsigned long size, unsigned long align,
>
>   void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot)
>   {
> -	return __vmalloc_node(size, 1, gfp_mask, prot, -1,
> +	return __vmalloc_node(size, 1, gfp_mask, prot, NUMA_NO_NODE,
>   				__builtin_return_address(0));
>   }
>   EXPORT_SYMBOL(__vmalloc);
> @@ -1746,7 +1747,8 @@ static inline void *__vmalloc_node_flags(unsigned long size,
>    */
>   void *vmalloc(unsigned long size)
>   {
> -	return __vmalloc_node_flags(size, -1, GFP_KERNEL | __GFP_HIGHMEM);
> +	return __vmalloc_node_flags(size, NUMA_NO_NODE,
> +				    GFP_KERNEL | __GFP_HIGHMEM);
>   }
>   EXPORT_SYMBOL(vmalloc);
>
> @@ -1762,7 +1764,7 @@ EXPORT_SYMBOL(vmalloc);
>    */
>   void *vzalloc(unsigned long size)
>   {
> -	return __vmalloc_node_flags(size, -1,
> +	return __vmalloc_node_flags(size, NUMA_NO_NODE,
>   				GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO);
>   }
>   EXPORT_SYMBOL(vzalloc);
> @@ -1781,7 +1783,8 @@ void *vmalloc_user(unsigned long size)
>
>   	ret = __vmalloc_node(size, SHMLBA,
>   			     GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO,
> -			     PAGE_KERNEL, -1, __builtin_return_address(0));
> +			     PAGE_KERNEL, NUMA_NO_NODE,
> +			     __builtin_return_address(0));
>   	if (ret) {
>   		area = find_vm_area(ret);
>   		area->flags |= VM_USERMAP;
> @@ -1846,7 +1849,7 @@ EXPORT_SYMBOL(vzalloc_node);
>   void *vmalloc_exec(unsigned long size)
>   {
>   	return __vmalloc_node(size, 1, GFP_KERNEL | __GFP_HIGHMEM, PAGE_KERNEL_EXEC,
> -			      -1, __builtin_return_address(0));
> +			      NUMA_NO_NODE, __builtin_return_address(0));
>   }
>
>   #if defined(CONFIG_64BIT) && defined(CONFIG_ZONE_DMA32)
> @@ -1867,7 +1870,7 @@ void *vmalloc_exec(unsigned long size)
>   void *vmalloc_32(unsigned long size)
>   {
>   	return __vmalloc_node(size, 1, GFP_VMALLOC32, PAGE_KERNEL,
> -			      -1, __builtin_return_address(0));
> +			      NUMA_NO_NODE, __builtin_return_address(0));
>   }
>   EXPORT_SYMBOL(vmalloc_32);
>
> @@ -1884,7 +1887,7 @@ void *vmalloc_32_user(unsigned long size)
>   	void *ret;
>
>   	ret = __vmalloc_node(size, 1, GFP_VMALLOC32 | __GFP_ZERO, PAGE_KERNEL,
> -			     -1, __builtin_return_address(0));
> +			     NUMA_NO_NODE, __builtin_return_address(0));
>   	if (ret) {
>   		area = find_vm_area(ret);
>   		area->flags |= VM_USERMAP;
>
> --
> To unsubscribe, send a message with 'unsubscribe linux-mm' in
> the body to majordomo@kvack.org.  For more info on Linux MM,
> see: http://www.linux-mm.org/ .
> Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
>


--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [patch] mm: use NUMA_NO_NODE
  2013-02-04 21:57 [patch] mm: use NUMA_NO_NODE David Rientjes
  2013-02-05  0:21 ` Yasuaki Ishimatsu
@ 2013-02-05  0:24 ` Yasuaki Ishimatsu
  2013-02-05  0:59   ` Yasuaki Ishimatsu
  1 sibling, 1 reply; 4+ messages in thread
From: Yasuaki Ishimatsu @ 2013-02-05  0:24 UTC (permalink / raw)
  To: David Rientjes; +Cc: Andrew Morton, linux-mm

2013/02/05 6:57, David Rientjes wrote:
> Make a sweep through mm/ and convert code that uses -1 directly to using
> the more appropriate NUMA_NO_NODE.
>
> Signed-off-by: David Rientjes <rientjes@google.com>
> ---

Reviewed-by: Yasauaki Ishimatsu <isimatu.yasuaki@jp.fujitsu.com>

Thanks,
Yasuaki Ishimatsu

>   mm/dmapool.c     |  2 +-
>   mm/huge_memory.c |  4 ++--
>   mm/mempolicy.c   | 10 +++++-----
>   mm/page_alloc.c  |  2 +-
>   mm/vmalloc.c     | 33 ++++++++++++++++++---------------
>   5 files changed, 27 insertions(+), 24 deletions(-)
>
> diff --git a/mm/dmapool.c b/mm/dmapool.c
> index 668f263..6a402c8 100644
> --- a/mm/dmapool.c
> +++ b/mm/dmapool.c
> @@ -157,7 +157,7 @@ struct dma_pool *dma_pool_create(const char *name, struct device *dev,
>   		return NULL;
>   	}
>
> -	node = WARN_ON(!dev) ? -1 : dev_to_node(dev);
> +	node = WARN_ON(!dev) ? NUMA_NO_NODE : dev_to_node(dev);
>
>   	retval = kmalloc_node(sizeof(*retval), GFP_KERNEL, node);
>   	if (!retval)
> diff --git a/mm/huge_memory.c b/mm/huge_memory.c
> index c63a21d..d41fa11 100644
> --- a/mm/huge_memory.c
> +++ b/mm/huge_memory.c
> @@ -2376,7 +2376,7 @@ static int khugepaged_scan_pmd(struct mm_struct *mm,
>   	struct page *page;
>   	unsigned long _address;
>   	spinlock_t *ptl;
> -	int node = -1;
> +	int node = NUMA_NO_NODE;
>
>   	VM_BUG_ON(address & ~HPAGE_PMD_MASK);
>
> @@ -2406,7 +2406,7 @@ static int khugepaged_scan_pmd(struct mm_struct *mm,
>   		 * be more sophisticated and look at more pages,
>   		 * but isn't for now.
>   		 */
> -		if (node == -1)
> +		if (node == NUMA_NO_NODE)
>   			node = page_to_nid(page);
>   		VM_BUG_ON(PageCompound(page));
>   		if (!PageLRU(page) || PageLocked(page) || !PageAnon(page))
> diff --git a/mm/mempolicy.c b/mm/mempolicy.c
> index 6f7979c..0a10d40 100644
> --- a/mm/mempolicy.c
> +++ b/mm/mempolicy.c
> @@ -26,7 +26,7 @@
>    *                the allocation to memory nodes instead
>    *
>    * preferred       Try a specific node first before normal fallback.
> - *                As a special case node -1 here means do the allocation
> + *                As a special case NUMA_NO_NODE here means do the allocation
>    *                on the local CPU. This is normally identical to default,
>    *                but useful to set in a VMA when you have a non default
>    *                process policy.
> @@ -127,7 +127,7 @@ static struct mempolicy *get_task_policy(struct task_struct *p)
>
>   	if (!pol) {
>   		node = numa_node_id();
> -		if (node != -1)
> +		if (node != NUMA_NO_NODE)
>   			pol = &preferred_node_policy[node];
>
>   		/* preferred_node_policy is not initialised early in boot */
> @@ -258,7 +258,7 @@ static struct mempolicy *mpol_new(unsigned short mode, unsigned short flags,
>   	struct mempolicy *policy;
>
>   	pr_debug("setting mode %d flags %d nodes[0] %lx\n",
> -		 mode, flags, nodes ? nodes_addr(*nodes)[0] : -1);
> +		 mode, flags, nodes ? nodes_addr(*nodes)[0] : NUMA_NO_NODE);
>
>   	if (mode == MPOL_DEFAULT) {
>   		if (nodes && !nodes_empty(*nodes))
> @@ -1223,7 +1223,7 @@ static long do_mbind(unsigned long start, unsigned long len,
>
>   	pr_debug("mbind %lx-%lx mode:%d flags:%d nodes:%lx\n",
>   		 start, start + len, mode, mode_flags,
> -		 nmask ? nodes_addr(*nmask)[0] : -1);
> +		 nmask ? nodes_addr(*nmask)[0] : NUMA_NO_NODE);
>
>   	if (flags & (MPOL_MF_MOVE | MPOL_MF_MOVE_ALL)) {
>
> @@ -2491,7 +2491,7 @@ int mpol_set_shared_policy(struct shared_policy *info,
>   		 vma->vm_pgoff,
>   		 sz, npol ? npol->mode : -1,
>   		 npol ? npol->flags : -1,
> -		 npol ? nodes_addr(npol->v.nodes)[0] : -1);
> +		 npol ? nodes_addr(npol->v.nodes)[0] : NUMA_NO_NODE);
>
>   	if (npol) {
>   		new = sp_alloc(vma->vm_pgoff, vma->vm_pgoff + sz, npol);
> diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> index 087845c..35d4714 100644
> --- a/mm/page_alloc.c
> +++ b/mm/page_alloc.c
> @@ -3246,7 +3246,7 @@ static int find_next_best_node(int node, nodemask_t *used_node_mask)
>   {
>   	int n, val;
>   	int min_val = INT_MAX;
> -	int best_node = -1;
> +	int best_node = NUMA_NO_NODE;
>   	const struct cpumask *tmp = cpumask_of_node(0);
>
>   	/* Use the local node if we haven't already */
> diff --git a/mm/vmalloc.c b/mm/vmalloc.c
> index 5123a16..0f751f2 100644
> --- a/mm/vmalloc.c
> +++ b/mm/vmalloc.c
> @@ -1376,8 +1376,8 @@ static struct vm_struct *__get_vm_area_node(unsigned long size,
>   struct vm_struct *__get_vm_area(unsigned long size, unsigned long flags,
>   				unsigned long start, unsigned long end)
>   {
> -	return __get_vm_area_node(size, 1, flags, start, end, -1, GFP_KERNEL,
> -						__builtin_return_address(0));
> +	return __get_vm_area_node(size, 1, flags, start, end, NUMA_NO_NODE,
> +				  GFP_KERNEL, __builtin_return_address(0));
>   }
>   EXPORT_SYMBOL_GPL(__get_vm_area);
>
> @@ -1385,8 +1385,8 @@ struct vm_struct *__get_vm_area_caller(unsigned long size, unsigned long flags,
>   				       unsigned long start, unsigned long end,
>   				       const void *caller)
>   {
> -	return __get_vm_area_node(size, 1, flags, start, end, -1, GFP_KERNEL,
> -				  caller);
> +	return __get_vm_area_node(size, 1, flags, start, end, NUMA_NO_NODE,
> +				  GFP_KERNEL, caller);
>   }
>
>   /**
> @@ -1401,14 +1401,15 @@ struct vm_struct *__get_vm_area_caller(unsigned long size, unsigned long flags,
>   struct vm_struct *get_vm_area(unsigned long size, unsigned long flags)
>   {
>   	return __get_vm_area_node(size, 1, flags, VMALLOC_START, VMALLOC_END,
> -				-1, GFP_KERNEL, __builtin_return_address(0));
> +				  NUMA_NO_NODE, GFP_KERNEL,
> +				  __builtin_return_address(0));
>   }
>
>   struct vm_struct *get_vm_area_caller(unsigned long size, unsigned long flags,
>   				const void *caller)
>   {
>   	return __get_vm_area_node(size, 1, flags, VMALLOC_START, VMALLOC_END,
> -						-1, GFP_KERNEL, caller);
> +				  NUMA_NO_NODE, GFP_KERNEL, caller);
>   }
>
>   /**
> @@ -1650,7 +1651,7 @@ fail:
>    *	@end:		vm area range end
>    *	@gfp_mask:	flags for the page level allocator
>    *	@prot:		protection mask for the allocated pages
> - *	@node:		node to use for allocation or -1
> + *	@node:		node to use for allocation or NUMA_NO_NODE
>    *	@caller:	caller's return address
>    *
>    *	Allocate enough pages to cover @size from the page level
> @@ -1706,7 +1707,7 @@ fail:
>    *	@align:		desired alignment
>    *	@gfp_mask:	flags for the page level allocator
>    *	@prot:		protection mask for the allocated pages
> - *	@node:		node to use for allocation or -1
> + *	@node:		node to use for allocation or NUMA_NO_NODE
>    *	@caller:	caller's return address
>    *
>    *	Allocate enough pages to cover @size from the page level
> @@ -1723,7 +1724,7 @@ static void *__vmalloc_node(unsigned long size, unsigned long align,
>
>   void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot)
>   {
> -	return __vmalloc_node(size, 1, gfp_mask, prot, -1,
> +	return __vmalloc_node(size, 1, gfp_mask, prot, NUMA_NO_NODE,
>   				__builtin_return_address(0));
>   }
>   EXPORT_SYMBOL(__vmalloc);
> @@ -1746,7 +1747,8 @@ static inline void *__vmalloc_node_flags(unsigned long size,
>    */
>   void *vmalloc(unsigned long size)
>   {
> -	return __vmalloc_node_flags(size, -1, GFP_KERNEL | __GFP_HIGHMEM);
> +	return __vmalloc_node_flags(size, NUMA_NO_NODE,
> +				    GFP_KERNEL | __GFP_HIGHMEM);
>   }
>   EXPORT_SYMBOL(vmalloc);
>
> @@ -1762,7 +1764,7 @@ EXPORT_SYMBOL(vmalloc);
>    */
>   void *vzalloc(unsigned long size)
>   {
> -	return __vmalloc_node_flags(size, -1,
> +	return __vmalloc_node_flags(size, NUMA_NO_NODE,
>   				GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO);
>   }
>   EXPORT_SYMBOL(vzalloc);
> @@ -1781,7 +1783,8 @@ void *vmalloc_user(unsigned long size)
>
>   	ret = __vmalloc_node(size, SHMLBA,
>   			     GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO,
> -			     PAGE_KERNEL, -1, __builtin_return_address(0));
> +			     PAGE_KERNEL, NUMA_NO_NODE,
> +			     __builtin_return_address(0));
>   	if (ret) {
>   		area = find_vm_area(ret);
>   		area->flags |= VM_USERMAP;
> @@ -1846,7 +1849,7 @@ EXPORT_SYMBOL(vzalloc_node);
>   void *vmalloc_exec(unsigned long size)
>   {
>   	return __vmalloc_node(size, 1, GFP_KERNEL | __GFP_HIGHMEM, PAGE_KERNEL_EXEC,
> -			      -1, __builtin_return_address(0));
> +			      NUMA_NO_NODE, __builtin_return_address(0));
>   }
>
>   #if defined(CONFIG_64BIT) && defined(CONFIG_ZONE_DMA32)
> @@ -1867,7 +1870,7 @@ void *vmalloc_exec(unsigned long size)
>   void *vmalloc_32(unsigned long size)
>   {
>   	return __vmalloc_node(size, 1, GFP_VMALLOC32, PAGE_KERNEL,
> -			      -1, __builtin_return_address(0));
> +			      NUMA_NO_NODE, __builtin_return_address(0));
>   }
>   EXPORT_SYMBOL(vmalloc_32);
>
> @@ -1884,7 +1887,7 @@ void *vmalloc_32_user(unsigned long size)
>   	void *ret;
>
>   	ret = __vmalloc_node(size, 1, GFP_VMALLOC32 | __GFP_ZERO, PAGE_KERNEL,
> -			     -1, __builtin_return_address(0));
> +			     NUMA_NO_NODE, __builtin_return_address(0));
>   	if (ret) {
>   		area = find_vm_area(ret);
>   		area->flags |= VM_USERMAP;
>
> --
> To unsubscribe, send a message with 'unsubscribe linux-mm' in
> the body to majordomo@kvack.org.  For more info on Linux MM,
> see: http://www.linux-mm.org/ .
> Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
>


--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [patch] mm: use NUMA_NO_NODE
  2013-02-05  0:24 ` Yasuaki Ishimatsu
@ 2013-02-05  0:59   ` Yasuaki Ishimatsu
  0 siblings, 0 replies; 4+ messages in thread
From: Yasuaki Ishimatsu @ 2013-02-05  0:59 UTC (permalink / raw)
  To: David Rientjes; +Cc: Andrew Morton, linux-mm

Sorry about the noise.

Thanks,
Yasuaki Ishimatsu

2013/02/05 9:24, Yasuaki Ishimatsu wrote:
> 2013/02/05 6:57, David Rientjes wrote:
>> Make a sweep through mm/ and convert code that uses -1 directly to using
>> the more appropriate NUMA_NO_NODE.
>>
>> Signed-off-by: David Rientjes <rientjes@google.com>
>> ---
>
> Reviewed-by: Yasauaki Ishimatsu <isimatu.yasuaki@jp.fujitsu.com>
>
> Thanks,
> Yasuaki Ishimatsu
>
>>   mm/dmapool.c     |  2 +-
>>   mm/huge_memory.c |  4 ++--
>>   mm/mempolicy.c   | 10 +++++-----
>>   mm/page_alloc.c  |  2 +-
>>   mm/vmalloc.c     | 33 ++++++++++++++++++---------------
>>   5 files changed, 27 insertions(+), 24 deletions(-)
>>
>> diff --git a/mm/dmapool.c b/mm/dmapool.c
>> index 668f263..6a402c8 100644
>> --- a/mm/dmapool.c
>> +++ b/mm/dmapool.c
>> @@ -157,7 +157,7 @@ struct dma_pool *dma_pool_create(const char *name, struct device *dev,
>>           return NULL;
>>       }
>>
>> -    node = WARN_ON(!dev) ? -1 : dev_to_node(dev);
>> +    node = WARN_ON(!dev) ? NUMA_NO_NODE : dev_to_node(dev);
>>
>>       retval = kmalloc_node(sizeof(*retval), GFP_KERNEL, node);
>>       if (!retval)
>> diff --git a/mm/huge_memory.c b/mm/huge_memory.c
>> index c63a21d..d41fa11 100644
>> --- a/mm/huge_memory.c
>> +++ b/mm/huge_memory.c
>> @@ -2376,7 +2376,7 @@ static int khugepaged_scan_pmd(struct mm_struct *mm,
>>       struct page *page;
>>       unsigned long _address;
>>       spinlock_t *ptl;
>> -    int node = -1;
>> +    int node = NUMA_NO_NODE;
>>
>>       VM_BUG_ON(address & ~HPAGE_PMD_MASK);
>>
>> @@ -2406,7 +2406,7 @@ static int khugepaged_scan_pmd(struct mm_struct *mm,
>>            * be more sophisticated and look at more pages,
>>            * but isn't for now.
>>            */
>> -        if (node == -1)
>> +        if (node == NUMA_NO_NODE)
>>               node = page_to_nid(page);
>>           VM_BUG_ON(PageCompound(page));
>>           if (!PageLRU(page) || PageLocked(page) || !PageAnon(page))
>> diff --git a/mm/mempolicy.c b/mm/mempolicy.c
>> index 6f7979c..0a10d40 100644
>> --- a/mm/mempolicy.c
>> +++ b/mm/mempolicy.c
>> @@ -26,7 +26,7 @@
>>    *                the allocation to memory nodes instead
>>    *
>>    * preferred       Try a specific node first before normal fallback.
>> - *                As a special case node -1 here means do the allocation
>> + *                As a special case NUMA_NO_NODE here means do the allocation
>>    *                on the local CPU. This is normally identical to default,
>>    *                but useful to set in a VMA when you have a non default
>>    *                process policy.
>> @@ -127,7 +127,7 @@ static struct mempolicy *get_task_policy(struct task_struct *p)
>>
>>       if (!pol) {
>>           node = numa_node_id();
>> -        if (node != -1)
>> +        if (node != NUMA_NO_NODE)
>>               pol = &preferred_node_policy[node];
>>
>>           /* preferred_node_policy is not initialised early in boot */
>> @@ -258,7 +258,7 @@ static struct mempolicy *mpol_new(unsigned short mode, unsigned short flags,
>>       struct mempolicy *policy;
>>
>>       pr_debug("setting mode %d flags %d nodes[0] %lx\n",
>> -         mode, flags, nodes ? nodes_addr(*nodes)[0] : -1);
>> +         mode, flags, nodes ? nodes_addr(*nodes)[0] : NUMA_NO_NODE);
>>
>>       if (mode == MPOL_DEFAULT) {
>>           if (nodes && !nodes_empty(*nodes))
>> @@ -1223,7 +1223,7 @@ static long do_mbind(unsigned long start, unsigned long len,
>>
>>       pr_debug("mbind %lx-%lx mode:%d flags:%d nodes:%lx\n",
>>            start, start + len, mode, mode_flags,
>> -         nmask ? nodes_addr(*nmask)[0] : -1);
>> +         nmask ? nodes_addr(*nmask)[0] : NUMA_NO_NODE);
>>
>>       if (flags & (MPOL_MF_MOVE | MPOL_MF_MOVE_ALL)) {
>>
>> @@ -2491,7 +2491,7 @@ int mpol_set_shared_policy(struct shared_policy *info,
>>            vma->vm_pgoff,
>>            sz, npol ? npol->mode : -1,
>>            npol ? npol->flags : -1,
>> -         npol ? nodes_addr(npol->v.nodes)[0] : -1);
>> +         npol ? nodes_addr(npol->v.nodes)[0] : NUMA_NO_NODE);
>>
>>       if (npol) {
>>           new = sp_alloc(vma->vm_pgoff, vma->vm_pgoff + sz, npol);
>> diff --git a/mm/page_alloc.c b/mm/page_alloc.c
>> index 087845c..35d4714 100644
>> --- a/mm/page_alloc.c
>> +++ b/mm/page_alloc.c
>> @@ -3246,7 +3246,7 @@ static int find_next_best_node(int node, nodemask_t *used_node_mask)
>>   {
>>       int n, val;
>>       int min_val = INT_MAX;
>> -    int best_node = -1;
>> +    int best_node = NUMA_NO_NODE;
>>       const struct cpumask *tmp = cpumask_of_node(0);
>>
>>       /* Use the local node if we haven't already */
>> diff --git a/mm/vmalloc.c b/mm/vmalloc.c
>> index 5123a16..0f751f2 100644
>> --- a/mm/vmalloc.c
>> +++ b/mm/vmalloc.c
>> @@ -1376,8 +1376,8 @@ static struct vm_struct *__get_vm_area_node(unsigned long size,
>>   struct vm_struct *__get_vm_area(unsigned long size, unsigned long flags,
>>                   unsigned long start, unsigned long end)
>>   {
>> -    return __get_vm_area_node(size, 1, flags, start, end, -1, GFP_KERNEL,
>> -                        __builtin_return_address(0));
>> +    return __get_vm_area_node(size, 1, flags, start, end, NUMA_NO_NODE,
>> +                  GFP_KERNEL, __builtin_return_address(0));
>>   }
>>   EXPORT_SYMBOL_GPL(__get_vm_area);
>>
>> @@ -1385,8 +1385,8 @@ struct vm_struct *__get_vm_area_caller(unsigned long size, unsigned long flags,
>>                          unsigned long start, unsigned long end,
>>                          const void *caller)
>>   {
>> -    return __get_vm_area_node(size, 1, flags, start, end, -1, GFP_KERNEL,
>> -                  caller);
>> +    return __get_vm_area_node(size, 1, flags, start, end, NUMA_NO_NODE,
>> +                  GFP_KERNEL, caller);
>>   }
>>
>>   /**
>> @@ -1401,14 +1401,15 @@ struct vm_struct *__get_vm_area_caller(unsigned long size, unsigned long flags,
>>   struct vm_struct *get_vm_area(unsigned long size, unsigned long flags)
>>   {
>>       return __get_vm_area_node(size, 1, flags, VMALLOC_START, VMALLOC_END,
>> -                -1, GFP_KERNEL, __builtin_return_address(0));
>> +                  NUMA_NO_NODE, GFP_KERNEL,
>> +                  __builtin_return_address(0));
>>   }
>>
>>   struct vm_struct *get_vm_area_caller(unsigned long size, unsigned long flags,
>>                   const void *caller)
>>   {
>>       return __get_vm_area_node(size, 1, flags, VMALLOC_START, VMALLOC_END,
>> -                        -1, GFP_KERNEL, caller);
>> +                  NUMA_NO_NODE, GFP_KERNEL, caller);
>>   }
>>
>>   /**
>> @@ -1650,7 +1651,7 @@ fail:
>>    *    @end:        vm area range end
>>    *    @gfp_mask:    flags for the page level allocator
>>    *    @prot:        protection mask for the allocated pages
>> - *    @node:        node to use for allocation or -1
>> + *    @node:        node to use for allocation or NUMA_NO_NODE
>>    *    @caller:    caller's return address
>>    *
>>    *    Allocate enough pages to cover @size from the page level
>> @@ -1706,7 +1707,7 @@ fail:
>>    *    @align:        desired alignment
>>    *    @gfp_mask:    flags for the page level allocator
>>    *    @prot:        protection mask for the allocated pages
>> - *    @node:        node to use for allocation or -1
>> + *    @node:        node to use for allocation or NUMA_NO_NODE
>>    *    @caller:    caller's return address
>>    *
>>    *    Allocate enough pages to cover @size from the page level
>> @@ -1723,7 +1724,7 @@ static void *__vmalloc_node(unsigned long size, unsigned long align,
>>
>>   void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot)
>>   {
>> -    return __vmalloc_node(size, 1, gfp_mask, prot, -1,
>> +    return __vmalloc_node(size, 1, gfp_mask, prot, NUMA_NO_NODE,
>>                   __builtin_return_address(0));
>>   }
>>   EXPORT_SYMBOL(__vmalloc);
>> @@ -1746,7 +1747,8 @@ static inline void *__vmalloc_node_flags(unsigned long size,
>>    */
>>   void *vmalloc(unsigned long size)
>>   {
>> -    return __vmalloc_node_flags(size, -1, GFP_KERNEL | __GFP_HIGHMEM);
>> +    return __vmalloc_node_flags(size, NUMA_NO_NODE,
>> +                    GFP_KERNEL | __GFP_HIGHMEM);
>>   }
>>   EXPORT_SYMBOL(vmalloc);
>>
>> @@ -1762,7 +1764,7 @@ EXPORT_SYMBOL(vmalloc);
>>    */
>>   void *vzalloc(unsigned long size)
>>   {
>> -    return __vmalloc_node_flags(size, -1,
>> +    return __vmalloc_node_flags(size, NUMA_NO_NODE,
>>                   GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO);
>>   }
>>   EXPORT_SYMBOL(vzalloc);
>> @@ -1781,7 +1783,8 @@ void *vmalloc_user(unsigned long size)
>>
>>       ret = __vmalloc_node(size, SHMLBA,
>>                    GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO,
>> -                 PAGE_KERNEL, -1, __builtin_return_address(0));
>> +                 PAGE_KERNEL, NUMA_NO_NODE,
>> +                 __builtin_return_address(0));
>>       if (ret) {
>>           area = find_vm_area(ret);
>>           area->flags |= VM_USERMAP;
>> @@ -1846,7 +1849,7 @@ EXPORT_SYMBOL(vzalloc_node);
>>   void *vmalloc_exec(unsigned long size)
>>   {
>>       return __vmalloc_node(size, 1, GFP_KERNEL | __GFP_HIGHMEM, PAGE_KERNEL_EXEC,
>> -                  -1, __builtin_return_address(0));
>> +                  NUMA_NO_NODE, __builtin_return_address(0));
>>   }
>>
>>   #if defined(CONFIG_64BIT) && defined(CONFIG_ZONE_DMA32)
>> @@ -1867,7 +1870,7 @@ void *vmalloc_exec(unsigned long size)
>>   void *vmalloc_32(unsigned long size)
>>   {
>>       return __vmalloc_node(size, 1, GFP_VMALLOC32, PAGE_KERNEL,
>> -                  -1, __builtin_return_address(0));
>> +                  NUMA_NO_NODE, __builtin_return_address(0));
>>   }
>>   EXPORT_SYMBOL(vmalloc_32);
>>
>> @@ -1884,7 +1887,7 @@ void *vmalloc_32_user(unsigned long size)
>>       void *ret;
>>
>>       ret = __vmalloc_node(size, 1, GFP_VMALLOC32 | __GFP_ZERO, PAGE_KERNEL,
>> -                 -1, __builtin_return_address(0));
>> +                 NUMA_NO_NODE, __builtin_return_address(0));
>>       if (ret) {
>>           area = find_vm_area(ret);
>>           area->flags |= VM_USERMAP;
>>
>> --
>> To unsubscribe, send a message with 'unsubscribe linux-mm' in
>> the body to majordomo@kvack.org.  For more info on Linux MM,
>> see: http://www.linux-mm.org/ .
>> Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
>>
>
>
> --
> To unsubscribe, send a message with 'unsubscribe linux-mm' in
> the body to majordomo@kvack.org.  For more info on Linux MM,
> see: http://www.linux-mm.org/ .
> Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>


--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2013-02-05  0:59 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2013-02-04 21:57 [patch] mm: use NUMA_NO_NODE David Rientjes
2013-02-05  0:21 ` Yasuaki Ishimatsu
2013-02-05  0:24 ` Yasuaki Ishimatsu
2013-02-05  0:59   ` Yasuaki Ishimatsu

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox