mirror of
				git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
				synced 2025-10-31 16:54:21 +00:00 
			
		
		
		
	[PATCH] NUMA: Add zone_to_nid function
There are many places where we need to determine the node of a zone. Currently we use a difficult to read sequence of pointer dereferencing. Put that into an inline function and use throughout VM. Maybe we can find a way to optimize the lookup in the future. Signed-off-by: Christoph Lameter <clameter@sgi.com> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
		
							parent
							
								
									4415cc8df6
								
							
						
					
					
						commit
						89fa30242f
					
				
					 9 changed files with 17 additions and 13 deletions
				
			
		|  | @ -422,7 +422,7 @@ void __init set_highmem_pages_init(int bad_ppro) | |||
| 		zone_end_pfn = zone_start_pfn + zone->spanned_pages; | ||||
| 
 | ||||
| 		printk("Initializing %s for node %d (%08lx:%08lx)\n", | ||||
| 				zone->name, zone->zone_pgdat->node_id, | ||||
| 				zone->name, zone_to_nid(zone), | ||||
| 				zone_start_pfn, zone_end_pfn); | ||||
| 
 | ||||
| 		for (node_pfn = zone_start_pfn; node_pfn < zone_end_pfn; node_pfn++) { | ||||
|  |  | |||
|  | @ -551,7 +551,7 @@ void show_mem(void) | |||
| 
 | ||||
| 				printk("Zone list for zone %d on node %d: ", j, i); | ||||
| 				for (k = 0; zl->zones[k] != NULL; k++)  | ||||
| 					printk("[%d/%s] ", zl->zones[k]->zone_pgdat->node_id, zl->zones[k]->name); | ||||
| 					printk("[%d/%s] ", zone_to_nid(zl->zones[k]), zl->zones[k]->name); | ||||
| 				printk("\n"); | ||||
| 			} | ||||
| 		} | ||||
|  |  | |||
|  | @ -499,12 +499,17 @@ static inline struct zone *page_zone(struct page *page) | |||
| 	return zone_table[page_zone_id(page)]; | ||||
| } | ||||
| 
 | ||||
| static inline unsigned long zone_to_nid(struct zone *zone) | ||||
| { | ||||
| 	return zone->zone_pgdat->node_id; | ||||
| } | ||||
| 
 | ||||
| static inline unsigned long page_to_nid(struct page *page) | ||||
| { | ||||
| 	if (FLAGS_HAS_NODE) | ||||
| 		return (page->flags >> NODES_PGSHIFT) & NODES_MASK; | ||||
| 	else | ||||
| 		return page_zone(page)->zone_pgdat->node_id; | ||||
| 		return zone_to_nid(page_zone(page)); | ||||
| } | ||||
| static inline unsigned long page_to_section(struct page *page) | ||||
| { | ||||
|  |  | |||
|  | @ -2245,7 +2245,7 @@ int cpuset_zonelist_valid_mems_allowed(struct zonelist *zl) | |||
| 	int i; | ||||
| 
 | ||||
| 	for (i = 0; zl->zones[i]; i++) { | ||||
| 		int nid = zl->zones[i]->zone_pgdat->node_id; | ||||
| 		int nid = zone_to_nid(zl->zones[i]); | ||||
| 
 | ||||
| 		if (node_isset(nid, current->mems_allowed)) | ||||
| 			return 1; | ||||
|  | @ -2318,7 +2318,7 @@ int __cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) | |||
| 
 | ||||
| 	if (in_interrupt() || (gfp_mask & __GFP_THISNODE)) | ||||
| 		return 1; | ||||
| 	node = z->zone_pgdat->node_id; | ||||
| 	node = zone_to_nid(z); | ||||
| 	might_sleep_if(!(gfp_mask & __GFP_HARDWALL)); | ||||
| 	if (node_isset(node, current->mems_allowed)) | ||||
| 		return 1; | ||||
|  |  | |||
|  | @ -72,7 +72,7 @@ static struct page *dequeue_huge_page(struct vm_area_struct *vma, | |||
| 	struct zone **z; | ||||
| 
 | ||||
| 	for (z = zonelist->zones; *z; z++) { | ||||
| 		nid = (*z)->zone_pgdat->node_id; | ||||
| 		nid = zone_to_nid(*z); | ||||
| 		if (cpuset_zone_allowed(*z, GFP_HIGHUSER) && | ||||
| 		    !list_empty(&hugepage_freelists[nid])) | ||||
| 			break; | ||||
|  |  | |||
|  | @ -487,7 +487,7 @@ static void get_zonemask(struct mempolicy *p, nodemask_t *nodes) | |||
| 	switch (p->policy) { | ||||
| 	case MPOL_BIND: | ||||
| 		for (i = 0; p->v.zonelist->zones[i]; i++) | ||||
| 			node_set(p->v.zonelist->zones[i]->zone_pgdat->node_id, | ||||
| 			node_set(zone_to_nid(p->v.zonelist->zones[i]), | ||||
| 				*nodes); | ||||
| 		break; | ||||
| 	case MPOL_DEFAULT: | ||||
|  | @ -1145,7 +1145,7 @@ unsigned slab_node(struct mempolicy *policy) | |||
| 		 * Follow bind policy behavior and start allocation at the | ||||
| 		 * first node. | ||||
| 		 */ | ||||
| 		return policy->v.zonelist->zones[0]->zone_pgdat->node_id; | ||||
| 		return zone_to_nid(policy->v.zonelist->zones[0]); | ||||
| 
 | ||||
| 	case MPOL_PREFERRED: | ||||
| 		if (policy->v.preferred_node >= 0) | ||||
|  | @ -1649,7 +1649,7 @@ void mpol_rebind_policy(struct mempolicy *pol, const nodemask_t *newmask) | |||
| 
 | ||||
| 		nodes_clear(nodes); | ||||
| 		for (z = pol->v.zonelist->zones; *z; z++) | ||||
| 			node_set((*z)->zone_pgdat->node_id, nodes); | ||||
| 			node_set(zone_to_nid(*z), nodes); | ||||
| 		nodes_remap(tmp, nodes, *mpolmask, *newmask); | ||||
| 		nodes = tmp; | ||||
| 
 | ||||
|  |  | |||
|  | @ -177,8 +177,7 @@ static inline int constrained_alloc(struct zonelist *zonelist, gfp_t gfp_mask) | |||
| 
 | ||||
| 	for (z = zonelist->zones; *z; z++) | ||||
| 		if (cpuset_zone_allowed(*z, gfp_mask)) | ||||
| 			node_clear((*z)->zone_pgdat->node_id, | ||||
| 					nodes); | ||||
| 			node_clear(zone_to_nid(*z), nodes); | ||||
| 		else | ||||
| 			return CONSTRAINT_CPUSET; | ||||
| 
 | ||||
|  |  | |||
|  | @ -1217,7 +1217,7 @@ unsigned int nr_free_pagecache_pages(void) | |||
| #ifdef CONFIG_NUMA | ||||
| static void show_node(struct zone *zone) | ||||
| { | ||||
| 	printk("Node %d ", zone->zone_pgdat->node_id); | ||||
| 	printk("Node %ld ", zone_to_nid(zone)); | ||||
| } | ||||
| #else | ||||
| #define show_node(zone)	do { } while (0) | ||||
|  |  | |||
|  | @ -1661,7 +1661,7 @@ int zone_reclaim(struct zone *zone, gfp_t gfp_mask, unsigned int order) | |||
| 	 * over remote processors and spread off node memory allocations | ||||
| 	 * as wide as possible. | ||||
| 	 */ | ||||
| 	node_id = zone->zone_pgdat->node_id; | ||||
| 	node_id = zone_to_nid(zone); | ||||
| 	mask = node_to_cpumask(node_id); | ||||
| 	if (!cpus_empty(mask) && node_id != numa_node_id()) | ||||
| 		return 0; | ||||
|  |  | |||
		Loading…
	
	Add table
		
		Reference in a new issue
	
	 Christoph Lameter
						Christoph Lameter