| /* |
| * Written by Kanoj Sarcar, SGI, Aug 1999 |
| */ |
| #include <linux/config.h> |
| #include <linux/kernel.h> |
| #include <linux/mm.h> |
| #include <linux/init.h> |
| #include <linux/bootmem.h> |
| #include <linux/mmzone.h> |
| #include <linux/spinlock.h> |
| |
| int numnodes = 1; /* Initialized for UMA platforms */ |
| |
| static bootmem_data_t contig_bootmem_data; |
| pg_data_t contig_page_data = { bdata: &contig_bootmem_data }; |
| |
| #ifndef CONFIG_DISCONTIGMEM |
| |
| /* |
| * This is meant to be invoked by platforms whose physical memory starts |
| * at a considerably higher value than 0. Examples are Super-H, ARM, m68k. |
| * Should be invoked with paramters (0, 0, unsigned long *[], start_paddr). |
| */ |
| void __init free_area_init_node(int nid, pg_data_t *pgdat, struct page *pmap, |
| unsigned long *zones_size, unsigned long zone_start_paddr, |
| unsigned long *zholes_size) |
| { |
| free_area_init_core(0, &contig_page_data, &mem_map, zones_size, |
| zone_start_paddr, zholes_size, pmap); |
| } |
| |
| #endif /* !CONFIG_DISCONTIGMEM */ |
| |
| struct page * alloc_pages_node(int nid, unsigned int gfp_mask, unsigned int order) |
| { |
| #ifdef CONFIG_NUMA |
| return __alloc_pages(gfp_mask, order, NODE_DATA(nid)->node_zonelists + (gfp_mask & GFP_ZONEMASK)); |
| #else |
| return alloc_pages(gfp_mask, order); |
| #endif |
| } |
| |
| #ifdef CONFIG_DISCONTIGMEM |
| |
| #define LONG_ALIGN(x) (((x)+(sizeof(long))-1)&~((sizeof(long))-1)) |
| |
| static spinlock_t node_lock = SPIN_LOCK_UNLOCKED; |
| |
| void show_free_areas_node(pg_data_t *pgdat) |
| { |
| unsigned long flags; |
| |
| spin_lock_irqsave(&node_lock, flags); |
| show_free_areas_core(pgdat); |
| spin_unlock_irqrestore(&node_lock, flags); |
| } |
| |
| /* |
| * Nodes can be initialized parallely, in no particular order. |
| */ |
| void __init free_area_init_node(int nid, pg_data_t *pgdat, struct page *pmap, |
| unsigned long *zones_size, unsigned long zone_start_paddr, |
| unsigned long *zholes_size) |
| { |
| int i, size = 0; |
| struct page *discard; |
| |
| if (mem_map == (mem_map_t *)NULL) |
| mem_map = (mem_map_t *)PAGE_OFFSET; |
| |
| free_area_init_core(nid, pgdat, &discard, zones_size, zone_start_paddr, |
| zholes_size, pmap); |
| pgdat->node_id = nid; |
| |
| /* |
| * Get space for the valid bitmap. |
| */ |
| for (i = 0; i < MAX_NR_ZONES; i++) |
| size += zones_size[i]; |
| size = LONG_ALIGN((size + 7) >> 3); |
| pgdat->valid_addr_bitmap = (unsigned long *)alloc_bootmem_node(pgdat, size); |
| memset(pgdat->valid_addr_bitmap, 0, size); |
| } |
| |
| static struct page * alloc_pages_pgdat(pg_data_t *pgdat, unsigned int gfp_mask, |
| unsigned int order) |
| { |
| return __alloc_pages(gfp_mask, order, pgdat->node_zonelists + (gfp_mask & GFP_ZONEMASK)); |
| } |
| |
| /* |
| * This can be refined. Currently, tries to do round robin, instead |
| * should do concentratic circle search, starting from current node. |
| */ |
| struct page * _alloc_pages(unsigned int gfp_mask, unsigned int order) |
| { |
| struct page *ret = 0; |
| pg_data_t *start, *temp; |
| #ifndef CONFIG_NUMA |
| unsigned long flags; |
| static pg_data_t *next = 0; |
| #endif |
| |
| if (order >= MAX_ORDER) |
| return NULL; |
| #ifdef CONFIG_NUMA |
| temp = NODE_DATA(numa_node_id()); |
| #else |
| spin_lock_irqsave(&node_lock, flags); |
| if (!next) next = pgdat_list; |
| temp = next; |
| next = next->node_next; |
| spin_unlock_irqrestore(&node_lock, flags); |
| #endif |
| start = temp; |
| while (temp) { |
| if ((ret = alloc_pages_pgdat(temp, gfp_mask, order))) |
| return(ret); |
| temp = temp->node_next; |
| } |
| temp = pgdat_list; |
| while (temp != start) { |
| if ((ret = alloc_pages_pgdat(temp, gfp_mask, order))) |
| return(ret); |
| temp = temp->node_next; |
| } |
| return(0); |
| } |
| |
| #endif /* CONFIG_DISCONTIGMEM */ |