[PATCH] Have Power use add_active_range() and free_area_init_nodes()

Size zones and holes in an architecture independent manner for Power.

[judith@osdl.org: build fix]
Signed-off-by: Mel Gorman <mel@csn.ul.ie>
Cc: Dave Hansen <haveblue@us.ibm.com>
Cc: Andy Whitcroft <apw@shadowen.org>
Cc: Andi Kleen <ak@muc.de>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Paul Mackerras <paulus@samba.org>
Cc: "Keith Mannthey" <kmannth@gmail.com>
Cc: "Luck, Tony" <tony.luck@intel.com>
Cc: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
Cc: Yasunori Goto <y-goto@jp.fujitsu.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
Mel Gorman 2006-09-27 01:49:49 -07:00 committed by Linus Torvalds
parent c713216dee
commit c67c3cb4c9
5 changed files with 52 additions and 191 deletions

View file

@ -731,11 +731,10 @@ config ARCH_SPARSEMEM_DEFAULT
def_bool y
depends on SMP && PPC_PSERIES
source "mm/Kconfig"
config HAVE_ARCH_EARLY_PFN_TO_NID
config ARCH_POPULATES_NODE_MAP
def_bool y
depends on NEED_MULTIPLE_NODES
source "mm/Kconfig"
config ARCH_MEMORY_PROBE
def_bool y

View file

@ -256,20 +256,22 @@ void __init do_init_bootmem(void)
boot_mapsize = init_bootmem(start >> PAGE_SHIFT, total_pages);
/* Add active regions with valid PFNs */
for (i = 0; i < lmb.memory.cnt; i++) {
unsigned long start_pfn, end_pfn;
start_pfn = lmb.memory.region[i].base >> PAGE_SHIFT;
end_pfn = start_pfn + lmb_size_pages(&lmb.memory, i);
add_active_range(0, start_pfn, end_pfn);
}
/* Add all physical memory to the bootmem map, mark each area
* present.
*/
for (i = 0; i < lmb.memory.cnt; i++) {
unsigned long base = lmb.memory.region[i].base;
unsigned long size = lmb_size_bytes(&lmb.memory, i);
#ifdef CONFIG_HIGHMEM
if (base >= total_lowmem)
continue;
if (base + size > total_lowmem)
size = total_lowmem - base;
free_bootmem_with_active_regions(0, total_lowmem >> PAGE_SHIFT);
#else
free_bootmem_with_active_regions(0, max_pfn);
#endif
free_bootmem(base, size);
}
/* reserve the sections we're already using */
for (i = 0; i < lmb.reserved.cnt; i++)
@ -277,9 +279,8 @@ void __init do_init_bootmem(void)
lmb_size_bytes(&lmb.reserved, i));
/* XXX need to clip this if using highmem? */
for (i = 0; i < lmb.memory.cnt; i++)
memory_present(0, lmb_start_pfn(&lmb.memory, i),
lmb_end_pfn(&lmb.memory, i));
sparse_memory_present_with_active_regions(0);
init_bootmem_done = 1;
}
@ -288,10 +289,9 @@ void __init do_init_bootmem(void)
*/
void __init paging_init(void)
{
unsigned long zones_size[MAX_NR_ZONES];
unsigned long zholes_size[MAX_NR_ZONES];
unsigned long total_ram = lmb_phys_mem_size();
unsigned long top_of_ram = lmb_end_of_DRAM();
unsigned long max_zone_pfns[MAX_NR_ZONES];
#ifdef CONFIG_HIGHMEM
map_page(PKMAP_BASE, 0, 0); /* XXX gross */
@ -307,26 +307,13 @@ void __init paging_init(void)
top_of_ram, total_ram);
printk(KERN_DEBUG "Memory hole size: %ldMB\n",
(top_of_ram - total_ram) >> 20);
/*
* All pages are DMA-able so we put them all in the DMA zone.
*/
memset(zones_size, 0, sizeof(zones_size));
memset(zholes_size, 0, sizeof(zholes_size));
zones_size[ZONE_DMA] = top_of_ram >> PAGE_SHIFT;
zholes_size[ZONE_DMA] = (top_of_ram - total_ram) >> PAGE_SHIFT;
#ifdef CONFIG_HIGHMEM
zones_size[ZONE_DMA] = total_lowmem >> PAGE_SHIFT;
zones_size[ZONE_HIGHMEM] = (total_memory - total_lowmem) >> PAGE_SHIFT;
zholes_size[ZONE_HIGHMEM] = (top_of_ram - total_ram) >> PAGE_SHIFT;
max_zone_pfns[0] = total_lowmem >> PAGE_SHIFT;
max_zone_pfns[1] = top_of_ram >> PAGE_SHIFT;
#else
zones_size[ZONE_DMA] = top_of_ram >> PAGE_SHIFT;
zholes_size[ZONE_DMA] = (top_of_ram - total_ram) >> PAGE_SHIFT;
#endif /* CONFIG_HIGHMEM */
free_area_init_node(0, NODE_DATA(0), zones_size,
__pa(PAGE_OFFSET) >> PAGE_SHIFT, zholes_size);
max_zone_pfns[0] = top_of_ram >> PAGE_SHIFT;
#endif
free_area_init_nodes(max_zone_pfns);
}
#endif /* ! CONFIG_NEED_MULTIPLE_NODES */

View file

@ -39,96 +39,6 @@ static bootmem_data_t __initdata plat_node_bdata[MAX_NUMNODES];
static int min_common_depth;
static int n_mem_addr_cells, n_mem_size_cells;
/*
* We need somewhere to store start/end/node for each region until we have
* allocated the real node_data structures.
*/
#define MAX_REGIONS (MAX_LMB_REGIONS*2)
static struct {
unsigned long start_pfn;
unsigned long end_pfn;
int nid;
} init_node_data[MAX_REGIONS] __initdata;
int __init early_pfn_to_nid(unsigned long pfn)
{
unsigned int i;
for (i = 0; init_node_data[i].end_pfn; i++) {
unsigned long start_pfn = init_node_data[i].start_pfn;
unsigned long end_pfn = init_node_data[i].end_pfn;
if ((start_pfn <= pfn) && (pfn < end_pfn))
return init_node_data[i].nid;
}
return -1;
}
void __init add_region(unsigned int nid, unsigned long start_pfn,
unsigned long pages)
{
unsigned int i;
dbg("add_region nid %d start_pfn 0x%lx pages 0x%lx\n",
nid, start_pfn, pages);
for (i = 0; init_node_data[i].end_pfn; i++) {
if (init_node_data[i].nid != nid)
continue;
if (init_node_data[i].end_pfn == start_pfn) {
init_node_data[i].end_pfn += pages;
return;
}
if (init_node_data[i].start_pfn == (start_pfn + pages)) {
init_node_data[i].start_pfn -= pages;
return;
}
}
/*
* Leave last entry NULL so we dont iterate off the end (we use
* entry.end_pfn to terminate the walk).
*/
if (i >= (MAX_REGIONS - 1)) {
printk(KERN_ERR "WARNING: too many memory regions in "
"numa code, truncating\n");
return;
}
init_node_data[i].start_pfn = start_pfn;
init_node_data[i].end_pfn = start_pfn + pages;
init_node_data[i].nid = nid;
}
/* We assume init_node_data has no overlapping regions */
void __init get_region(unsigned int nid, unsigned long *start_pfn,
unsigned long *end_pfn, unsigned long *pages_present)
{
unsigned int i;
*start_pfn = -1UL;
*end_pfn = *pages_present = 0;
for (i = 0; init_node_data[i].end_pfn; i++) {
if (init_node_data[i].nid != nid)
continue;
*pages_present += init_node_data[i].end_pfn -
init_node_data[i].start_pfn;
if (init_node_data[i].start_pfn < *start_pfn)
*start_pfn = init_node_data[i].start_pfn;
if (init_node_data[i].end_pfn > *end_pfn)
*end_pfn = init_node_data[i].end_pfn;
}
/* We didnt find a matching region, return start/end as 0 */
if (*start_pfn == -1UL)
*start_pfn = 0;
}
static void __cpuinit map_cpu_to_node(int cpu, int node)
{
numa_cpu_lookup_table[cpu] = node;
@ -468,8 +378,8 @@ static int __init parse_numa_properties(void)
continue;
}
add_region(nid, start >> PAGE_SHIFT,
size >> PAGE_SHIFT);
add_active_range(nid, start >> PAGE_SHIFT,
(start >> PAGE_SHIFT) + (size >> PAGE_SHIFT));
if (--ranges)
goto new_range;
@ -482,6 +392,7 @@ static void __init setup_nonnuma(void)
{
unsigned long top_of_ram = lmb_end_of_DRAM();
unsigned long total_ram = lmb_phys_mem_size();
unsigned long start_pfn, end_pfn;
unsigned int i;
printk(KERN_DEBUG "Top of RAM: 0x%lx, Total RAM: 0x%lx\n",
@ -489,9 +400,11 @@ static void __init setup_nonnuma(void)
printk(KERN_DEBUG "Memory hole size: %ldMB\n",
(top_of_ram - total_ram) >> 20);
for (i = 0; i < lmb.memory.cnt; ++i)
add_region(0, lmb.memory.region[i].base >> PAGE_SHIFT,
lmb_size_pages(&lmb.memory, i));
for (i = 0; i < lmb.memory.cnt; ++i) {
start_pfn = lmb.memory.region[i].base >> PAGE_SHIFT;
end_pfn = start_pfn + lmb_size_pages(&lmb.memory, i);
add_active_range(0, start_pfn, end_pfn);
}
node_set_online(0);
}
@ -630,11 +543,11 @@ void __init do_init_bootmem(void)
(void *)(unsigned long)boot_cpuid);
for_each_online_node(nid) {
unsigned long start_pfn, end_pfn, pages_present;
unsigned long start_pfn, end_pfn;
unsigned long bootmem_paddr;
unsigned long bootmap_pages;
get_region(nid, &start_pfn, &end_pfn, &pages_present);
get_pfn_range_for_nid(nid, &start_pfn, &end_pfn);
/* Allocate the node structure node local if possible */
NODE_DATA(nid) = careful_allocation(nid,
@ -667,19 +580,7 @@ void __init do_init_bootmem(void)
init_bootmem_node(NODE_DATA(nid), bootmem_paddr >> PAGE_SHIFT,
start_pfn, end_pfn);
/* Add free regions on this node */
for (i = 0; init_node_data[i].end_pfn; i++) {
unsigned long start, end;
if (init_node_data[i].nid != nid)
continue;
start = init_node_data[i].start_pfn << PAGE_SHIFT;
end = init_node_data[i].end_pfn << PAGE_SHIFT;
dbg("free_bootmem %lx %lx\n", start, end - start);
free_bootmem_node(NODE_DATA(nid), start, end - start);
}
free_bootmem_with_active_regions(nid, end_pfn);
/* Mark reserved regions on this node */
for (i = 0; i < lmb.reserved.cnt; i++) {
@ -710,44 +611,16 @@ void __init do_init_bootmem(void)
}
}
/* Add regions into sparsemem */
for (i = 0; init_node_data[i].end_pfn; i++) {
unsigned long start, end;
if (init_node_data[i].nid != nid)
continue;
start = init_node_data[i].start_pfn;
end = init_node_data[i].end_pfn;
memory_present(nid, start, end);
}
sparse_memory_present_with_active_regions(nid);
}
}
void __init paging_init(void)
{
unsigned long zones_size[MAX_NR_ZONES];
unsigned long zholes_size[MAX_NR_ZONES];
int nid;
memset(zones_size, 0, sizeof(zones_size));
memset(zholes_size, 0, sizeof(zholes_size));
for_each_online_node(nid) {
unsigned long start_pfn, end_pfn, pages_present;
get_region(nid, &start_pfn, &end_pfn, &pages_present);
zones_size[ZONE_DMA] = end_pfn - start_pfn;
zholes_size[ZONE_DMA] = zones_size[ZONE_DMA] - pages_present;
dbg("free_area_init node %d %lx %lx (hole: %lx)\n", nid,
zones_size[ZONE_DMA], start_pfn, zholes_size[ZONE_DMA]);
free_area_init_node(nid, NODE_DATA(nid), zones_size, start_pfn,
zholes_size);
}
unsigned long max_zone_pfns[MAX_NR_ZONES] = {
lmb_end_of_DRAM() >> PAGE_SHIFT
};
free_area_init_nodes(max_zone_pfns);
}
static int __init early_numa(char *p)

View file

@ -953,6 +953,9 @@ config NR_CPUS
config HIGHMEM
bool "High memory support"
config ARCH_POPULATES_NODE_MAP
def_bool y
source kernel/Kconfig.hz
source kernel/Kconfig.preempt
source "mm/Kconfig"

View file

@ -358,8 +358,8 @@ void __init do_init_bootmem(void)
*/
void __init paging_init(void)
{
unsigned long zones_size[MAX_NR_ZONES], i;
unsigned long start_pfn, end_pfn;
unsigned long max_zone_pfns[MAX_NR_ZONES];
#ifdef CONFIG_HIGHMEM
map_page(PKMAP_BASE, 0, 0); /* XXX gross */
pkmap_page_table = pte_offset_kernel(pmd_offset(pgd_offset_k
@ -369,19 +369,18 @@ void __init paging_init(void)
(KMAP_FIX_BEGIN), KMAP_FIX_BEGIN), KMAP_FIX_BEGIN);
kmap_prot = PAGE_KERNEL;
#endif /* CONFIG_HIGHMEM */
/*
* All pages are DMA-able so we put them all in the DMA zone.
*/
zones_size[ZONE_DMA] = total_lowmem >> PAGE_SHIFT;
for (i = 1; i < MAX_NR_ZONES; i++)
zones_size[i] = 0;
/* All pages are DMA-able so we put them all in the DMA zone. */
start_pfn = __pa(PAGE_OFFSET) >> PAGE_SHIFT;
end_pfn = start_pfn + (total_memory >> PAGE_SHIFT);
add_active_range(0, start_pfn, end_pfn);
#ifdef CONFIG_HIGHMEM
zones_size[ZONE_HIGHMEM] = (total_memory - total_lowmem) >> PAGE_SHIFT;
max_zone_pfns[0] = total_lowmem >> PAGE_SHIFT;
max_zone_pfns[1] = total_memory >> PAGE_SHIFT;
#else
max_zone_pfns[0] = total_memory >> PAGE_SHIFT;
#endif /* CONFIG_HIGHMEM */
free_area_init(zones_size);
free_area_init_nodes(max_zone_pfns);
}
void __init mem_init(void)