Fix sparsemem_vmemmap init. sorry if known bug. This patch fixes page table handling in sparsemem_vmammap.
Without this, part of vmem_map is not mapped because each section's start addr of mem_map is not aligned to PGD/PMD/PUD. (In ia64, secion's mem_map size is 3670016bytes. ) for example, addr pmd_addr_end(addr_end) addr + PMD_SIZE |XXXXXXXXXX|??????????????????????????????|XXXXXXXXXXXXXXXXXX X ... initialized vmem_map ? ... not intialized Signed-off-by: KAMEZAWA Hiroyuki <[EMAIL PROTECTED]> --- mm/sparse.c | 24 +++++++++++++----------- 1 file changed, 13 insertions(+), 11 deletions(-) Index: devel-2.6.23-rc1-mm1/mm/sparse.c =================================================================== --- devel-2.6.23-rc1-mm1.orig/mm/sparse.c +++ devel-2.6.23-rc1-mm1/mm/sparse.c @@ -320,7 +320,7 @@ static int __meminit vmemmap_populate_pt { pte_t *pte; - for (pte = pte_offset_map(pmd, addr); addr < end; + for (pte = pte_offset_kernel(pmd, addr); addr < end; pte++, addr += PAGE_SIZE) if (pte_none(*pte)) { pte_t entry; @@ -345,9 +345,10 @@ int __meminit vmemmap_populate_pmd(pud_t { pmd_t *pmd; int error = 0; + unsigned long next; for (pmd = pmd_offset(pud, addr); addr < end && !error; - pmd++, addr += PMD_SIZE) { + pmd++, addr = next) { if (pmd_none(*pmd)) { void *p = vmemmap_alloc_block(PAGE_SIZE, node); if (!p) @@ -357,9 +358,8 @@ int __meminit vmemmap_populate_pmd(pud_t } else vmemmap_verify((pte_t *)pmd, node, pmd_addr_end(addr, end), end); - - error = vmemmap_populate_pte(pmd, addr, - pmd_addr_end(addr, end), node); + next = pmd_addr_end(addr, end); + error = vmemmap_populate_pte(pmd, addr, next, node); } return error; } @@ -370,9 +370,10 @@ static int __meminit vmemmap_populate_pu { pud_t *pud; int error = 0; + unsigned long next; for (pud = pud_offset(pgd, addr); addr < end && !error; - pud++, addr += PUD_SIZE) { + pud++, addr = next) { if (pud_none(*pud)) { void *p = vmemmap_alloc_block(PAGE_SIZE, node); if (!p) @@ -380,8 +381,8 @@ static int __meminit vmemmap_populate_pu pud_populate(&init_mm, pud, p); } - error = vmemmap_populate_pmd(pud, addr, - pud_addr_end(addr, end), node); + next = pud_addr_end(addr, end); + error = vmemmap_populate_pmd(pud, addr, next, node); } return error; } @@ -392,13 +393,14 @@ int __meminit vmemmap_populate(struct pa pgd_t *pgd; unsigned long addr = (unsigned long)start_page; unsigned long end = (unsigned long)(start_page + nr); + unsigned long next; int error = 0; printk(KERN_DEBUG "[%lx-%lx] Virtual memory section" " (%ld pages) node %d\n", addr, end - 1, nr, node); for (pgd = pgd_offset_k(addr); addr < end && !error; - pgd++, addr += PGDIR_SIZE) { + pgd++, addr = next) { if (pgd_none(*pgd)) { void *p = vmemmap_alloc_block(PAGE_SIZE, node); if (!p) @@ -406,8 +408,8 @@ int __meminit vmemmap_populate(struct pa pgd_populate(&init_mm, pgd, p); } - error = vmemmap_populate_pud(pgd, addr, - pgd_addr_end(addr, end), node); + next = pgd_addr_end(addr,end); + error = vmemmap_populate_pud(pgd, addr, next, node); } return error; } - To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/