On Tue, 28 Jul 2020, 07:16 Mike Rapoport, <r...@kernel.org> wrote: > From: Mike Rapoport <r...@linux.ibm.com> > > There are several occurrences of the following pattern: > > for_each_memblock(memory, reg) { > start = __pfn_to_phys(memblock_region_memory_base_pfn(reg); > end = __pfn_to_phys(memblock_region_memory_end_pfn(reg)); > > /* do something with start and end */ > } > > Using for_each_mem_range() iterator is more appropriate in such cases and > allows simpler and cleaner code. > > Signed-off-by: Mike Rapoport <r...@linux.ibm.com> > --- > arch/arm/kernel/setup.c | 18 +++++++---- > arch/arm/mm/mmu.c | 39 ++++++++---------------- > arch/arm/mm/pmsa-v7.c | 20 ++++++------ > arch/arm/mm/pmsa-v8.c | 17 +++++------ > arch/arm/xen/mm.c | 7 +++-- > arch/arm64/mm/kasan_init.c | 8 ++--- > arch/arm64/mm/mmu.c | 11 ++----- > arch/c6x/kernel/setup.c | 9 +++--- > arch/microblaze/mm/init.c | 9 +++--- > arch/mips/cavium-octeon/dma-octeon.c | 12 ++++---- > arch/mips/kernel/setup.c | 31 +++++++++---------- > arch/openrisc/mm/init.c | 8 +++-- > arch/powerpc/kernel/fadump.c | 27 +++++++--------- > arch/powerpc/mm/book3s64/hash_utils.c | 16 +++++----- > arch/powerpc/mm/book3s64/radix_pgtable.c | 11 +++---- > arch/powerpc/mm/kasan/kasan_init_32.c | 8 ++--- > arch/powerpc/mm/mem.c | 16 ++++++---- > arch/powerpc/mm/pgtable_32.c | 8 ++--- > arch/riscv/mm/init.c | 24 ++++++--------- > arch/riscv/mm/kasan_init.c | 10 +++--- > arch/s390/kernel/setup.c | 27 ++++++++++------ > arch/s390/mm/vmem.c | 16 +++++----- > arch/sparc/mm/init_64.c | 12 +++----- > drivers/bus/mvebu-mbus.c | 12 ++++---- > drivers/s390/char/zcore.c | 9 +++--- > 25 files changed, 187 insertions(+), 198 deletions(-) > > diff --git a/arch/arm/kernel/setup.c b/arch/arm/kernel/setup.c > index d8e18cdd96d3..3f65d0ac9f63 100644 > --- a/arch/arm/kernel/setup.c > +++ b/arch/arm/kernel/setup.c > @@ -843,19 +843,25 @@ early_param("mem", early_mem); > > static void __init request_standard_resources(const struct machine_desc > *mdesc) > { > - struct memblock_region *region; > + phys_addr_t start, end, res_end; > struct resource *res; > + u64 i; > > kernel_code.start = virt_to_phys(_text); > kernel_code.end = virt_to_phys(__init_begin - 1); > kernel_data.start = virt_to_phys(_sdata); > kernel_data.end = virt_to_phys(_end - 1); > > - for_each_memblock(memory, region) { > - phys_addr_t start = > __pfn_to_phys(memblock_region_memory_base_pfn(region)); > - phys_addr_t end = > __pfn_to_phys(memblock_region_memory_end_pfn(region)) - 1; > + for_each_mem_range(i, &start, &end) { > unsigned long boot_alias_start; > > + /* > + * In memblock, end points to the first byte after the > + * range while in resourses, end points to the last byte in > + * the range. > + */ > + res_end = end - 1; > + > /* > * Some systems have a special memory alias which is only > * used for booting. We need to advertise this region to > @@ -869,7 +875,7 @@ static void __init request_standard_resources(const > struct machine_desc *mdesc) > __func__, sizeof(*res)); > res->name = "System RAM (boot alias)"; > res->start = boot_alias_start; > - res->end = phys_to_idmap(end); > + res->end = phys_to_idmap(res_end); > res->flags = IORESOURCE_MEM | IORESOURCE_BUSY; > request_resource(&iomem_resource, res); > } > @@ -880,7 +886,7 @@ static void __init request_standard_resources(const > struct machine_desc *mdesc) > sizeof(*res)); > res->name = "System RAM"; > res->start = start; > - res->end = end; > + res->end = res_end; > res->flags = IORESOURCE_SYSTEM_RAM | IORESOURCE_BUSY; > > request_resource(&iomem_resource, res); > diff --git a/arch/arm/mm/mmu.c b/arch/arm/mm/mmu.c > index 628028bfbb92..a149d9cb4fdb 100644 > --- a/arch/arm/mm/mmu.c > +++ b/arch/arm/mm/mmu.c > @@ -1155,9 +1155,8 @@ phys_addr_t arm_lowmem_limit __initdata = 0; > > void __init adjust_lowmem_bounds(void) > { > - phys_addr_t memblock_limit = 0; > - u64 vmalloc_limit; > - struct memblock_region *reg; > + phys_addr_t block_start, block_end, memblock_limit = 0; > + u64 vmalloc_limit, i; > phys_addr_t lowmem_limit = 0; > > /* > @@ -1173,26 +1172,18 @@ void __init adjust_lowmem_bounds(void) > * The first usable region must be PMD aligned. Mark its start > * as MEMBLOCK_NOMAP if it isn't > */ > - for_each_memblock(memory, reg) { > - if (!memblock_is_nomap(reg)) { > - if (!IS_ALIGNED(reg->base, PMD_SIZE)) { > - phys_addr_t len; > + for_each_mem_range(i, &block_start, &block_end) { > + if (!IS_ALIGNED(block_start, PMD_SIZE)) { > + phys_addr_t len; > > - len = round_up(reg->base, PMD_SIZE) - > reg->base; > - memblock_mark_nomap(reg->base, len); > - } > - break; > + len = round_up(block_start, PMD_SIZE) - > block_start; > + memblock_mark_nomap(block_start, len); > } > + break; > } > > - for_each_memblock(memory, reg) { > - phys_addr_t block_start = reg->base; > - phys_addr_t block_end = reg->base + reg->size; > - > - if (memblock_is_nomap(reg)) > - continue; > - > - if (reg->base < vmalloc_limit) { > + for_each_mem_range(i, &block_start, &block_end) { > + if (block_start < vmalloc_limit) { > if (block_end > lowmem_limit) > /* > * Compare as u64 to ensure vmalloc_limit > does > @@ -1441,19 +1432,15 @@ static void __init kmap_init(void) > > static void __init map_lowmem(void) > { > - struct memblock_region *reg; > phys_addr_t kernel_x_start = round_down(__pa(KERNEL_START), > SECTION_SIZE); > phys_addr_t kernel_x_end = round_up(__pa(__init_end), > SECTION_SIZE); > + phys_addr_t start, end; > + u64 i; > > /* Map all the lowmem memory banks. */ > - for_each_memblock(memory, reg) { > - phys_addr_t start = reg->base; > - phys_addr_t end = start + reg->size; > + for_each_mem_range(i, &start, &end) { > struct map_desc map; > > - if (memblock_is_nomap(reg)) > - continue; > - > if (end > arm_lowmem_limit) > end = arm_lowmem_limit; > if (start >= end) > diff --git a/arch/arm/mm/pmsa-v7.c b/arch/arm/mm/pmsa-v7.c > index 699fa2e88725..44b7644a4237 100644 > --- a/arch/arm/mm/pmsa-v7.c > +++ b/arch/arm/mm/pmsa-v7.c > @@ -231,10 +231,9 @@ static int __init allocate_region(phys_addr_t base, > phys_addr_t size, > void __init pmsav7_adjust_lowmem_bounds(void) > { > phys_addr_t specified_mem_size = 0, total_mem_size = 0; > - struct memblock_region *reg; > - bool first = true; > phys_addr_t mem_start; > phys_addr_t mem_end; > + phys_addr_t reg_start, reg_end; > unsigned int mem_max_regions; > int num, i; > > @@ -262,20 +261,19 @@ void __init pmsav7_adjust_lowmem_bounds(void) > mem_max_regions -= num; > #endif > > - for_each_memblock(memory, reg) { > - if (first) { > + for_each_mem_range(i, ®_start, ®_end) { > + if (i == 0) { > phys_addr_t phys_offset = PHYS_OFFSET; > > /* > * Initially only use memory continuous from > * PHYS_OFFSET */ > - if (reg->base != phys_offset) > + if (reg_start != phys_offset) > panic("First memory bank must be > contiguous from PHYS_OFFSET"); > > - mem_start = reg->base; > - mem_end = reg->base + reg->size; > - specified_mem_size = reg->size; > - first = false; > + mem_start = reg_start; > + mem_end = reg_end > + specified_mem_size = mem_end - mem_start; > } else { > /* > * memblock auto merges contiguous blocks, remove > @@ -283,8 +281,8 @@ void __init pmsav7_adjust_lowmem_bounds(void) > * blocks separately while iterating) > */ > pr_notice("Ignoring RAM after %pa, memory at %pa > ignored\n", > - &mem_end, ®->base); > - memblock_remove(reg->base, 0 - reg->base); > + &mem_end, ®_start); > + memblock_remove(reg_start, 0 - reg_start); > break; > } > } > diff --git a/arch/arm/mm/pmsa-v8.c b/arch/arm/mm/pmsa-v8.c > index 0d7d5fb59247..b39e74b48437 100644 > --- a/arch/arm/mm/pmsa-v8.c > +++ b/arch/arm/mm/pmsa-v8.c > @@ -94,20 +94,19 @@ static __init bool is_region_fixed(int number) > void __init pmsav8_adjust_lowmem_bounds(void) > { > phys_addr_t mem_end; > - struct memblock_region *reg; > - bool first = true; > + phys_addr_t reg_start, reg_end; > + int i; > > - for_each_memblock(memory, reg) { > - if (first) { > + for_each_mem_range(i, ®_start, ®_end) { > + if (i == 0) { > phys_addr_t phys_offset = PHYS_OFFSET; > > /* > * Initially only use memory continuous from > * PHYS_OFFSET */ > - if (reg->base != phys_offset) > + if (reg_start != phys_offset) > panic("First memory bank must be > contiguous from PHYS_OFFSET"); > - mem_end = reg->base + reg->size; > - first = false; > + mem_end = reg_end; > } else { > /* > * memblock auto merges contiguous blocks, remove > @@ -115,8 +114,8 @@ void __init pmsav8_adjust_lowmem_bounds(void) > * blocks separately while iterating) > */ > pr_notice("Ignoring RAM after %pa, memory at %pa > ignored\n", > - &mem_end, ®->base); > - memblock_remove(reg->base, 0 - reg->base); > + &mem_end, ®_start); > + memblock_remove(reg_start, 0 - reg_start); > break; > } > } > diff --git a/arch/arm/xen/mm.c b/arch/arm/xen/mm.c > index d40e9e5fc52b..05f24ff41e36 100644 > --- a/arch/arm/xen/mm.c > +++ b/arch/arm/xen/mm.c > @@ -24,11 +24,12 @@ > > unsigned long xen_get_swiotlb_free_pages(unsigned int order) > { > - struct memblock_region *reg; > + phys_addr_t base; > gfp_t flags = __GFP_NOWARN|__GFP_KSWAPD_RECLAIM; > + u64 i; > > - for_each_memblock(memory, reg) { > - if (reg->base < (phys_addr_t)0xffffffff) { > + for_each_mem_range(i, &base, NULL) { > + if (base < (phys_addr_t)0xffffffff) { > if (IS_ENABLED(CONFIG_ZONE_DMA32)) > flags |= __GFP_DMA32; > else > diff --git a/arch/arm64/mm/kasan_init.c b/arch/arm64/mm/kasan_init.c > index 7291b26ce788..1faa086f9193 100644 > --- a/arch/arm64/mm/kasan_init.c > +++ b/arch/arm64/mm/kasan_init.c > @@ -212,7 +212,7 @@ void __init kasan_init(void) > { > u64 kimg_shadow_start, kimg_shadow_end; > u64 mod_shadow_start, mod_shadow_end; > - struct memblock_region *reg; > + phys_addr_t _start, _end; > int i; > > kimg_shadow_start = (u64)kasan_mem_to_shadow(_text) & PAGE_MASK; > @@ -246,9 +246,9 @@ void __init kasan_init(void) > kasan_populate_early_shadow((void *)mod_shadow_end, > (void *)kimg_shadow_start); > > - for_each_memblock(memory, reg) { > - void *start = (void *)__phys_to_virt(reg->base); > - void *end = (void *)__phys_to_virt(reg->base + reg->size); > + for_each_mem_range(i, &start, &end) { > + void *_start = (void *)__phys_to_virt(_start); > + void *end = (void *)__phys_to_virt(_end); > > if (start >= end) > break; > diff --git a/arch/arm64/mm/mmu.c b/arch/arm64/mm/mmu.c > index 1df25f26571d..327264fb83fb 100644 > --- a/arch/arm64/mm/mmu.c > +++ b/arch/arm64/mm/mmu.c > @@ -461,8 +461,9 @@ static void __init map_mem(pgd_t *pgdp) > { > phys_addr_t kernel_start = __pa_symbol(_text); > phys_addr_t kernel_end = __pa_symbol(__init_begin); > - struct memblock_region *reg; > + phys_addr_t start, end; > int flags = 0; > + u64 i; > > if (rodata_full || debug_pagealloc_enabled()) > flags = NO_BLOCK_MAPPINGS | NO_CONT_MAPPINGS; > @@ -481,15 +482,9 @@ static void __init map_mem(pgd_t *pgdp) > #endif > > /* map all the memory banks */ > - for_each_memblock(memory, reg) { > - phys_addr_t start = reg->base; > - phys_addr_t end = start + reg->size; > - > + for_each_mem_range(i, &start, &end) { > if (start >= end) > break; > - if (memblock_is_nomap(reg)) > - continue; > - > __map_memblock(pgdp, start, end, PAGE_KERNEL, flags); > } > > diff --git a/arch/c6x/kernel/setup.c b/arch/c6x/kernel/setup.c > index 8ef35131f999..9254c3b794a5 100644 > --- a/arch/c6x/kernel/setup.c > +++ b/arch/c6x/kernel/setup.c > @@ -287,7 +287,8 @@ notrace void __init machine_init(unsigned long dt_ptr) > > void __init setup_arch(char **cmdline_p) > { > - struct memblock_region *reg; > + phys_addr_t start, end; > + u64 i; > > printk(KERN_INFO "Initializing kernel\n"); > > @@ -351,9 +352,9 @@ void __init setup_arch(char **cmdline_p) > disable_caching(ram_start, ram_end - 1); > > /* Set caching of external RAM used by Linux */ > - for_each_memblock(memory, reg) > - enable_caching(CACHE_REGION_START(reg->base), > - CACHE_REGION_START(reg->base + reg->size - > 1)); > + for_each_mem_range(i, &start, &end) > + enable_caching(CACHE_REGION_START(start), > + CACHE_REGION_START(end - 1)); > > #ifdef CONFIG_BLK_DEV_INITRD > /* > diff --git a/arch/microblaze/mm/init.c b/arch/microblaze/mm/init.c > index 49e0c241f9b1..15403b5adfcf 100644 > --- a/arch/microblaze/mm/init.c > +++ b/arch/microblaze/mm/init.c > @@ -106,13 +106,14 @@ static void __init paging_init(void) > void __init setup_memory(void) > { > #ifndef CONFIG_MMU > - struct memblock_region *reg; > u32 kernel_align_start, kernel_align_size; > + phys_addr_t start, end; > + u64 i; > > /* Find main memory where is the kernel */ > - for_each_memblock(memory, reg) { > - memory_start = (u32)reg->base; > - lowmem_size = reg->size; > + for_each_mem_range(i, &start, &end) { > + memory_start = start; > + lowmem_size = end - start; > if ((memory_start <= (u32)_text) && > ((u32)_text <= (memory_start + lowmem_size - 1))) { > memory_size = lowmem_size; > diff --git a/arch/mips/cavium-octeon/dma-octeon.c > b/arch/mips/cavium-octeon/dma-octeon.c > index 14ea680d180e..d938c1f7c1e1 100644 > --- a/arch/mips/cavium-octeon/dma-octeon.c > +++ b/arch/mips/cavium-octeon/dma-octeon.c > @@ -190,25 +190,25 @@ char *octeon_swiotlb; > > void __init plat_swiotlb_setup(void) > { > - struct memblock_region *mem; > + phys_addr_t start, end; > phys_addr_t max_addr; > phys_addr_t addr_size; > size_t swiotlbsize; > unsigned long swiotlb_nslabs; > + u64 i; > > max_addr = 0; > addr_size = 0; > > - for_each_memblock(memory, mem) { > + for_each_mem_range(i, &start, &end) { > /* These addresses map low for PCI. */ > if (mem->base > 0x410000000ull && !OCTEON_IS_OCTEON2()) > continue; > > - addr_size += mem->size; > - > - if (max_addr < mem->base + mem->size) > - max_addr = mem->base + mem->size; > + addr_size += (end - start); > > + if (max_addr < end) > + max_addr = end; > } > > swiotlbsize = PAGE_SIZE; > diff --git a/arch/mips/kernel/setup.c b/arch/mips/kernel/setup.c > index 7b537fa2035d..eaac1b66026d 100644 > --- a/arch/mips/kernel/setup.c > +++ b/arch/mips/kernel/setup.c > @@ -300,8 +300,9 @@ static void __init bootmem_init(void) > > static void __init bootmem_init(void) > { > - struct memblock_region *mem; > phys_addr_t ramstart, ramend; > + phys_addr_t start, end; > + u64 i; > > ramstart = memblock_start_of_DRAM(); > ramend = memblock_end_of_DRAM(); > @@ -338,18 +339,13 @@ static void __init bootmem_init(void) > > min_low_pfn = ARCH_PFN_OFFSET; > max_pfn = PFN_DOWN(ramend); > - for_each_memblock(memory, mem) { > - unsigned long start = memblock_region_memory_base_pfn(mem); > - unsigned long end = memblock_region_memory_end_pfn(mem); > - > + for_each_mem_range(i, &start, &end) { > /* > * Skip highmem here so we get an accurate max_low_pfn if > low > * memory stops short of high memory. > * If the region overlaps HIGHMEM_START, end is clipped so > * max_pfn excludes the highmem portion. > */ > - if (memblock_is_nomap(mem)) > - continue; > if (start >= PFN_DOWN(HIGHMEM_START)) > continue; > if (end > PFN_DOWN(HIGHMEM_START)) > @@ -458,13 +454,12 @@ early_param("memmap", early_parse_memmap); > unsigned long setup_elfcorehdr, setup_elfcorehdr_size; > static int __init early_parse_elfcorehdr(char *p) > { > - struct memblock_region *mem; > + phys_addr_t start, end; > + u64 i; > > setup_elfcorehdr = memparse(p, &p); > > - for_each_memblock(memory, mem) { > - unsigned long start = mem->base; > - unsigned long end = start + mem->size; > + for_each_mem_range(i, &start, &end) { > if (setup_elfcorehdr >= start && setup_elfcorehdr < end) { > /* > * Reserve from the elf core header to the end of > @@ -728,7 +723,8 @@ static void __init arch_mem_init(char **cmdline_p) > > static void __init resource_init(void) > { > - struct memblock_region *region; > + phys_addr_t start, end; > + u64 i; > > if (UNCAC_BASE != IO_BASE) > return; > @@ -740,9 +736,7 @@ static void __init resource_init(void) > bss_resource.start = __pa_symbol(&__bss_start); > bss_resource.end = __pa_symbol(&__bss_stop) - 1; > > - for_each_memblock(memory, region) { > - phys_addr_t start = > PFN_PHYS(memblock_region_memory_base_pfn(region)); > - phys_addr_t end = > PFN_PHYS(memblock_region_memory_end_pfn(region)) - 1; > + for_each_mem_range(i, &start, &end) { > struct resource *res; > > res = memblock_alloc(sizeof(struct resource), > SMP_CACHE_BYTES); > @@ -751,7 +745,12 @@ static void __init resource_init(void) > sizeof(struct resource)); > > res->start = start; > - res->end = end; > + /* > + * In memblock, end points to the first byte after the > + * range while in resourses, end points to the last byte in > + * the range. > + */ > + res->end = end - 1; > res->flags = IORESOURCE_SYSTEM_RAM | IORESOURCE_BUSY; > res->name = "System RAM"; > > diff --git a/arch/openrisc/mm/init.c b/arch/openrisc/mm/init.c > index 3d7c79c7745d..8348feaaf46e 100644 > --- a/arch/openrisc/mm/init.c > +++ b/arch/openrisc/mm/init.c > @@ -64,6 +64,7 @@ extern const char _s_kernel_ro[], _e_kernel_ro[]; > */ > static void __init map_ram(void) > { > + phys_addr_t start, end; > unsigned long v, p, e; > pgprot_t prot; > pgd_t *pge; > @@ -71,6 +72,7 @@ static void __init map_ram(void) > pud_t *pue; > pmd_t *pme; > pte_t *pte; > + u64 i; > /* These mark extents of read-only kernel pages... > * ...from vmlinux.lds.S > */ > @@ -78,9 +80,9 @@ static void __init map_ram(void) > > v = PAGE_OFFSET; > > - for_each_memblock(memory, region) { > - p = (u32) region->base & PAGE_MASK; > - e = p + (u32) region->size; > + for_each_mem_range(i, &start, &end) { > + p = (u32) start & PAGE_MASK; > + e = (u32) end; > > v = (u32) __va(p); > pge = pgd_offset_k(v); > diff --git a/arch/powerpc/kernel/fadump.c b/arch/powerpc/kernel/fadump.c > index fdbafe417139..435b98d069eb 100644 > --- a/arch/powerpc/kernel/fadump.c > +++ b/arch/powerpc/kernel/fadump.c > @@ -180,13 +180,13 @@ int is_fadump_active(void) > */ > static bool is_fadump_mem_area_contiguous(u64 d_start, u64 d_end) > { > - struct memblock_region *reg; > + phys_addr_t reg_start, reg_end; > bool ret = false; > - u64 start, end; > + u64 i, start, end; > > - for_each_memblock(memory, reg) { > - start = max_t(u64, d_start, reg->base); > - end = min_t(u64, d_end, (reg->base + reg->size)); > + for_each_mem_range(i, ®_start, ®_end) { > + start = max_t(u64, d_start, reg_start); > + end = min_t(u64, d_end, reg_end)); > if (d_start < end) { > /* Memory hole from d_start to start */ > if (start > d_start) > @@ -413,7 +413,7 @@ static int __init fadump_get_boot_mem_regions(void) > { > unsigned long base, size, cur_size, hole_size, last_end; > unsigned long mem_size = fw_dump.boot_memory_size; > - struct memblock_region *reg; > + phys_addr_t reg_start, reg_end; > int ret = 1; > > fw_dump.boot_mem_regs_cnt = 0; > @@ -421,9 +421,8 @@ static int __init fadump_get_boot_mem_regions(void) > last_end = 0; > hole_size = 0; > cur_size = 0; > - for_each_memblock(memory, reg) { > - base = reg->base; > - size = reg->size; > + for_each_mem_range(i, ®_start, ®_end) { > + size = reg_end - reg_start; > hole_size += (base - last_end); > > if ((cur_size + size) >= mem_size) { > @@ -959,9 +958,8 @@ static int fadump_init_elfcore_header(char *bufp) > */ > static int fadump_setup_crash_memory_ranges(void) > { > - struct memblock_region *reg; > - u64 start, end; > - int i, ret; > + u64 i, start, end; > + int ret; > > pr_debug("Setup crash memory ranges.\n"); > crash_mrange_info.mem_range_cnt = 0; > @@ -979,10 +977,7 @@ static int fadump_setup_crash_memory_ranges(void) > return ret; > } > > - for_each_memblock(memory, reg) { > - start = (u64)reg->base; > - end = start + (u64)reg->size; > - > + for_each_mem_range(i, &start, end) { >
I don't know anything about this code, but from pure pattern matching it looks like you missed a & here. /* > * skip the memory chunk that is already added > * (0 through boot_memory_top). > diff --git a/arch/powerpc/mm/book3s64/hash_utils.c > b/arch/powerpc/mm/book3s64/hash_utils.c > index 468169e33c86..9ba76b075b11 100644 > --- a/arch/powerpc/mm/book3s64/hash_utils.c > +++ b/arch/powerpc/mm/book3s64/hash_utils.c > @@ -7,7 +7,7 @@ > * > * SMP scalability work: > * Copyright (C) 2001 Anton Blanchard <an...@au.ibm.com>, IBM > - * > + * > * Module name: htab.c > * > * Description: > @@ -862,8 +862,8 @@ static void __init htab_initialize(void) > unsigned long table; > unsigned long pteg_count; > unsigned long prot; > - unsigned long base = 0, size = 0; > - struct memblock_region *reg; > + phys_addr_t base = 0, size = 0, end; > + u64 i; > > DBG(" -> htab_initialize()\n"); > > @@ -879,7 +879,7 @@ static void __init htab_initialize(void) > /* > * Calculate the required size of the htab. We want the number of > * PTEGs to equal one half the number of real pages. > - */ > + */ > htab_size_bytes = htab_get_table_size(); > pteg_count = htab_size_bytes >> 7; > > @@ -889,7 +889,7 @@ static void __init htab_initialize(void) > firmware_has_feature(FW_FEATURE_PS3_LV1)) { > /* Using a hypervisor which owns the htab */ > htab_address = NULL; > - _SDR1 = 0; > + _SDR1 = 0; > #ifdef CONFIG_FA_DUMP > /* > * If firmware assisted dump is active firmware preserves > @@ -955,9 +955,9 @@ static void __init htab_initialize(void) > #endif /* CONFIG_DEBUG_PAGEALLOC */ > > /* create bolted the linear mapping in the hash table */ > - for_each_memblock(memory, reg) { > - base = (unsigned long)__va(reg->base); > - size = reg->size; > + for_each_mem_range(i, &base, &end) { > + size = end - base; > + base = (unsigned long)__va(base); > > DBG("creating mapping for region: %lx..%lx (prot: %lx)\n", > base, size, prot); > diff --git a/arch/powerpc/mm/book3s64/radix_pgtable.c > b/arch/powerpc/mm/book3s64/radix_pgtable.c > index bb00e0cba119..65657b920847 100644 > --- a/arch/powerpc/mm/book3s64/radix_pgtable.c > +++ b/arch/powerpc/mm/book3s64/radix_pgtable.c > @@ -318,28 +318,27 @@ static int __meminit > create_physical_mapping(unsigned long start, > static void __init radix_init_pgtable(void) > { > unsigned long rts_field; > - struct memblock_region *reg; > + phys_addr_t start, end; > + u64 i; > > /* We don't support slb for radix */ > mmu_slb_size = 0; > /* > * Create the linear mapping, using standard page size for now > */ > - for_each_memblock(memory, reg) { > + for_each_mem_range(i, &start, &end) { > /* > * The memblock allocator is up at this point, so the > * page tables will be allocated within the range. No > * need or a node (which we don't have yet). > */ > > - if ((reg->base + reg->size) >= RADIX_VMALLOC_START) { > + if (end >= RADIX_VMALLOC_START) { > pr_warn("Outside the supported range\n"); > continue; > } > > - WARN_ON(create_physical_mapping(reg->base, > - reg->base + reg->size, > - -1, PAGE_KERNEL)); > + WARN_ON(create_physical_mapping(start, end, -1, > PAGE_KERNEL)); > } > > /* Find out how many PID bits are supported */ > diff --git a/arch/powerpc/mm/kasan/kasan_init_32.c > b/arch/powerpc/mm/kasan/kasan_init_32.c > index 0760e1e754e4..6e73434e4e41 100644 > --- a/arch/powerpc/mm/kasan/kasan_init_32.c > +++ b/arch/powerpc/mm/kasan/kasan_init_32.c > @@ -120,11 +120,11 @@ static void __init > kasan_unmap_early_shadow_vmalloc(void) > static void __init kasan_mmu_init(void) > { > int ret; > - struct memblock_region *reg; > + phys_addr_t base, end; > + u64 i; > > - for_each_memblock(memory, reg) { > - phys_addr_t base = reg->base; > - phys_addr_t top = min(base + reg->size, total_lowmem); > + for_each_mem_range(i, &base, &end) { > + phys_addr_t top = min(end, total_lowmem); > > if (base >= top) > continue; > diff --git a/arch/powerpc/mm/mem.c b/arch/powerpc/mm/mem.c > index 38d1acd7c8ef..0248b6d58fcd 100644 > --- a/arch/powerpc/mm/mem.c > +++ b/arch/powerpc/mm/mem.c > @@ -593,20 +593,24 @@ void flush_icache_user_page(struct vm_area_struct > *vma, struct page *page, > */ > static int __init add_system_ram_resources(void) > { > - struct memblock_region *reg; > + phys_addr_t start, end; > + u64 i; > > - for_each_memblock(memory, reg) { > + for_each_mem_range(i, &start, &end) { > struct resource *res; > - unsigned long base = reg->base; > - unsigned long size = reg->size; > > res = kzalloc(sizeof(struct resource), GFP_KERNEL); > WARN_ON(!res); > > if (res) { > res->name = "System RAM"; > - res->start = base; > - res->end = base + size - 1; > + res->start = start; > + /* > + * In memblock, end points to the first byte after > + * the range while in resourses, end points to the > + * last byte in the range. > + */ > + res->end = end - 1; > res->flags = IORESOURCE_SYSTEM_RAM | > IORESOURCE_BUSY; > WARN_ON(request_resource(&iomem_resource, res) < > 0); > } > diff --git a/arch/powerpc/mm/pgtable_32.c b/arch/powerpc/mm/pgtable_32.c > index 6eb4eab79385..079159e97bca 100644 > --- a/arch/powerpc/mm/pgtable_32.c > +++ b/arch/powerpc/mm/pgtable_32.c > @@ -123,11 +123,11 @@ static void __init __mapin_ram_chunk(unsigned long > offset, unsigned long top) > > void __init mapin_ram(void) > { > - struct memblock_region *reg; > + phys_addr_t base, end; > + u64 i; > > - for_each_memblock(memory, reg) { > - phys_addr_t base = reg->base; > - phys_addr_t top = min(base + reg->size, total_lowmem); > + for_each_mem_range(i, &base, &end) { > + phys_addr_t top = min(end, total_lowmem); > > if (base >= top) > continue; > diff --git a/arch/riscv/mm/init.c b/arch/riscv/mm/init.c > index 7440ba2cdaaa..2abe1165fe56 100644 > --- a/arch/riscv/mm/init.c > +++ b/arch/riscv/mm/init.c > @@ -145,21 +145,22 @@ static phys_addr_t dtb_early_pa __initdata; > > void __init setup_bootmem(void) > { > - struct memblock_region *reg; > + phys_addr_t start, end; > phys_addr_t mem_size = 0; > phys_addr_t total_mem = 0; > phys_addr_t mem_start, end = 0; > phys_addr_t vmlinux_end = __pa_symbol(&_end); > phys_addr_t vmlinux_start = __pa_symbol(&_start); > + u64 i; > > /* Find the memory region containing the kernel */ > - for_each_memblock(memory, reg) { > - end = reg->base + reg->size; > + for_each_mem_range(i, &start, &end) { > + phys_addr_t size = end - start; > if (!total_mem) > - mem_start = reg->base; > - if (reg->base <= vmlinux_start && vmlinux_end <= end) > - BUG_ON(reg->size == 0); > - total_mem = total_mem + reg->size; > + mem_start = start; > + if (start <= vmlinux_start && vmlinux_end <= end) > + BUG_ON(size == 0); > + total_mem = total_mem + size; > } > > /* > @@ -456,7 +457,7 @@ static void __init setup_vm_final(void) > { > uintptr_t va, map_size; > phys_addr_t pa, start, end; > - struct memblock_region *reg; > + u64 i; > > /* Set mmu_enabled flag */ > mmu_enabled = true; > @@ -467,14 +468,9 @@ static void __init setup_vm_final(void) > PGDIR_SIZE, PAGE_TABLE); > > /* Map all memory banks */ > - for_each_memblock(memory, reg) { > - start = reg->base; > - end = start + reg->size; > - > + for_each_mem_range(i, &start, &end) { > if (start >= end) > break; > - if (memblock_is_nomap(reg)) > - continue; > if (start <= __pa(PAGE_OFFSET) && > __pa(PAGE_OFFSET) < end) > start = __pa(PAGE_OFFSET); > diff --git a/arch/riscv/mm/kasan_init.c b/arch/riscv/mm/kasan_init.c > index 87b4ab3d3c77..12ddd1f6bf70 100644 > --- a/arch/riscv/mm/kasan_init.c > +++ b/arch/riscv/mm/kasan_init.c > @@ -85,16 +85,16 @@ static void __init populate(void *start, void *end) > > void __init kasan_init(void) > { > - struct memblock_region *reg; > - unsigned long i; > + phys_addr_t _start, _end; > + u64 i; > > kasan_populate_early_shadow((void *)KASAN_SHADOW_START, > (void *)kasan_mem_to_shadow((void *) > > VMALLOC_END)); > > - for_each_memblock(memory, reg) { > - void *start = (void *)__va(reg->base); > - void *end = (void *)__va(reg->base + reg->size); > + for_each_mem_range(i, &_start, &_end) { > + void *start = (void *)_start; > + void *end = (void *)_end; > > if (start >= end) > break; > diff --git a/arch/s390/kernel/setup.c b/arch/s390/kernel/setup.c > index 8b284cf6e199..b6c4a0c5ff86 100644 > --- a/arch/s390/kernel/setup.c > +++ b/arch/s390/kernel/setup.c > @@ -198,7 +198,7 @@ static void __init conmode_default(void) > cpcmd("QUERY TERM", query_buffer, 1024, NULL); > ptr = strstr(query_buffer, "CONMODE"); > /* > - * Set the conmode to 3215 so that the device recognition > + * Set the conmode to 3215 so that the device recognition > * will set the cu_type of the console to 3215. If the > * conmode is 3270 and we don't set it back then both > * 3215 and the 3270 driver will try to access the console > @@ -258,7 +258,7 @@ static inline void setup_zfcpdump(void) {} > > /* > * Reboot, halt and power_off stubs. They just call _machine_restart, > - * _machine_halt or _machine_power_off. > + * _machine_halt or _machine_power_off. > */ > > void machine_restart(char *command) > @@ -484,8 +484,9 @@ static struct resource __initdata > *standard_resources[] = { > static void __init setup_resources(void) > { > struct resource *res, *std_res, *sub_res; > - struct memblock_region *reg; > + phys_addr_t start, end; > int j; > + u64 i; > > code_resource.start = (unsigned long) _text; > code_resource.end = (unsigned long) _etext - 1; > @@ -494,7 +495,7 @@ static void __init setup_resources(void) > bss_resource.start = (unsigned long) __bss_start; > bss_resource.end = (unsigned long) __bss_stop - 1; > > - for_each_memblock(memory, reg) { > + for_each_mem_range(i, &start, &end) { > res = memblock_alloc(sizeof(*res), 8); > if (!res) > panic("%s: Failed to allocate %zu bytes > align=0x%x\n", > @@ -502,8 +503,13 @@ static void __init setup_resources(void) > res->flags = IORESOURCE_BUSY | IORESOURCE_SYSTEM_RAM; > > res->name = "System RAM"; > - res->start = reg->base; > - res->end = reg->base + reg->size - 1; > + res->start = start; > + /* > + * In memblock, end points to the first byte after the > + * range while in resourses, end points to the last byte in > + * the range. > + */ > + res->end = end - 1; > request_resource(&iomem_resource, res); > > for (j = 0; j < ARRAY_SIZE(standard_resources); j++) { > @@ -819,14 +825,15 @@ static void __init reserve_kernel(void) > > static void __init setup_memory(void) > { > - struct memblock_region *reg; > + phys_addr_t start, end; > + u64 i; > > /* > * Init storage key for present memory > */ > - for_each_memblock(memory, reg) { > - storage_key_init_range(reg->base, reg->base + reg->size); > - } > + for_each_mem_range(i, &start, &end) > + storage_key_init_range(start, end); > + > psw_set_key(PAGE_DEFAULT_KEY); > > /* Only cosmetics */ > diff --git a/arch/s390/mm/vmem.c b/arch/s390/mm/vmem.c > index 8b6282cf7d13..30076ecc3eb7 100644 > --- a/arch/s390/mm/vmem.c > +++ b/arch/s390/mm/vmem.c > @@ -399,10 +399,11 @@ int vmem_add_mapping(unsigned long start, unsigned > long size) > */ > void __init vmem_map_init(void) > { > - struct memblock_region *reg; > + phys_addr_t start, end; > + u64 i; > > - for_each_memblock(memory, reg) > - vmem_add_mem(reg->base, reg->size); > + for_each_mem_range(i, &start, &end) > + vmem_add_mem(start, end - start); > __set_memory((unsigned long)_stext, > (unsigned long)(_etext - _stext) >> PAGE_SHIFT, > SET_MEMORY_RO | SET_MEMORY_X); > @@ -428,16 +429,17 @@ void __init vmem_map_init(void) > */ > static int __init vmem_convert_memory_chunk(void) > { > - struct memblock_region *reg; > + phys_addr_t start, end; > struct memory_segment *seg; > + u64 i; > > mutex_lock(&vmem_mutex); > - for_each_memblock(memory, reg) { > + for_each_mem_range(i, &start, &end) { > seg = kzalloc(sizeof(*seg), GFP_KERNEL); > if (!seg) > panic("Out of memory...\n"); > - seg->start = reg->base; > - seg->size = reg->size; > + seg->start = start; > + seg->size = end - start; > insert_memory_segment(seg); > } > mutex_unlock(&vmem_mutex); > diff --git a/arch/sparc/mm/init_64.c b/arch/sparc/mm/init_64.c > index 02e6e5e0f106..de63c002638e 100644 > --- a/arch/sparc/mm/init_64.c > +++ b/arch/sparc/mm/init_64.c > @@ -1192,18 +1192,14 @@ int of_node_to_nid(struct device_node *dp) > > static void __init add_node_ranges(void) > { > - struct memblock_region *reg; > + phys_addr_t start, end; > unsigned long prev_max; > + u64 i; > > memblock_resized: > prev_max = memblock.memory.max; > > - for_each_memblock(memory, reg) { > - unsigned long size = reg->size; > - unsigned long start, end; > - > - start = reg->base; > - end = start + size; > + for_each_mem_range(i, &start, &end) { > while (start < end) { > unsigned long this_end; > int nid; > @@ -1211,7 +1207,7 @@ static void __init add_node_ranges(void) > this_end = memblock_nid_range(start, end, &nid); > > numadbg("Setting memblock NUMA node nid[%d] " > - "start[%lx] end[%lx]\n", > + "start[%llx] end[%lx]\n", > nid, start, this_end); > > memblock_set_node(start, this_end - start, > diff --git a/drivers/bus/mvebu-mbus.c b/drivers/bus/mvebu-mbus.c > index 5b2a11a88951..2519ceede64b 100644 > --- a/drivers/bus/mvebu-mbus.c > +++ b/drivers/bus/mvebu-mbus.c > @@ -610,23 +610,23 @@ static unsigned int > armada_xp_mbus_win_remap_offset(int win) > static void __init > mvebu_mbus_find_bridge_hole(uint64_t *start, uint64_t *end) > { > - struct memblock_region *r; > - uint64_t s = 0; > + phys_addr_t reg_start, reg_end; > + uint64_t i, s = 0; > > - for_each_memblock(memory, r) { > + for_each_mem_range(i, ®_start, ®_end) { > /* > * This part of the memory is above 4 GB, so we don't > * care for the MBus bridge hole. > */ > - if (r->base >= 0x100000000ULL) > + if (reg_start >= 0x100000000ULL) > continue; > > /* > * The MBus bridge hole is at the end of the RAM under > * the 4 GB limit. > */ > - if (r->base + r->size > s) > - s = r->base + r->size; > + if (reg_end > s) > + s = reg_end; > } > > *start = s; > diff --git a/drivers/s390/char/zcore.c b/drivers/s390/char/zcore.c > index 08f812475f5e..484b1ec9a1bc 100644 > --- a/drivers/s390/char/zcore.c > +++ b/drivers/s390/char/zcore.c > @@ -148,18 +148,19 @@ static ssize_t zcore_memmap_read(struct file *filp, > char __user *buf, > > static int zcore_memmap_open(struct inode *inode, struct file *filp) > { > - struct memblock_region *reg; > + phys_addr_t start, end; > char *buf; > int i = 0; > + u64 r; > > buf = kcalloc(memblock.memory.cnt, CHUNK_INFO_SIZE, GFP_KERNEL); > if (!buf) { > return -ENOMEM; > } > - for_each_memblock(memory, reg) { > + for_each_mem_range(r, &start, &end) { > sprintf(buf + (i++ * CHUNK_INFO_SIZE), "%016llx %016llx ", > - (unsigned long long) reg->base, > - (unsigned long long) reg->size); > + (unsigned long long) start, > + (unsigned long long) (end - start)); > } > filp->private_data = buf; > return nonseekable_open(inode, filp); > -- > 2.26.2 > > > _______________________________________________ > linux-riscv mailing list > linux-ri...@lists.infradead.org > http://lists.infradead.org/mailman/listinfo/linux-riscv >
_______________________________________________ iommu mailing list iommu@lists.linux-foundation.org https://lists.linuxfoundation.org/mailman/listinfo/iommu