end_pfn 220 arch/alpha/mm/numa.c unsigned long end_pfn = start_pfn + NODE_DATA(nid)->node_present_pages; end_pfn 222 arch/alpha/mm/numa.c if (dma_local_pfn >= end_pfn - start_pfn) end_pfn 223 arch/alpha/mm/numa.c zones_size[ZONE_DMA] = end_pfn - start_pfn; end_pfn 226 arch/alpha/mm/numa.c zones_size[ZONE_NORMAL] = (end_pfn - start_pfn) - dma_local_pfn; end_pfn 327 arch/arm/mm/init.c free_memmap(unsigned long start_pfn, unsigned long end_pfn) end_pfn 336 arch/arm/mm/init.c end_pg = pfn_to_page(end_pfn - 1) + 1; end_pfn 462 arch/arm64/mm/init.c static inline void free_memmap(unsigned long start_pfn, unsigned long end_pfn) end_pfn 471 arch/arm64/mm/init.c end_pg = pfn_to_page(end_pfn - 1) + 1; end_pfn 218 arch/arm64/mm/numa.c static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) end_pfn 225 arch/arm64/mm/numa.c if (start_pfn >= end_pfn) end_pfn 246 arch/arm64/mm/numa.c NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; end_pfn 363 arch/arm64/mm/numa.c unsigned long start_pfn, end_pfn; end_pfn 365 arch/arm64/mm/numa.c get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); end_pfn 366 arch/arm64/mm/numa.c setup_node_data(nid, start_pfn, end_pfn); end_pfn 174 arch/microblaze/mm/init.c unsigned long start_pfn, end_pfn; end_pfn 177 arch/microblaze/mm/init.c end_pfn = memblock_region_memory_end_pfn(reg); end_pfn 179 arch/microblaze/mm/init.c (end_pfn - start_pfn) << PAGE_SHIFT, end_pfn 123 arch/mips/loongson64/loongson-3/numa.c u64 node_id, node_psize, start_pfn, end_pfn, mem_start, mem_size; end_pfn 139 arch/mips/loongson64/loongson-3/numa.c end_pfn = start_pfn + node_psize; end_pfn 144 arch/mips/loongson64/loongson-3/numa.c start_pfn, end_pfn, num_physpages); end_pfn 146 arch/mips/loongson64/loongson-3/numa.c PFN_PHYS(end_pfn - start_pfn), node); end_pfn 151 arch/mips/loongson64/loongson-3/numa.c end_pfn = start_pfn + node_psize; end_pfn 156 arch/mips/loongson64/loongson-3/numa.c start_pfn, end_pfn, num_physpages); end_pfn 158 arch/mips/loongson64/loongson-3/numa.c PFN_PHYS(end_pfn - start_pfn), node); end_pfn 173 arch/mips/loongson64/loongson-3/numa.c unsigned long start_pfn, end_pfn; end_pfn 179 arch/mips/loongson64/loongson-3/numa.c get_pfn_range_for_nid(node, &start_pfn, &end_pfn); end_pfn 181 arch/mips/loongson64/loongson-3/numa.c node, start_pfn, end_pfn); end_pfn 186 arch/mips/loongson64/loongson-3/numa.c NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; end_pfn 193 arch/mips/loongson64/loongson-3/numa.c max_low_pfn = end_pfn; end_pfn 387 arch/mips/sgi-ip27/ip27-memory.c unsigned long start_pfn, end_pfn; end_pfn 389 arch/mips/sgi-ip27/ip27-memory.c get_pfn_range_for_nid(node, &start_pfn, &end_pfn); end_pfn 398 arch/mips/sgi-ip27/ip27-memory.c NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; end_pfn 405 arch/mips/sgi-ip27/ip27-memory.c free_bootmem_with_active_regions(node, end_pfn); end_pfn 229 arch/parisc/mm/init.c unsigned long end_pfn; end_pfn 233 arch/parisc/mm/init.c end_pfn = pmem_ranges[0].start_pfn + pmem_ranges[0].pages; end_pfn 236 arch/parisc/mm/init.c hole_pages = pmem_ranges[i].start_pfn - end_pfn; end_pfn 238 arch/parisc/mm/init.c pmem_holes[npmem_holes].start_pfn = end_pfn; end_pfn 240 arch/parisc/mm/init.c end_pfn += hole_pages; end_pfn 242 arch/parisc/mm/init.c end_pfn += pmem_ranges[i].pages; end_pfn 245 arch/parisc/mm/init.c pmem_ranges[0].pages = end_pfn - pmem_ranges[0].start_pfn; end_pfn 1112 arch/powerpc/kernel/fadump.c unsigned long end_pfn) end_pfn 1118 arch/powerpc/kernel/fadump.c PFN_PHYS(start_pfn), PFN_PHYS(end_pfn)); end_pfn 1120 arch/powerpc/kernel/fadump.c for (pfn = start_pfn; pfn < end_pfn; pfn++) { end_pfn 181 arch/powerpc/mm/init_64.c if ((start_pfn + nr_pfn) > altmap->end_pfn) end_pfn 85 arch/powerpc/mm/numa.c static int __init fake_numa_create_new_node(unsigned long end_pfn, end_pfn 116 arch/powerpc/mm/numa.c if ((end_pfn << PAGE_SHIFT) > mem) { end_pfn 744 arch/powerpc/mm/numa.c unsigned long start_pfn, end_pfn; end_pfn 755 arch/powerpc/mm/numa.c end_pfn = memblock_region_memory_end_pfn(reg); end_pfn 757 arch/powerpc/mm/numa.c fake_numa_create_new_node(end_pfn, &nid); end_pfn 759 arch/powerpc/mm/numa.c PFN_PHYS(end_pfn - start_pfn), end_pfn 801 arch/powerpc/mm/numa.c static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) end_pfn 803 arch/powerpc/mm/numa.c u64 spanned_pages = end_pfn - start_pfn; end_pfn 891 arch/powerpc/mm/numa.c unsigned long start_pfn, end_pfn; end_pfn 893 arch/powerpc/mm/numa.c get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); end_pfn 894 arch/powerpc/mm/numa.c setup_node_data(nid, start_pfn, end_pfn); end_pfn 97 arch/powerpc/platforms/powernv/memtrace.c u64 start_pfn, end_pfn, nr_pages, pfn; end_pfn 105 arch/powerpc/platforms/powernv/memtrace.c end_pfn = node_end_pfn(nid); end_pfn 109 arch/powerpc/platforms/powernv/memtrace.c end_pfn = round_down(end_pfn - nr_pages, nr_pages); end_pfn 112 arch/powerpc/platforms/powernv/memtrace.c for (base_pfn = end_pfn; base_pfn > start_pfn; base_pfn -= nr_pages) { end_pfn 120 arch/powerpc/platforms/powernv/memtrace.c end_pfn = base_pfn + nr_pages; end_pfn 121 arch/powerpc/platforms/powernv/memtrace.c for (pfn = base_pfn; pfn < end_pfn; pfn += bytes>> PAGE_SHIFT) { end_pfn 138 arch/riscv/mm/init.c unsigned long end_pfn = memblock_region_memory_end_pfn(reg); end_pfn 141 arch/riscv/mm/init.c PFN_PHYS(end_pfn - start_pfn), end_pfn 56 arch/s390/kernel/early.c unsigned long end_pfn, init_pfn; end_pfn 58 arch/s390/kernel/early.c end_pfn = PFN_UP(__pa(_end)); end_pfn 60 arch/s390/kernel/early.c for (init_pfn = 0 ; init_pfn < end_pfn; init_pfn++) end_pfn 103 arch/s390/numa/numa.c unsigned long start_pfn, end_pfn; end_pfn 108 arch/s390/numa/numa.c end_pfn = 0; end_pfn 112 arch/s390/numa/numa.c if (t_end > end_pfn) end_pfn 113 arch/s390/numa/numa.c end_pfn = t_end; end_pfn 115 arch/s390/numa/numa.c NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; end_pfn 43 arch/sh/include/asm/mmzone.h unsigned long end_pfn); end_pfn 197 arch/sh/kernel/setup.c unsigned long end_pfn) end_pfn 205 arch/sh/kernel/setup.c end = end_pfn << PAGE_SHIFT; end_pfn 214 arch/sh/kernel/setup.c start_pfn, end_pfn); end_pfn 238 arch/sh/kernel/setup.c memblock_set_node(PFN_PHYS(start_pfn), PFN_PHYS(end_pfn - start_pfn), end_pfn 22 arch/sh/kernel/swsusp.c unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; end_pfn 24 arch/sh/kernel/swsusp.c return (pfn >= begin_pfn) && (pfn < end_pfn); end_pfn 201 arch/sh/mm/init.c unsigned long start_pfn, end_pfn; end_pfn 203 arch/sh/mm/init.c get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); end_pfn 215 arch/sh/mm/init.c NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; end_pfn 224 arch/sh/mm/init.c unsigned long start_pfn, end_pfn; end_pfn 226 arch/sh/mm/init.c end_pfn = memblock_region_memory_end_pfn(reg); end_pfn 227 arch/sh/mm/init.c __add_active_range(0, start_pfn, end_pfn); end_pfn 28 arch/sh/mm/numa.c unsigned long start_pfn, end_pfn; end_pfn 34 arch/sh/mm/numa.c end_pfn = PFN_DOWN(end); end_pfn 41 arch/sh/mm/numa.c __add_active_range(nid, start_pfn, end_pfn); end_pfn 52 arch/sh/mm/numa.c NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; end_pfn 67 arch/sparc/mm/init_32.c unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; end_pfn 69 arch/sparc/mm/init_32.c if (end_pfn <= max_low_pfn) end_pfn 75 arch/sparc/mm/init_32.c nr += end_pfn - start_pfn; end_pfn 233 arch/sparc/mm/init_32.c static void map_high_region(unsigned long start_pfn, unsigned long end_pfn) end_pfn 238 arch/sparc/mm/init_32.c printk("mapping high region %08lx - %08lx\n", start_pfn, end_pfn); end_pfn 241 arch/sparc/mm/init_32.c for (tmp = start_pfn; tmp < end_pfn; tmp++) end_pfn 282 arch/sparc/mm/init_32.c unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; end_pfn 284 arch/sparc/mm/init_32.c if (end_pfn <= highstart_pfn) end_pfn 290 arch/sparc/mm/init_32.c map_high_region(start_pfn, end_pfn); end_pfn 1090 arch/sparc/mm/init_64.c unsigned long start_pfn, end_pfn; end_pfn 1105 arch/sparc/mm/init_64.c get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); end_pfn 1107 arch/sparc/mm/init_64.c p->node_spanned_pages = end_pfn - start_pfn; end_pfn 1628 arch/sparc/mm/init_64.c unsigned long end_pfn; end_pfn 1630 arch/sparc/mm/init_64.c end_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT; end_pfn 1631 arch/sparc/mm/init_64.c max_pfn = max_low_pfn = end_pfn; end_pfn 1645 arch/sparc/mm/init_64.c return end_pfn; end_pfn 2291 arch/sparc/mm/init_64.c unsigned long end_pfn, shift, phys_base; end_pfn 2460 arch/sparc/mm/init_64.c last_valid_pfn = end_pfn = bootmem_init(phys_base); end_pfn 2469 arch/sparc/mm/init_64.c max_zone_pfns[ZONE_NORMAL] = end_pfn; end_pfn 145 arch/unicore32/kernel/hibernate.c unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; end_pfn 147 arch/unicore32/kernel/hibernate.c return (pfn >= begin_pfn) && (pfn < end_pfn); end_pfn 200 arch/unicore32/mm/init.c free_memmap(unsigned long start_pfn, unsigned long end_pfn) end_pfn 209 arch/unicore32/mm/init.c end_pg = pfn_to_page(end_pfn); end_pfn 76 arch/x86/include/asm/highmem.h unsigned long end_pfn); end_pfn 50 arch/x86/include/asm/mtrr.h extern int mtrr_trim_uncached_memory(unsigned long end_pfn); end_pfn 80 arch/x86/include/asm/mtrr.h static inline int mtrr_trim_uncached_memory(unsigned long end_pfn) end_pfn 74 arch/x86/include/asm/page_types.h bool pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn); end_pfn 715 arch/x86/kernel/amd_gart_64.c unsigned long start_pfn, end_pfn; end_pfn 745 arch/x86/kernel/amd_gart_64.c end_pfn = (aper_base>>PAGE_SHIFT) + (aper_size>>PAGE_SHIFT); end_pfn 748 arch/x86/kernel/amd_gart_64.c if (!pfn_range_is_mapped(start_pfn, end_pfn)) end_pfn 749 arch/x86/kernel/amd_gart_64.c init_memory_mapping(start_pfn<<PAGE_SHIFT, end_pfn<<PAGE_SHIFT); end_pfn 877 arch/x86/kernel/cpu/mtrr/cleanup.c int __init mtrr_trim_uncached_memory(unsigned long end_pfn) end_pfn 969 arch/x86/kernel/cpu/mtrr/cleanup.c if (range[i].end < end_pfn) end_pfn 971 arch/x86/kernel/cpu/mtrr/cleanup.c end_pfn); end_pfn 824 arch/x86/kernel/e820.c unsigned long end_pfn; end_pfn 830 arch/x86/kernel/e820.c end_pfn = (entry->addr + entry->size) >> PAGE_SHIFT; end_pfn 834 arch/x86/kernel/e820.c if (end_pfn > limit_pfn) { end_pfn 838 arch/x86/kernel/e820.c if (end_pfn > last_pfn) end_pfn 839 arch/x86/kernel/e820.c last_pfn = end_pfn; end_pfn 262 arch/x86/mm/init.c unsigned long start_pfn, unsigned long end_pfn, end_pfn 265 arch/x86/mm/init.c if (start_pfn < end_pfn) { end_pfn 269 arch/x86/mm/init.c mr[nr_range].end = end_pfn<<PAGE_SHIFT; end_pfn 340 arch/x86/mm/init.c unsigned long start_pfn, end_pfn, limit_pfn; end_pfn 356 arch/x86/mm/init.c end_pfn = PFN_DOWN(PMD_SIZE); end_pfn 358 arch/x86/mm/init.c end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); end_pfn 360 arch/x86/mm/init.c end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); end_pfn 362 arch/x86/mm/init.c if (end_pfn > limit_pfn) end_pfn 363 arch/x86/mm/init.c end_pfn = limit_pfn; end_pfn 364 arch/x86/mm/init.c if (start_pfn < end_pfn) { end_pfn 365 arch/x86/mm/init.c nr_range = save_mr(mr, nr_range, start_pfn, end_pfn, 0); end_pfn 366 arch/x86/mm/init.c pfn = end_pfn; end_pfn 372 arch/x86/mm/init.c end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); end_pfn 374 arch/x86/mm/init.c end_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE)); end_pfn 375 arch/x86/mm/init.c if (end_pfn > round_down(limit_pfn, PFN_DOWN(PMD_SIZE))) end_pfn 376 arch/x86/mm/init.c end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); end_pfn 379 arch/x86/mm/init.c if (start_pfn < end_pfn) { end_pfn 380 arch/x86/mm/init.c nr_range = save_mr(mr, nr_range, start_pfn, end_pfn, end_pfn 382 arch/x86/mm/init.c pfn = end_pfn; end_pfn 388 arch/x86/mm/init.c end_pfn = round_down(limit_pfn, PFN_DOWN(PUD_SIZE)); end_pfn 389 arch/x86/mm/init.c if (start_pfn < end_pfn) { end_pfn 390 arch/x86/mm/init.c nr_range = save_mr(mr, nr_range, start_pfn, end_pfn, end_pfn 393 arch/x86/mm/init.c pfn = end_pfn; end_pfn 398 arch/x86/mm/init.c end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); end_pfn 399 arch/x86/mm/init.c if (start_pfn < end_pfn) { end_pfn 400 arch/x86/mm/init.c nr_range = save_mr(mr, nr_range, start_pfn, end_pfn, end_pfn 402 arch/x86/mm/init.c pfn = end_pfn; end_pfn 408 arch/x86/mm/init.c end_pfn = limit_pfn; end_pfn 409 arch/x86/mm/init.c nr_range = save_mr(mr, nr_range, start_pfn, end_pfn, 0); end_pfn 439 arch/x86/mm/init.c static void add_pfn_range_mapped(unsigned long start_pfn, unsigned long end_pfn) end_pfn 442 arch/x86/mm/init.c nr_pfn_mapped, start_pfn, end_pfn); end_pfn 445 arch/x86/mm/init.c max_pfn_mapped = max(max_pfn_mapped, end_pfn); end_pfn 449 arch/x86/mm/init.c min(end_pfn, 1UL<<(32-PAGE_SHIFT))); end_pfn 452 arch/x86/mm/init.c bool pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn) end_pfn 458 arch/x86/mm/init.c (end_pfn <= pfn_mapped[i].end)) end_pfn 508 arch/x86/mm/init.c unsigned long start_pfn, end_pfn; end_pfn 512 arch/x86/mm/init.c for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) { end_pfn 514 arch/x86/mm/init.c u64 end = clamp_val(PFN_PHYS(end_pfn), r_start, r_end); end_pfn 899 arch/x86/mm/init.c unsigned long start_pfn, end_pfn; end_pfn 909 arch/x86/mm/init.c for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) { end_pfn 911 arch/x86/mm/init.c end_pfn = min(end_pfn, MAX_DMA_PFN); end_pfn 913 arch/x86/mm/init.c nr_pages += end_pfn - start_pfn; end_pfn 924 arch/x86/mm/init.c end_pfn = min_t(unsigned long, PFN_DOWN(end_addr), MAX_DMA_PFN); end_pfn 926 arch/x86/mm/init.c if (start_pfn < end_pfn) end_pfn 927 arch/x86/mm/init.c nr_free_pages += end_pfn - start_pfn; end_pfn 259 arch/x86/mm/init_32.c unsigned long start_pfn, end_pfn; end_pfn 270 arch/x86/mm/init_32.c end_pfn = end >> PAGE_SHIFT; end_pfn 299 arch/x86/mm/init_32.c if (pfn >= end_pfn) end_pfn 307 arch/x86/mm/init_32.c for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn; end_pfn 347 arch/x86/mm/init_32.c for (; pte_ofs < PTRS_PER_PTE && pfn < end_pfn; end_pfn 435 arch/x86/mm/init_32.c unsigned long start_pfn, unsigned long end_pfn) end_pfn 442 arch/x86/mm/init_32.c start_pfn, end_pfn); end_pfn 444 arch/x86/mm/init_32.c start_pfn, end_pfn); end_pfn 837 arch/x86/mm/init_64.c unsigned long end_pfn = PFN_UP(start + size); end_pfn 839 arch/x86/mm/init_64.c if (end_pfn > max_pfn) { end_pfn 840 arch/x86/mm/init_64.c max_pfn = end_pfn; end_pfn 841 arch/x86/mm/init_64.c max_low_pfn = end_pfn; end_pfn 34 arch/x86/mm/numa_emulation.c unsigned long end_pfn = PFN_DOWN(end); end_pfn 36 arch/x86/mm/numa_emulation.c if (start_pfn < end_pfn) end_pfn 37 arch/x86/mm/numa_emulation.c return PFN_PHYS(absent_pages_in_range(start_pfn, end_pfn)); end_pfn 428 arch/x86/mm/pat.c unsigned long end_pfn = (end + PAGE_SIZE - 1) >> PAGE_SHIFT; end_pfn 441 arch/x86/mm/pat.c if (start_pfn < end_pfn) { end_pfn 442 arch/x86/mm/pat.c ret = walk_system_ram_range(start_pfn, end_pfn - start_pfn, end_pfn 614 arch/x86/platform/efi/efi.c u64 start_pfn, end_pfn, end; end_pfn 621 arch/x86/platform/efi/efi.c end_pfn = PFN_UP(end); end_pfn 623 arch/x86/platform/efi/efi.c if (pfn_range_is_mapped(start_pfn, end_pfn)) { end_pfn 255 arch/x86/xen/setup.c unsigned long end_pfn, unsigned long nr_pages) end_pfn 260 arch/x86/xen/setup.c WARN_ON(start_pfn > end_pfn); end_pfn 263 arch/x86/xen/setup.c end = min(end_pfn, nr_pages); end_pfn 282 arch/x86/xen/setup.c set_phys_range_identity(start_pfn, end_pfn); end_pfn 388 arch/x86/xen/setup.c unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, end_pfn 393 arch/x86/xen/setup.c unsigned long n = end_pfn - start_pfn; end_pfn 435 arch/x86/xen/setup.c for (pfn = start_pfn; pfn <= max_pfn_mapped && pfn < end_pfn; pfn++) end_pfn 444 arch/x86/xen/setup.c unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, end_pfn 450 arch/x86/xen/setup.c return remap_pages + min(end_pfn, nr_pages) - start_pfn; end_pfn 454 arch/x86/xen/setup.c unsigned long (*func)(unsigned long start_pfn, unsigned long end_pfn, end_pfn 477 arch/x86/xen/setup.c unsigned long end_pfn = PFN_UP(end); end_pfn 480 arch/x86/xen/setup.c end_pfn = PFN_UP(entry->addr); end_pfn 482 arch/x86/xen/setup.c if (start_pfn < end_pfn) end_pfn 483 arch/x86/xen/setup.c ret_val = func(start_pfn, end_pfn, nr_pages, end_pfn 767 drivers/base/node.c unsigned long end_pfn = start_pfn + memory_block_pfns - 1; end_pfn 771 drivers/base/node.c for (pfn = start_pfn; pfn <= end_pfn; pfn++) { end_pfn 832 drivers/base/node.c int link_mem_sections(int nid, unsigned long start_pfn, unsigned long end_pfn) end_pfn 835 drivers/base/node.c PFN_PHYS(end_pfn - start_pfn), (void *)&nid, end_pfn 77 drivers/dax/super.c pfn_t pfn, end_pfn; end_pfn 105 drivers/dax/super.c len2 = dax_direct_access(dax_dev, pgoff_end, 1, &end_kaddr, &end_pfn); end_pfn 125 drivers/dax/super.c } else if (pfn_t_devmap(pfn) && pfn_t_devmap(end_pfn)) { end_pfn 129 drivers/dax/super.c end_pgmap = get_dev_pagemap(pfn_t_to_pfn(end_pfn), NULL); end_pfn 132 drivers/dax/super.c && pfn_t_to_page(end_pfn)->pgmap == pgmap end_pfn 134 drivers/dax/super.c && pfn_t_to_pfn(end_pfn) == PHYS_PFN(__pa(end_kaddr))) end_pfn 436 drivers/hv/hv_balloon.c unsigned long end_pfn; end_pfn 446 drivers/hv/hv_balloon.c unsigned long end_pfn; end_pfn 583 drivers/hv/hv_balloon.c if ((pfn >= gap->start_pfn) && (pfn < gap->end_pfn)) end_pfn 605 drivers/hv/hv_balloon.c (pfn < has->end_pfn) && end_pfn 779 drivers/hv/hv_balloon.c (pfn + (1UL << order) > has->end_pfn)) end_pfn 802 drivers/hv/hv_balloon.c if (start_pfn < has->start_pfn || start_pfn >= has->end_pfn) end_pfn 818 drivers/hv/hv_balloon.c gap->end_pfn = start_pfn; end_pfn 828 drivers/hv/hv_balloon.c if ((start_pfn + pfn_cnt) > has->end_pfn) { end_pfn 829 drivers/hv/hv_balloon.c residual = (start_pfn + pfn_cnt - has->end_pfn); end_pfn 837 drivers/hv/hv_balloon.c has->end_pfn += new_inc; end_pfn 868 drivers/hv/hv_balloon.c if (start_pfn < has->start_pfn || start_pfn >= has->end_pfn) end_pfn 899 drivers/hv/hv_balloon.c if ((has->ha_end_pfn < has->end_pfn) && (pfn_cnt > 0)) { end_pfn 907 drivers/hv/hv_balloon.c size = (has->end_pfn - has->ha_end_pfn); end_pfn 968 drivers/hv/hv_balloon.c ha_region->end_pfn = rg_start + rg_size; end_pfn 2266 drivers/iommu/intel-iommu.c unsigned long nr_superpages, end_pfn; end_pfn 2272 drivers/iommu/intel-iommu.c end_pfn = iov_pfn + nr_superpages * lvl_pages - 1; end_pfn 2280 drivers/iommu/intel-iommu.c dma_pte_free_pagetable(domain, iov_pfn, end_pfn, end_pfn 2753 drivers/iommu/intel-iommu.c unsigned long start_pfn, end_pfn; end_pfn 2756 drivers/iommu/intel-iommu.c for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { end_pfn 2758 drivers/iommu/intel-iommu.c PFN_PHYS(start_pfn), PFN_PHYS(end_pfn)); end_pfn 689 drivers/net/ethernet/ibm/ehea/ehea_qmr.c unsigned long pfn, start_pfn, end_pfn, nr_pages; end_pfn 697 drivers/net/ethernet/ibm/ehea/ehea_qmr.c end_pfn = initial_pfn + total_nr_pages; end_pfn 700 drivers/net/ethernet/ibm/ehea/ehea_qmr.c while (pfn < end_pfn) { end_pfn 679 drivers/nvdimm/pfn_devs.c .end_pfn = PHYS_PFN(end), end_pfn 239 fs/proc/kcore.c unsigned long end_pfn; end_pfn 243 fs/proc/kcore.c end_pfn = 0; end_pfn 247 fs/proc/kcore.c if (end_pfn < node_end) end_pfn 248 fs/proc/kcore.c end_pfn = node_end; end_pfn 251 fs/proc/kcore.c ret = walk_system_ram_range(0, end_pfn, list, kclist_add_private); end_pfn 485 include/linux/hugetlb.h unsigned long end_pfn); end_pfn 707 include/linux/hugetlb.h unsigned long end_pfn) end_pfn 221 include/linux/memblock.h unsigned long *end_pfn); end_pfn 98 include/linux/memory_hotplug.h extern int test_pages_in_a_zone(unsigned long start_pfn, unsigned long end_pfn, end_pfn 101 include/linux/memory_hotplug.h unsigned long end_pfn); end_pfn 20 include/linux/memremap.h const unsigned long end_pfn; end_pfn 2176 include/linux/mm.h unsigned long end_pfn); end_pfn 2178 include/linux/mm.h unsigned long end_pfn); end_pfn 2180 include/linux/mm.h unsigned long *start_pfn, unsigned long *end_pfn); end_pfn 1312 include/linux/mmzone.h void online_mem_sections(unsigned long start_pfn, unsigned long end_pfn); end_pfn 1314 include/linux/mmzone.h void offline_mem_sections(unsigned long start_pfn, unsigned long end_pfn); end_pfn 103 include/linux/node.h unsigned long end_pfn); end_pfn 106 include/linux/node.h unsigned long end_pfn) end_pfn 125 include/linux/node.h unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; end_pfn 131 include/linux/node.h error = link_mem_sections(nid, start_pfn, end_pfn); end_pfn 46 include/linux/page-isolation.h start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, end_pfn 54 include/linux/page-isolation.h undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, end_pfn 60 include/linux/page-isolation.h int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, end_pfn 18 include/trace/events/compaction.h unsigned long end_pfn, end_pfn 22 include/trace/events/compaction.h TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken), end_pfn 26 include/trace/events/compaction.h __field(unsigned long, end_pfn) end_pfn 33 include/trace/events/compaction.h __entry->end_pfn = end_pfn; end_pfn 40 include/trace/events/compaction.h __entry->end_pfn, end_pfn 49 include/trace/events/compaction.h unsigned long end_pfn, end_pfn 53 include/trace/events/compaction.h TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) end_pfn 60 include/trace/events/compaction.h unsigned long end_pfn, end_pfn 64 include/trace/events/compaction.h TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) end_pfn 14 include/trace/events/page_isolation.h unsigned long end_pfn, end_pfn 17 include/trace/events/page_isolation.h TP_ARGS(start_pfn, end_pfn, fin_pfn), end_pfn 21 include/trace/events/page_isolation.h __field(unsigned long, end_pfn) end_pfn 27 include/trace/events/page_isolation.h __entry->end_pfn = end_pfn; end_pfn 32 include/trace/events/page_isolation.h __entry->start_pfn, __entry->end_pfn, __entry->fin_pfn, end_pfn 33 include/trace/events/page_isolation.h __entry->end_pfn <= __entry->fin_pfn ? "success" : "fail") end_pfn 362 kernel/power/snapshot.c unsigned long end_pfn; /* Zone end page frame + 1 */ end_pfn 516 kernel/power/snapshot.c zone->end_pfn = end; end_pfn 716 kernel/power/snapshot.c if (pfn >= zone->start_pfn && pfn < zone->end_pfn) end_pfn 723 kernel/power/snapshot.c if (pfn >= curr->start_pfn && pfn < curr->end_pfn) { end_pfn 888 kernel/power/snapshot.c pages = bm->cur.zone->end_pfn - bm->cur.zone->start_pfn; end_pfn 909 kernel/power/snapshot.c unsigned long end_pfn; end_pfn 949 kernel/power/snapshot.c unsigned long end_pfn, int use_kmalloc) end_pfn 953 kernel/power/snapshot.c if (start_pfn >= end_pfn) end_pfn 960 kernel/power/snapshot.c if (region->end_pfn == start_pfn) { end_pfn 961 kernel/power/snapshot.c region->end_pfn = end_pfn; end_pfn 978 kernel/power/snapshot.c region->end_pfn = end_pfn; end_pfn 983 kernel/power/snapshot.c ((unsigned long long) end_pfn << PAGE_SHIFT) - 1); end_pfn 1055 kernel/power/snapshot.c ((unsigned long long) region->end_pfn << PAGE_SHIFT) end_pfn 1058 kernel/power/snapshot.c for (pfn = region->start_pfn; pfn < region->end_pfn; pfn++) end_pfn 481 kernel/resource.c unsigned long pfn, end_pfn; end_pfn 491 kernel/resource.c end_pfn = PFN_DOWN(res.end + 1); end_pfn 492 kernel/resource.c if (end_pfn > pfn) end_pfn 493 kernel/resource.c ret = (*func)(pfn, end_pfn - pfn, arg); end_pfn 537 mm/compaction.c unsigned long end_pfn, end_pfn 556 mm/compaction.c for (; blockpfn < end_pfn; blockpfn += stride, cursor += stride) { end_pfn 644 mm/compaction.c if (unlikely(blockpfn > end_pfn)) end_pfn 645 mm/compaction.c blockpfn = end_pfn; end_pfn 658 mm/compaction.c if (strict && blockpfn < end_pfn) end_pfn 683 mm/compaction.c unsigned long start_pfn, unsigned long end_pfn) end_pfn 694 mm/compaction.c for (; pfn < end_pfn; pfn += isolated, end_pfn 700 mm/compaction.c block_end_pfn = min(block_end_pfn, end_pfn); end_pfn 710 mm/compaction.c block_end_pfn = min(block_end_pfn, end_pfn); end_pfn 738 mm/compaction.c if (pfn < end_pfn) { end_pfn 783 mm/compaction.c unsigned long end_pfn, isolate_mode_t isolate_mode) end_pfn 820 mm/compaction.c for (; low_pfn < end_pfn; low_pfn++) { end_pfn 870 mm/compaction.c low_pfn = end_pfn; end_pfn 1042 mm/compaction.c if (unlikely(low_pfn > end_pfn)) end_pfn 1043 mm/compaction.c low_pfn = end_pfn; end_pfn 1057 mm/compaction.c if (low_pfn == end_pfn && (!nr_isolated || cc->rescan)) { end_pfn 1086 mm/compaction.c unsigned long end_pfn) end_pfn 1097 mm/compaction.c for (; pfn < end_pfn; pfn = block_end_pfn, end_pfn 1101 mm/compaction.c block_end_pfn = min(block_end_pfn, end_pfn); end_pfn 1224 mm/compaction.c unsigned long start_pfn, end_pfn; end_pfn 1237 mm/compaction.c end_pfn = min(pageblock_end_pfn(pfn), zone_end_pfn(cc->zone)) - 1; end_pfn 1248 mm/compaction.c if (start_pfn < end_pfn) end_pfn 1249 mm/compaction.c isolate_freepages_block(cc, &start_pfn, end_pfn, &cc->freepages, 1, false); end_pfn 2076 mm/compaction.c unsigned long end_pfn = zone_end_pfn(cc->zone); end_pfn 2118 mm/compaction.c cc->free_pfn = pageblock_start_pfn(end_pfn - 1); end_pfn 2122 mm/compaction.c if (cc->free_pfn < start_pfn || cc->free_pfn >= end_pfn) { end_pfn 2123 mm/compaction.c cc->free_pfn = pageblock_start_pfn(end_pfn - 1); end_pfn 2126 mm/compaction.c if (cc->migrate_pfn < start_pfn || cc->migrate_pfn >= end_pfn) { end_pfn 2150 mm/compaction.c cc->free_pfn, end_pfn, sync); end_pfn 2283 mm/compaction.c cc->free_pfn, end_pfn, sync, ret); end_pfn 1076 mm/hugetlb.c unsigned long end_pfn = start_pfn + nr_pages; end_pfn 1077 mm/hugetlb.c return alloc_contig_range(start_pfn, end_pfn, MIGRATE_MOVABLE, end_pfn 1084 mm/hugetlb.c unsigned long i, end_pfn = start_pfn + nr_pages; end_pfn 1087 mm/hugetlb.c for (i = start_pfn; i < end_pfn; i++) { end_pfn 1649 mm/hugetlb.c int dissolve_free_huge_pages(unsigned long start_pfn, unsigned long end_pfn) end_pfn 1658 mm/hugetlb.c for (pfn = start_pfn; pfn < end_pfn; pfn += 1 << minimum_order) { end_pfn 148 mm/internal.h unsigned long end_pfn, struct zone *zone); end_pfn 151 mm/internal.h unsigned long end_pfn, struct zone *zone) end_pfn 156 mm/internal.h return __pageblock_pfn_to_page(start_pfn, end_pfn, zone); end_pfn 219 mm/internal.h unsigned long start_pfn, unsigned long end_pfn); end_pfn 222 mm/internal.h unsigned long low_pfn, unsigned long end_pfn); end_pfn 463 mm/internal.h unsigned long *end_pfn); end_pfn 466 mm/internal.h unsigned long *end_pfn) end_pfn 1444 mm/kmemleak.c unsigned long end_pfn = node_end_pfn(i); end_pfn 1447 mm/kmemleak.c for (pfn = start_pfn; pfn < end_pfn; pfn++) { end_pfn 2725 mm/ksm.c unsigned long end_pfn) end_pfn 2728 mm/ksm.c stable_node->kpfn < end_pfn) { end_pfn 2741 mm/ksm.c unsigned long end_pfn, end_pfn 2750 mm/ksm.c end_pfn); end_pfn 2756 mm/ksm.c stable_node_dup_remove_range(dup, start_pfn, end_pfn); end_pfn 2766 mm/ksm.c unsigned long end_pfn) end_pfn 2777 mm/ksm.c start_pfn, end_pfn, end_pfn 2788 mm/ksm.c stable_node->kpfn < end_pfn) end_pfn 1601 mm/memblock.c unsigned long start_pfn, end_pfn; end_pfn 1605 mm/memblock.c end_pfn = memblock_region_memory_end_pfn(r); end_pfn 1607 mm/memblock.c end_pfn = min_t(unsigned long, end_pfn, limit_pfn); end_pfn 1608 mm/memblock.c pages += end_pfn - start_pfn; end_pfn 1751 mm/memblock.c unsigned long *start_pfn, unsigned long *end_pfn) end_pfn 1760 mm/memblock.c *end_pfn = PFN_DOWN(type->regions[mid].base + type->regions[mid].size); end_pfn 1912 mm/memblock.c unsigned long end_pfn = min_t(unsigned long, end_pfn 1915 mm/memblock.c if (start_pfn >= end_pfn) end_pfn 1918 mm/memblock.c __free_pages_memory(start_pfn, end_pfn); end_pfn 1920 mm/memblock.c return end_pfn - start_pfn; end_pfn 227 mm/memory_hotplug.c unsigned long i, pfn, end_pfn, nr_pages; end_pfn 238 mm/memory_hotplug.c end_pfn = pgdat_end_pfn(pgdat); end_pfn 241 mm/memory_hotplug.c for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { end_pfn 331 mm/memory_hotplug.c unsigned long end_pfn) end_pfn 333 mm/memory_hotplug.c for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) { end_pfn 352 mm/memory_hotplug.c unsigned long end_pfn) end_pfn 357 mm/memory_hotplug.c pfn = end_pfn - 1; end_pfn 375 mm/memory_hotplug.c unsigned long end_pfn) end_pfn 391 mm/memory_hotplug.c pfn = find_smallest_section_pfn(nid, zone, end_pfn, end_pfn 397 mm/memory_hotplug.c } else if (zone_end_pfn == end_pfn) { end_pfn 425 mm/memory_hotplug.c if (pfn >= start_pfn && pfn < end_pfn) end_pfn 619 mm/memory_hotplug.c const unsigned long end_pfn = start_pfn + nr_pages; end_pfn 628 mm/memory_hotplug.c for (pfn = start_pfn; pfn < end_pfn; pfn += 1ul << order) { end_pfn 629 mm/memory_hotplug.c order = min(MAX_ORDER - 1, get_order(PFN_PHYS(end_pfn - pfn))); end_pfn 637 mm/memory_hotplug.c online_mem_sections(start_pfn, end_pfn); end_pfn 1195 mm/memory_hotplug.c unsigned long end_pfn, pfn; end_pfn 1197 mm/memory_hotplug.c end_pfn = min(start_pfn + nr_pages, end_pfn 1201 mm/memory_hotplug.c for (pfn = start_pfn; pfn < end_pfn; pfn = next_active_pageblock(pfn)) { end_pfn 1215 mm/memory_hotplug.c int test_pages_in_a_zone(unsigned long start_pfn, unsigned long end_pfn, end_pfn 1224 mm/memory_hotplug.c pfn < end_pfn; end_pfn 1229 mm/memory_hotplug.c for (; pfn < sec_end_pfn && pfn < end_pfn; end_pfn 1236 mm/memory_hotplug.c if (i == MAX_ORDER_NR_PAGES || pfn + i >= end_pfn) end_pfn 1253 mm/memory_hotplug.c *valid_end = min(end, end_pfn); end_pfn 1311 mm/memory_hotplug.c do_migrate_range(unsigned long start_pfn, unsigned long end_pfn) end_pfn 1318 mm/memory_hotplug.c for (pfn = start_pfn; pfn < end_pfn; pfn++) { end_pfn 1488 mm/memory_hotplug.c unsigned long end_pfn) end_pfn 1503 mm/memory_hotplug.c if (!test_pages_in_a_zone(start_pfn, end_pfn, &valid_start, end_pfn 1512 mm/memory_hotplug.c nr_pages = end_pfn - start_pfn; end_pfn 1515 mm/memory_hotplug.c ret = start_isolate_page_range(start_pfn, end_pfn, end_pfn 1546 mm/memory_hotplug.c pfn = scan_movable_pages(pfn, end_pfn); end_pfn 1552 mm/memory_hotplug.c do_migrate_range(pfn, end_pfn); end_pfn 1561 mm/memory_hotplug.c ret = dissolve_free_huge_pages(start_pfn, end_pfn); end_pfn 1567 mm/memory_hotplug.c ret = walk_system_ram_range(start_pfn, end_pfn - start_pfn, end_pfn 1573 mm/memory_hotplug.c walk_system_ram_range(start_pfn, end_pfn - start_pfn, end_pfn 1616 mm/memory_hotplug.c undo_isolate_page_range(start_pfn, end_pfn, MIGRATE_MOVABLE); end_pfn 1621 mm/memory_hotplug.c ((unsigned long long) end_pfn << PAGE_SHIFT) - 1, end_pfn 405 mm/page_alloc.c defer_init(int nid, unsigned long pfn, unsigned long end_pfn) end_pfn 413 mm/page_alloc.c if (prev_end_pfn != end_pfn) { end_pfn 414 mm/page_alloc.c prev_end_pfn = end_pfn; end_pfn 419 mm/page_alloc.c if (end_pfn < pgdat_end_pfn(NODE_DATA(nid))) end_pfn 442 mm/page_alloc.c static inline bool defer_init(int nid, unsigned long pfn, unsigned long end_pfn) end_pfn 1393 mm/page_alloc.c unsigned long end_pfn = PFN_UP(end); end_pfn 1395 mm/page_alloc.c for (; start_pfn < end_pfn; start_pfn++) { end_pfn 1516 mm/page_alloc.c unsigned long end_pfn, struct zone *zone) end_pfn 1522 mm/page_alloc.c end_pfn--; end_pfn 1524 mm/page_alloc.c if (!pfn_valid(start_pfn) || !pfn_valid(end_pfn)) end_pfn 1534 mm/page_alloc.c end_page = pfn_to_page(end_pfn); end_pfn 1631 mm/page_alloc.c unsigned long end_pfn) end_pfn 1636 mm/page_alloc.c for (; pfn < end_pfn; pfn++) { end_pfn 1659 mm/page_alloc.c unsigned long end_pfn) end_pfn 1667 mm/page_alloc.c for (; pfn < end_pfn; pfn++) { end_pfn 1725 mm/page_alloc.c unsigned long *end_pfn) end_pfn 1728 mm/page_alloc.c unsigned long spfn = *start_pfn, epfn = *end_pfn; end_pfn 1733 mm/page_alloc.c for_each_free_mem_pfn_range_in_zone_from(j, zone, start_pfn, end_pfn) { end_pfn 1739 mm/page_alloc.c t = min(mo_pfn, *end_pfn); end_pfn 1742 mm/page_alloc.c if (mo_pfn < *end_pfn) { end_pfn 2282 mm/page_alloc.c unsigned long start_pfn, end_pfn; end_pfn 2292 mm/page_alloc.c end_pfn = start_pfn + pageblock_nr_pages - 1; end_pfn 2297 mm/page_alloc.c if (!zone_spans_pfn(zone, end_pfn)) end_pfn 5881 mm/page_alloc.c unsigned long pfn, end_pfn = start_pfn + size; end_pfn 5884 mm/page_alloc.c if (highest_memmap_pfn < end_pfn - 1) end_pfn 5885 mm/page_alloc.c highest_memmap_pfn = end_pfn - 1; end_pfn 5901 mm/page_alloc.c end_pfn = altmap->base_pfn + vmem_altmap_offset(altmap); end_pfn 5905 mm/page_alloc.c for (pfn = start_pfn; pfn < end_pfn; pfn++) { end_pfn 5917 mm/page_alloc.c if (defer_init(nid, pfn, end_pfn)) end_pfn 5951 mm/page_alloc.c unsigned long pfn, end_pfn = start_pfn + size; end_pfn 5968 mm/page_alloc.c size = end_pfn - start_pfn; end_pfn 5971 mm/page_alloc.c for (pfn = start_pfn; pfn < end_pfn; pfn++) { end_pfn 6238 mm/page_alloc.c unsigned long start_pfn, end_pfn; end_pfn 6244 mm/page_alloc.c nid = memblock_search_pfn_nid(pfn, &start_pfn, &end_pfn); end_pfn 6247 mm/page_alloc.c state->last_end = end_pfn; end_pfn 6266 mm/page_alloc.c unsigned long start_pfn, end_pfn; end_pfn 6269 mm/page_alloc.c for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, &this_nid) { end_pfn 6271 mm/page_alloc.c end_pfn = min(end_pfn, max_low_pfn); end_pfn 6273 mm/page_alloc.c if (start_pfn < end_pfn) end_pfn 6275 mm/page_alloc.c (end_pfn - start_pfn) << PAGE_SHIFT, end_pfn 6289 mm/page_alloc.c unsigned long start_pfn, end_pfn; end_pfn 6292 mm/page_alloc.c for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, &this_nid) end_pfn 6293 mm/page_alloc.c memory_present(this_nid, start_pfn, end_pfn); end_pfn 6308 mm/page_alloc.c unsigned long *start_pfn, unsigned long *end_pfn) end_pfn 6314 mm/page_alloc.c *end_pfn = 0; end_pfn 6318 mm/page_alloc.c *end_pfn = max(*end_pfn, this_end_pfn); end_pfn 6429 mm/page_alloc.c unsigned long start_pfn, end_pfn; end_pfn 6432 mm/page_alloc.c for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { end_pfn 6434 mm/page_alloc.c end_pfn = clamp(end_pfn, range_start_pfn, range_end_pfn); end_pfn 6435 mm/page_alloc.c nr_absent -= end_pfn - start_pfn; end_pfn 6448 mm/page_alloc.c unsigned long end_pfn) end_pfn 6450 mm/page_alloc.c return __absent_pages_in_range(MAX_NUMNODES, start_pfn, end_pfn); end_pfn 6483 mm/page_alloc.c unsigned long start_pfn, end_pfn; end_pfn 6489 mm/page_alloc.c end_pfn = clamp(memblock_region_memory_end_pfn(r), end_pfn 6494 mm/page_alloc.c nr_absent += end_pfn - start_pfn; end_pfn 6498 mm/page_alloc.c nr_absent += end_pfn - start_pfn; end_pfn 6887 mm/page_alloc.c unsigned long end_pfn = 0; end_pfn 6896 mm/page_alloc.c get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); end_pfn 6899 mm/page_alloc.c end_pfn ? ((u64)end_pfn << PAGE_SHIFT) - 1 : 0); end_pfn 6903 mm/page_alloc.c calculate_node_totalpages(pgdat, start_pfn, end_pfn, end_pfn 7086 mm/page_alloc.c unsigned long start_pfn, end_pfn; end_pfn 7089 mm/page_alloc.c for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { end_pfn 7090 mm/page_alloc.c unsigned long pages = end_pfn - start_pfn; end_pfn 7217 mm/page_alloc.c unsigned long start_pfn, end_pfn; end_pfn 7235 mm/page_alloc.c for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { end_pfn 7239 mm/page_alloc.c if (start_pfn >= end_pfn) end_pfn 7245 mm/page_alloc.c kernel_pages = min(end_pfn, usable_startpfn) end_pfn 7254 mm/page_alloc.c if (end_pfn <= usable_startpfn) { end_pfn 7262 mm/page_alloc.c zone_movable_pfn[nid] = end_pfn; end_pfn 7273 mm/page_alloc.c size_pages = end_pfn - start_pfn; end_pfn 7344 mm/page_alloc.c unsigned long start_pfn, end_pfn; end_pfn 7359 mm/page_alloc.c end_pfn = max(max_zone_pfn[i], start_pfn); end_pfn 7361 mm/page_alloc.c arch_zone_highest_possible_pfn[i] = end_pfn; end_pfn 7363 mm/page_alloc.c start_pfn = end_pfn; end_pfn 7401 mm/page_alloc.c for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { end_pfn 7404 mm/page_alloc.c ((u64)end_pfn << PAGE_SHIFT) - 1); end_pfn 7405 mm/page_alloc.c subsection_map_init(start_pfn, end_pfn - start_pfn); end_pfn 8567 mm/page_alloc.c __offline_isolated_pages(unsigned long start_pfn, unsigned long end_pfn) end_pfn 8577 mm/page_alloc.c for (pfn = start_pfn; pfn < end_pfn; pfn++) end_pfn 8580 mm/page_alloc.c if (pfn == end_pfn) end_pfn 8583 mm/page_alloc.c offline_mem_sections(pfn, end_pfn); end_pfn 8587 mm/page_alloc.c while (pfn < end_pfn) { end_pfn 8610 mm/page_alloc.c pfn, 1 << order, end_pfn); end_pfn 375 mm/page_ext.c unsigned long start_pfn, end_pfn; end_pfn 378 mm/page_ext.c end_pfn = node_end_pfn(nid); end_pfn 384 mm/page_ext.c for (pfn = start_pfn; pfn < end_pfn; end_pfn 127 mm/page_idle.c unsigned long pfn, end_pfn; end_pfn 137 mm/page_idle.c end_pfn = pfn + count * BITS_PER_BYTE; end_pfn 138 mm/page_idle.c if (end_pfn > max_pfn) end_pfn 139 mm/page_idle.c end_pfn = max_pfn; end_pfn 141 mm/page_idle.c for (; pfn < end_pfn; pfn++) { end_pfn 172 mm/page_idle.c unsigned long pfn, end_pfn; end_pfn 182 mm/page_idle.c end_pfn = pfn + count * BITS_PER_BYTE; end_pfn 183 mm/page_idle.c if (end_pfn > max_pfn) end_pfn 184 mm/page_idle.c end_pfn = max_pfn; end_pfn 186 mm/page_idle.c for (; pfn < end_pfn; pfn++) { end_pfn 193 mm/page_isolation.c int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, end_pfn 202 mm/page_isolation.c BUG_ON(!IS_ALIGNED(end_pfn, pageblock_nr_pages)); end_pfn 205 mm/page_isolation.c pfn < end_pfn; end_pfn 233 mm/page_isolation.c void undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, end_pfn 240 mm/page_isolation.c BUG_ON(!IS_ALIGNED(end_pfn, pageblock_nr_pages)); end_pfn 243 mm/page_isolation.c pfn < end_pfn; end_pfn 259 mm/page_isolation.c __test_page_isolated_in_pageblock(unsigned long pfn, unsigned long end_pfn, end_pfn 264 mm/page_isolation.c while (pfn < end_pfn) { end_pfn 288 mm/page_isolation.c int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, end_pfn 300 mm/page_isolation.c for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { end_pfn 305 mm/page_isolation.c page = __first_valid_page(start_pfn, end_pfn - start_pfn); end_pfn 306 mm/page_isolation.c if ((pfn < end_pfn) || !page) end_pfn 311 mm/page_isolation.c pfn = __test_page_isolated_in_pageblock(start_pfn, end_pfn, end_pfn 315 mm/page_isolation.c trace_test_pages_isolated(start_pfn, end_pfn, pfn); end_pfn 317 mm/page_isolation.c return pfn < end_pfn ? -EBUSY : 0; end_pfn 260 mm/page_owner.c unsigned long end_pfn = pfn + zone->spanned_pages; end_pfn 273 mm/page_owner.c for (; pfn < end_pfn; ) { end_pfn 281 mm/page_owner.c block_end_pfn = min(block_end_pfn, end_pfn); end_pfn 549 mm/page_owner.c unsigned long end_pfn = zone_end_pfn(zone); end_pfn 557 mm/page_owner.c for (; pfn < end_pfn; ) { end_pfn 566 mm/page_owner.c block_end_pfn = min(block_end_pfn, end_pfn); end_pfn 109 mm/shuffle.c unsigned long end_pfn = zone_end_pfn(z); end_pfn 115 mm/shuffle.c for (i = start_pfn; i < end_pfn; i += order_pages) { end_pfn 157 mm/sparse.c unsigned long *end_pfn) end_pfn 168 mm/sparse.c *start_pfn, *end_pfn, max_sparsemem_pfn); end_pfn 171 mm/sparse.c *end_pfn = max_sparsemem_pfn; end_pfn 172 mm/sparse.c } else if (*end_pfn > max_sparsemem_pfn) { end_pfn 175 mm/sparse.c *start_pfn, *end_pfn, max_sparsemem_pfn); end_pfn 177 mm/sparse.c *end_pfn = max_sparsemem_pfn; end_pfn 609 mm/sparse.c void online_mem_sections(unsigned long start_pfn, unsigned long end_pfn) end_pfn 613 mm/sparse.c for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { end_pfn 628 mm/sparse.c void offline_mem_sections(unsigned long start_pfn, unsigned long end_pfn) end_pfn 632 mm/sparse.c for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { end_pfn 1437 mm/vmstat.c unsigned long end_pfn = zone_end_pfn(zone); end_pfn 1440 mm/vmstat.c for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) {