end_pfn           220 arch/alpha/mm/numa.c 		unsigned long end_pfn = start_pfn + NODE_DATA(nid)->node_present_pages;
end_pfn           222 arch/alpha/mm/numa.c 		if (dma_local_pfn >= end_pfn - start_pfn)
end_pfn           223 arch/alpha/mm/numa.c 			zones_size[ZONE_DMA] = end_pfn - start_pfn;
end_pfn           226 arch/alpha/mm/numa.c 			zones_size[ZONE_NORMAL] = (end_pfn - start_pfn) - dma_local_pfn;
end_pfn           327 arch/arm/mm/init.c free_memmap(unsigned long start_pfn, unsigned long end_pfn)
end_pfn           336 arch/arm/mm/init.c 	end_pg = pfn_to_page(end_pfn - 1) + 1;
end_pfn           462 arch/arm64/mm/init.c static inline void free_memmap(unsigned long start_pfn, unsigned long end_pfn)
end_pfn           471 arch/arm64/mm/init.c 	end_pg = pfn_to_page(end_pfn - 1) + 1;
end_pfn           218 arch/arm64/mm/numa.c static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn)
end_pfn           225 arch/arm64/mm/numa.c 	if (start_pfn >= end_pfn)
end_pfn           246 arch/arm64/mm/numa.c 	NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn;
end_pfn           363 arch/arm64/mm/numa.c 		unsigned long start_pfn, end_pfn;
end_pfn           365 arch/arm64/mm/numa.c 		get_pfn_range_for_nid(nid, &start_pfn, &end_pfn);
end_pfn           366 arch/arm64/mm/numa.c 		setup_node_data(nid, start_pfn, end_pfn);
end_pfn           174 arch/microblaze/mm/init.c 		unsigned long start_pfn, end_pfn;
end_pfn           177 arch/microblaze/mm/init.c 		end_pfn = memblock_region_memory_end_pfn(reg);
end_pfn           179 arch/microblaze/mm/init.c 				  (end_pfn - start_pfn) << PAGE_SHIFT,
end_pfn           123 arch/mips/loongson64/loongson-3/numa.c 	u64 node_id, node_psize, start_pfn, end_pfn, mem_start, mem_size;
end_pfn           139 arch/mips/loongson64/loongson-3/numa.c 			end_pfn  = start_pfn + node_psize;
end_pfn           144 arch/mips/loongson64/loongson-3/numa.c 				start_pfn, end_pfn, num_physpages);
end_pfn           146 arch/mips/loongson64/loongson-3/numa.c 				PFN_PHYS(end_pfn - start_pfn), node);
end_pfn           151 arch/mips/loongson64/loongson-3/numa.c 			end_pfn  = start_pfn + node_psize;
end_pfn           156 arch/mips/loongson64/loongson-3/numa.c 				start_pfn, end_pfn, num_physpages);
end_pfn           158 arch/mips/loongson64/loongson-3/numa.c 				PFN_PHYS(end_pfn - start_pfn), node);
end_pfn           173 arch/mips/loongson64/loongson-3/numa.c 	unsigned long start_pfn, end_pfn;
end_pfn           179 arch/mips/loongson64/loongson-3/numa.c 	get_pfn_range_for_nid(node, &start_pfn, &end_pfn);
end_pfn           181 arch/mips/loongson64/loongson-3/numa.c 		node, start_pfn, end_pfn);
end_pfn           186 arch/mips/loongson64/loongson-3/numa.c 	NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn;
end_pfn           193 arch/mips/loongson64/loongson-3/numa.c 		max_low_pfn = end_pfn;
end_pfn           387 arch/mips/sgi-ip27/ip27-memory.c 	unsigned long start_pfn, end_pfn;
end_pfn           389 arch/mips/sgi-ip27/ip27-memory.c 	get_pfn_range_for_nid(node, &start_pfn, &end_pfn);
end_pfn           398 arch/mips/sgi-ip27/ip27-memory.c 	NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn;
end_pfn           405 arch/mips/sgi-ip27/ip27-memory.c 	free_bootmem_with_active_regions(node, end_pfn);
end_pfn           229 arch/parisc/mm/init.c 		unsigned long end_pfn;
end_pfn           233 arch/parisc/mm/init.c 		end_pfn = pmem_ranges[0].start_pfn + pmem_ranges[0].pages;
end_pfn           236 arch/parisc/mm/init.c 			hole_pages = pmem_ranges[i].start_pfn - end_pfn;
end_pfn           238 arch/parisc/mm/init.c 				pmem_holes[npmem_holes].start_pfn = end_pfn;
end_pfn           240 arch/parisc/mm/init.c 				end_pfn += hole_pages;
end_pfn           242 arch/parisc/mm/init.c 			end_pfn += pmem_ranges[i].pages;
end_pfn           245 arch/parisc/mm/init.c 		pmem_ranges[0].pages = end_pfn - pmem_ranges[0].start_pfn;
end_pfn          1112 arch/powerpc/kernel/fadump.c 					unsigned long end_pfn)
end_pfn          1118 arch/powerpc/kernel/fadump.c 		PFN_PHYS(start_pfn), PFN_PHYS(end_pfn));
end_pfn          1120 arch/powerpc/kernel/fadump.c 	for (pfn = start_pfn; pfn < end_pfn; pfn++) {
end_pfn           181 arch/powerpc/mm/init_64.c 	if ((start_pfn + nr_pfn) > altmap->end_pfn)
end_pfn            85 arch/powerpc/mm/numa.c static int __init fake_numa_create_new_node(unsigned long end_pfn,
end_pfn           116 arch/powerpc/mm/numa.c 	if ((end_pfn << PAGE_SHIFT) > mem) {
end_pfn           744 arch/powerpc/mm/numa.c 	unsigned long start_pfn, end_pfn;
end_pfn           755 arch/powerpc/mm/numa.c 		end_pfn = memblock_region_memory_end_pfn(reg);
end_pfn           757 arch/powerpc/mm/numa.c 		fake_numa_create_new_node(end_pfn, &nid);
end_pfn           759 arch/powerpc/mm/numa.c 				  PFN_PHYS(end_pfn - start_pfn),
end_pfn           801 arch/powerpc/mm/numa.c static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn)
end_pfn           803 arch/powerpc/mm/numa.c 	u64 spanned_pages = end_pfn - start_pfn;
end_pfn           891 arch/powerpc/mm/numa.c 		unsigned long start_pfn, end_pfn;
end_pfn           893 arch/powerpc/mm/numa.c 		get_pfn_range_for_nid(nid, &start_pfn, &end_pfn);
end_pfn           894 arch/powerpc/mm/numa.c 		setup_node_data(nid, start_pfn, end_pfn);
end_pfn            97 arch/powerpc/platforms/powernv/memtrace.c 	u64 start_pfn, end_pfn, nr_pages, pfn;
end_pfn           105 arch/powerpc/platforms/powernv/memtrace.c 	end_pfn = node_end_pfn(nid);
end_pfn           109 arch/powerpc/platforms/powernv/memtrace.c 	end_pfn = round_down(end_pfn - nr_pages, nr_pages);
end_pfn           112 arch/powerpc/platforms/powernv/memtrace.c 	for (base_pfn = end_pfn; base_pfn > start_pfn; base_pfn -= nr_pages) {
end_pfn           120 arch/powerpc/platforms/powernv/memtrace.c 			end_pfn = base_pfn + nr_pages;
end_pfn           121 arch/powerpc/platforms/powernv/memtrace.c 			for (pfn = base_pfn; pfn < end_pfn; pfn += bytes>> PAGE_SHIFT) {
end_pfn           138 arch/riscv/mm/init.c 		unsigned long end_pfn = memblock_region_memory_end_pfn(reg);
end_pfn           141 arch/riscv/mm/init.c 				  PFN_PHYS(end_pfn - start_pfn),
end_pfn            56 arch/s390/kernel/early.c 	unsigned long end_pfn, init_pfn;
end_pfn            58 arch/s390/kernel/early.c 	end_pfn = PFN_UP(__pa(_end));
end_pfn            60 arch/s390/kernel/early.c 	for (init_pfn = 0 ; init_pfn < end_pfn; init_pfn++)
end_pfn           103 arch/s390/numa/numa.c 		unsigned long start_pfn, end_pfn;
end_pfn           108 arch/s390/numa/numa.c 		end_pfn = 0;
end_pfn           112 arch/s390/numa/numa.c 			if (t_end > end_pfn)
end_pfn           113 arch/s390/numa/numa.c 				end_pfn = t_end;
end_pfn           115 arch/s390/numa/numa.c 		NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn;
end_pfn            43 arch/sh/include/asm/mmzone.h 			       unsigned long end_pfn);
end_pfn           197 arch/sh/kernel/setup.c 						unsigned long end_pfn)
end_pfn           205 arch/sh/kernel/setup.c 	end = end_pfn << PAGE_SHIFT;
end_pfn           214 arch/sh/kernel/setup.c 		       start_pfn, end_pfn);
end_pfn           238 arch/sh/kernel/setup.c 	memblock_set_node(PFN_PHYS(start_pfn), PFN_PHYS(end_pfn - start_pfn),
end_pfn            22 arch/sh/kernel/swsusp.c 	unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT;
end_pfn            24 arch/sh/kernel/swsusp.c 	return (pfn >= begin_pfn) && (pfn < end_pfn);
end_pfn           201 arch/sh/mm/init.c 	unsigned long start_pfn, end_pfn;
end_pfn           203 arch/sh/mm/init.c 	get_pfn_range_for_nid(nid, &start_pfn, &end_pfn);
end_pfn           215 arch/sh/mm/init.c 	NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn;
end_pfn           224 arch/sh/mm/init.c 		unsigned long start_pfn, end_pfn;
end_pfn           226 arch/sh/mm/init.c 		end_pfn = memblock_region_memory_end_pfn(reg);
end_pfn           227 arch/sh/mm/init.c 		__add_active_range(0, start_pfn, end_pfn);
end_pfn            28 arch/sh/mm/numa.c 	unsigned long start_pfn, end_pfn;
end_pfn            34 arch/sh/mm/numa.c 	end_pfn = PFN_DOWN(end);
end_pfn            41 arch/sh/mm/numa.c 	__add_active_range(nid, start_pfn, end_pfn);
end_pfn            52 arch/sh/mm/numa.c 	NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn;
end_pfn            67 arch/sparc/mm/init_32.c 		unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT;
end_pfn            69 arch/sparc/mm/init_32.c 		if (end_pfn <= max_low_pfn)
end_pfn            75 arch/sparc/mm/init_32.c 		nr += end_pfn - start_pfn;
end_pfn           233 arch/sparc/mm/init_32.c static void map_high_region(unsigned long start_pfn, unsigned long end_pfn)
end_pfn           238 arch/sparc/mm/init_32.c 	printk("mapping high region %08lx - %08lx\n", start_pfn, end_pfn);
end_pfn           241 arch/sparc/mm/init_32.c 	for (tmp = start_pfn; tmp < end_pfn; tmp++)
end_pfn           282 arch/sparc/mm/init_32.c 		unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT;
end_pfn           284 arch/sparc/mm/init_32.c 		if (end_pfn <= highstart_pfn)
end_pfn           290 arch/sparc/mm/init_32.c 		map_high_region(start_pfn, end_pfn);
end_pfn          1090 arch/sparc/mm/init_64.c 	unsigned long start_pfn, end_pfn;
end_pfn          1105 arch/sparc/mm/init_64.c 	get_pfn_range_for_nid(nid, &start_pfn, &end_pfn);
end_pfn          1107 arch/sparc/mm/init_64.c 	p->node_spanned_pages = end_pfn - start_pfn;
end_pfn          1628 arch/sparc/mm/init_64.c 	unsigned long end_pfn;
end_pfn          1630 arch/sparc/mm/init_64.c 	end_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT;
end_pfn          1631 arch/sparc/mm/init_64.c 	max_pfn = max_low_pfn = end_pfn;
end_pfn          1645 arch/sparc/mm/init_64.c 	return end_pfn;
end_pfn          2291 arch/sparc/mm/init_64.c 	unsigned long end_pfn, shift, phys_base;
end_pfn          2460 arch/sparc/mm/init_64.c 	last_valid_pfn = end_pfn = bootmem_init(phys_base);
end_pfn          2469 arch/sparc/mm/init_64.c 		max_zone_pfns[ZONE_NORMAL] = end_pfn;
end_pfn           145 arch/unicore32/kernel/hibernate.c 	unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT;
end_pfn           147 arch/unicore32/kernel/hibernate.c 	return (pfn >= begin_pfn) && (pfn < end_pfn);
end_pfn           200 arch/unicore32/mm/init.c free_memmap(unsigned long start_pfn, unsigned long end_pfn)
end_pfn           209 arch/unicore32/mm/init.c 	end_pg = pfn_to_page(end_pfn);
end_pfn            76 arch/x86/include/asm/highmem.h 					unsigned long end_pfn);
end_pfn            50 arch/x86/include/asm/mtrr.h extern int mtrr_trim_uncached_memory(unsigned long end_pfn);
end_pfn            80 arch/x86/include/asm/mtrr.h static inline int mtrr_trim_uncached_memory(unsigned long end_pfn)
end_pfn            74 arch/x86/include/asm/page_types.h bool pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn);
end_pfn           715 arch/x86/kernel/amd_gart_64.c 	unsigned long start_pfn, end_pfn;
end_pfn           745 arch/x86/kernel/amd_gart_64.c 	end_pfn		= (aper_base>>PAGE_SHIFT) + (aper_size>>PAGE_SHIFT);
end_pfn           748 arch/x86/kernel/amd_gart_64.c 	if (!pfn_range_is_mapped(start_pfn, end_pfn))
end_pfn           749 arch/x86/kernel/amd_gart_64.c 		init_memory_mapping(start_pfn<<PAGE_SHIFT, end_pfn<<PAGE_SHIFT);
end_pfn           877 arch/x86/kernel/cpu/mtrr/cleanup.c int __init mtrr_trim_uncached_memory(unsigned long end_pfn)
end_pfn           969 arch/x86/kernel/cpu/mtrr/cleanup.c 	if (range[i].end < end_pfn)
end_pfn           971 arch/x86/kernel/cpu/mtrr/cleanup.c 							 end_pfn);
end_pfn           824 arch/x86/kernel/e820.c 		unsigned long end_pfn;
end_pfn           830 arch/x86/kernel/e820.c 		end_pfn = (entry->addr + entry->size) >> PAGE_SHIFT;
end_pfn           834 arch/x86/kernel/e820.c 		if (end_pfn > limit_pfn) {
end_pfn           838 arch/x86/kernel/e820.c 		if (end_pfn > last_pfn)
end_pfn           839 arch/x86/kernel/e820.c 			last_pfn = end_pfn;
end_pfn           262 arch/x86/mm/init.c 			     unsigned long start_pfn, unsigned long end_pfn,
end_pfn           265 arch/x86/mm/init.c 	if (start_pfn < end_pfn) {
end_pfn           269 arch/x86/mm/init.c 		mr[nr_range].end   = end_pfn<<PAGE_SHIFT;
end_pfn           340 arch/x86/mm/init.c 	unsigned long start_pfn, end_pfn, limit_pfn;
end_pfn           356 arch/x86/mm/init.c 		end_pfn = PFN_DOWN(PMD_SIZE);
end_pfn           358 arch/x86/mm/init.c 		end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE));
end_pfn           360 arch/x86/mm/init.c 	end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE));
end_pfn           362 arch/x86/mm/init.c 	if (end_pfn > limit_pfn)
end_pfn           363 arch/x86/mm/init.c 		end_pfn = limit_pfn;
end_pfn           364 arch/x86/mm/init.c 	if (start_pfn < end_pfn) {
end_pfn           365 arch/x86/mm/init.c 		nr_range = save_mr(mr, nr_range, start_pfn, end_pfn, 0);
end_pfn           366 arch/x86/mm/init.c 		pfn = end_pfn;
end_pfn           372 arch/x86/mm/init.c 	end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE));
end_pfn           374 arch/x86/mm/init.c 	end_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE));
end_pfn           375 arch/x86/mm/init.c 	if (end_pfn > round_down(limit_pfn, PFN_DOWN(PMD_SIZE)))
end_pfn           376 arch/x86/mm/init.c 		end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE));
end_pfn           379 arch/x86/mm/init.c 	if (start_pfn < end_pfn) {
end_pfn           380 arch/x86/mm/init.c 		nr_range = save_mr(mr, nr_range, start_pfn, end_pfn,
end_pfn           382 arch/x86/mm/init.c 		pfn = end_pfn;
end_pfn           388 arch/x86/mm/init.c 	end_pfn = round_down(limit_pfn, PFN_DOWN(PUD_SIZE));
end_pfn           389 arch/x86/mm/init.c 	if (start_pfn < end_pfn) {
end_pfn           390 arch/x86/mm/init.c 		nr_range = save_mr(mr, nr_range, start_pfn, end_pfn,
end_pfn           393 arch/x86/mm/init.c 		pfn = end_pfn;
end_pfn           398 arch/x86/mm/init.c 	end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE));
end_pfn           399 arch/x86/mm/init.c 	if (start_pfn < end_pfn) {
end_pfn           400 arch/x86/mm/init.c 		nr_range = save_mr(mr, nr_range, start_pfn, end_pfn,
end_pfn           402 arch/x86/mm/init.c 		pfn = end_pfn;
end_pfn           408 arch/x86/mm/init.c 	end_pfn = limit_pfn;
end_pfn           409 arch/x86/mm/init.c 	nr_range = save_mr(mr, nr_range, start_pfn, end_pfn, 0);
end_pfn           439 arch/x86/mm/init.c static void add_pfn_range_mapped(unsigned long start_pfn, unsigned long end_pfn)
end_pfn           442 arch/x86/mm/init.c 					     nr_pfn_mapped, start_pfn, end_pfn);
end_pfn           445 arch/x86/mm/init.c 	max_pfn_mapped = max(max_pfn_mapped, end_pfn);
end_pfn           449 arch/x86/mm/init.c 					 min(end_pfn, 1UL<<(32-PAGE_SHIFT)));
end_pfn           452 arch/x86/mm/init.c bool pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn)
end_pfn           458 arch/x86/mm/init.c 		    (end_pfn <= pfn_mapped[i].end))
end_pfn           508 arch/x86/mm/init.c 	unsigned long start_pfn, end_pfn;
end_pfn           512 arch/x86/mm/init.c 	for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) {
end_pfn           514 arch/x86/mm/init.c 		u64 end = clamp_val(PFN_PHYS(end_pfn), r_start, r_end);
end_pfn           899 arch/x86/mm/init.c 	unsigned long start_pfn, end_pfn;
end_pfn           909 arch/x86/mm/init.c 	for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) {
end_pfn           911 arch/x86/mm/init.c 		end_pfn   = min(end_pfn,   MAX_DMA_PFN);
end_pfn           913 arch/x86/mm/init.c 		nr_pages += end_pfn - start_pfn;
end_pfn           924 arch/x86/mm/init.c 		end_pfn   = min_t(unsigned long, PFN_DOWN(end_addr), MAX_DMA_PFN);
end_pfn           926 arch/x86/mm/init.c 		if (start_pfn < end_pfn)
end_pfn           927 arch/x86/mm/init.c 			nr_free_pages += end_pfn - start_pfn;
end_pfn           259 arch/x86/mm/init_32.c 	unsigned long start_pfn, end_pfn;
end_pfn           270 arch/x86/mm/init_32.c 	end_pfn = end >> PAGE_SHIFT;
end_pfn           299 arch/x86/mm/init_32.c 		if (pfn >= end_pfn)
end_pfn           307 arch/x86/mm/init_32.c 		for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn;
end_pfn           347 arch/x86/mm/init_32.c 			for (; pte_ofs < PTRS_PER_PTE && pfn < end_pfn;
end_pfn           435 arch/x86/mm/init_32.c 			 unsigned long start_pfn, unsigned long end_pfn)
end_pfn           442 arch/x86/mm/init_32.c 					    start_pfn, end_pfn);
end_pfn           444 arch/x86/mm/init_32.c 					      start_pfn, end_pfn);
end_pfn           837 arch/x86/mm/init_64.c 	unsigned long end_pfn = PFN_UP(start + size);
end_pfn           839 arch/x86/mm/init_64.c 	if (end_pfn > max_pfn) {
end_pfn           840 arch/x86/mm/init_64.c 		max_pfn = end_pfn;
end_pfn           841 arch/x86/mm/init_64.c 		max_low_pfn = end_pfn;
end_pfn            34 arch/x86/mm/numa_emulation.c 	unsigned long end_pfn = PFN_DOWN(end);
end_pfn            36 arch/x86/mm/numa_emulation.c 	if (start_pfn < end_pfn)
end_pfn            37 arch/x86/mm/numa_emulation.c 		return PFN_PHYS(absent_pages_in_range(start_pfn, end_pfn));
end_pfn           428 arch/x86/mm/pat.c 	unsigned long end_pfn = (end + PAGE_SIZE - 1) >> PAGE_SHIFT;
end_pfn           441 arch/x86/mm/pat.c 	if (start_pfn < end_pfn) {
end_pfn           442 arch/x86/mm/pat.c 		ret = walk_system_ram_range(start_pfn, end_pfn - start_pfn,
end_pfn           614 arch/x86/platform/efi/efi.c 	u64 start_pfn, end_pfn, end;
end_pfn           621 arch/x86/platform/efi/efi.c 	end_pfn   = PFN_UP(end);
end_pfn           623 arch/x86/platform/efi/efi.c 	if (pfn_range_is_mapped(start_pfn, end_pfn)) {
end_pfn           255 arch/x86/xen/setup.c 			unsigned long end_pfn, unsigned long nr_pages)
end_pfn           260 arch/x86/xen/setup.c 	WARN_ON(start_pfn > end_pfn);
end_pfn           263 arch/x86/xen/setup.c 	end = min(end_pfn, nr_pages);
end_pfn           282 arch/x86/xen/setup.c 	set_phys_range_identity(start_pfn, end_pfn);
end_pfn           388 arch/x86/xen/setup.c 	unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages,
end_pfn           393 arch/x86/xen/setup.c 	unsigned long n = end_pfn - start_pfn;
end_pfn           435 arch/x86/xen/setup.c 	for (pfn = start_pfn; pfn <= max_pfn_mapped && pfn < end_pfn; pfn++)
end_pfn           444 arch/x86/xen/setup.c 	unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages,
end_pfn           450 arch/x86/xen/setup.c 	return remap_pages + min(end_pfn, nr_pages) - start_pfn;
end_pfn           454 arch/x86/xen/setup.c 	unsigned long (*func)(unsigned long start_pfn, unsigned long end_pfn,
end_pfn           477 arch/x86/xen/setup.c 			unsigned long end_pfn = PFN_UP(end);
end_pfn           480 arch/x86/xen/setup.c 				end_pfn = PFN_UP(entry->addr);
end_pfn           482 arch/x86/xen/setup.c 			if (start_pfn < end_pfn)
end_pfn           483 arch/x86/xen/setup.c 				ret_val = func(start_pfn, end_pfn, nr_pages,
end_pfn           767 drivers/base/node.c 	unsigned long end_pfn = start_pfn + memory_block_pfns - 1;
end_pfn           771 drivers/base/node.c 	for (pfn = start_pfn; pfn <= end_pfn; pfn++) {
end_pfn           832 drivers/base/node.c int link_mem_sections(int nid, unsigned long start_pfn, unsigned long end_pfn)
end_pfn           835 drivers/base/node.c 				  PFN_PHYS(end_pfn - start_pfn), (void *)&nid,
end_pfn            77 drivers/dax/super.c 	pfn_t pfn, end_pfn;
end_pfn           105 drivers/dax/super.c 	len2 = dax_direct_access(dax_dev, pgoff_end, 1, &end_kaddr, &end_pfn);
end_pfn           125 drivers/dax/super.c 	} else if (pfn_t_devmap(pfn) && pfn_t_devmap(end_pfn)) {
end_pfn           129 drivers/dax/super.c 		end_pgmap = get_dev_pagemap(pfn_t_to_pfn(end_pfn), NULL);
end_pfn           132 drivers/dax/super.c 				&& pfn_t_to_page(end_pfn)->pgmap == pgmap
end_pfn           134 drivers/dax/super.c 				&& pfn_t_to_pfn(end_pfn) == PHYS_PFN(__pa(end_kaddr)))
end_pfn           436 drivers/hv/hv_balloon.c 	unsigned long end_pfn;
end_pfn           446 drivers/hv/hv_balloon.c 	unsigned long end_pfn;
end_pfn           583 drivers/hv/hv_balloon.c 		if ((pfn >= gap->start_pfn) && (pfn < gap->end_pfn))
end_pfn           605 drivers/hv/hv_balloon.c 			       (pfn < has->end_pfn) &&
end_pfn           779 drivers/hv/hv_balloon.c 				(pfn + (1UL << order) > has->end_pfn))
end_pfn           802 drivers/hv/hv_balloon.c 		if (start_pfn < has->start_pfn || start_pfn >= has->end_pfn)
end_pfn           818 drivers/hv/hv_balloon.c 			gap->end_pfn = start_pfn;
end_pfn           828 drivers/hv/hv_balloon.c 		if ((start_pfn + pfn_cnt) > has->end_pfn) {
end_pfn           829 drivers/hv/hv_balloon.c 			residual = (start_pfn + pfn_cnt - has->end_pfn);
end_pfn           837 drivers/hv/hv_balloon.c 			has->end_pfn += new_inc;
end_pfn           868 drivers/hv/hv_balloon.c 		if (start_pfn < has->start_pfn || start_pfn >= has->end_pfn)
end_pfn           899 drivers/hv/hv_balloon.c 		if ((has->ha_end_pfn < has->end_pfn) && (pfn_cnt > 0)) {
end_pfn           907 drivers/hv/hv_balloon.c 			size = (has->end_pfn - has->ha_end_pfn);
end_pfn           968 drivers/hv/hv_balloon.c 		ha_region->end_pfn = rg_start + rg_size;
end_pfn          2266 drivers/iommu/intel-iommu.c 				unsigned long nr_superpages, end_pfn;
end_pfn          2272 drivers/iommu/intel-iommu.c 				end_pfn = iov_pfn + nr_superpages * lvl_pages - 1;
end_pfn          2280 drivers/iommu/intel-iommu.c 				dma_pte_free_pagetable(domain, iov_pfn, end_pfn,
end_pfn          2753 drivers/iommu/intel-iommu.c 		unsigned long start_pfn, end_pfn;
end_pfn          2756 drivers/iommu/intel-iommu.c 		for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) {
end_pfn          2758 drivers/iommu/intel-iommu.c 					PFN_PHYS(start_pfn), PFN_PHYS(end_pfn));
end_pfn           689 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	unsigned long pfn, start_pfn, end_pfn, nr_pages;
end_pfn           697 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	end_pfn = initial_pfn + total_nr_pages;
end_pfn           700 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	while (pfn < end_pfn) {
end_pfn           679 drivers/nvdimm/pfn_devs.c 		.end_pfn = PHYS_PFN(end),
end_pfn           239 fs/proc/kcore.c 	unsigned long end_pfn;
end_pfn           243 fs/proc/kcore.c 	end_pfn = 0;
end_pfn           247 fs/proc/kcore.c 		if (end_pfn < node_end)
end_pfn           248 fs/proc/kcore.c 			end_pfn = node_end;
end_pfn           251 fs/proc/kcore.c 	ret = walk_system_ram_range(0, end_pfn, list, kclist_add_private);
end_pfn           485 include/linux/hugetlb.h 				    unsigned long end_pfn);
end_pfn           707 include/linux/hugetlb.h 					   unsigned long end_pfn)
end_pfn           221 include/linux/memblock.h 			    unsigned long  *end_pfn);
end_pfn            98 include/linux/memory_hotplug.h extern int test_pages_in_a_zone(unsigned long start_pfn, unsigned long end_pfn,
end_pfn           101 include/linux/memory_hotplug.h 						unsigned long end_pfn);
end_pfn            20 include/linux/memremap.h 	const unsigned long end_pfn;
end_pfn          2176 include/linux/mm.h 						unsigned long end_pfn);
end_pfn          2178 include/linux/mm.h 						unsigned long end_pfn);
end_pfn          2180 include/linux/mm.h 			unsigned long *start_pfn, unsigned long *end_pfn);
end_pfn          1312 include/linux/mmzone.h void online_mem_sections(unsigned long start_pfn, unsigned long end_pfn);
end_pfn          1314 include/linux/mmzone.h void offline_mem_sections(unsigned long start_pfn, unsigned long end_pfn);
end_pfn           103 include/linux/node.h 			     unsigned long end_pfn);
end_pfn           106 include/linux/node.h 				    unsigned long end_pfn)
end_pfn           125 include/linux/node.h 		unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages;
end_pfn           131 include/linux/node.h 		error = link_mem_sections(nid, start_pfn, end_pfn);
end_pfn            46 include/linux/page-isolation.h start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn,
end_pfn            54 include/linux/page-isolation.h undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn,
end_pfn            60 include/linux/page-isolation.h int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
end_pfn            18 include/trace/events/compaction.h 		unsigned long end_pfn,
end_pfn            22 include/trace/events/compaction.h 	TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken),
end_pfn            26 include/trace/events/compaction.h 		__field(unsigned long, end_pfn)
end_pfn            33 include/trace/events/compaction.h 		__entry->end_pfn = end_pfn;
end_pfn            40 include/trace/events/compaction.h 		__entry->end_pfn,
end_pfn            49 include/trace/events/compaction.h 		unsigned long end_pfn,
end_pfn            53 include/trace/events/compaction.h 	TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken)
end_pfn            60 include/trace/events/compaction.h 		unsigned long end_pfn,
end_pfn            64 include/trace/events/compaction.h 	TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken)
end_pfn            14 include/trace/events/page_isolation.h 		unsigned long end_pfn,
end_pfn            17 include/trace/events/page_isolation.h 	TP_ARGS(start_pfn, end_pfn, fin_pfn),
end_pfn            21 include/trace/events/page_isolation.h 		__field(unsigned long, end_pfn)
end_pfn            27 include/trace/events/page_isolation.h 		__entry->end_pfn = end_pfn;
end_pfn            32 include/trace/events/page_isolation.h 		__entry->start_pfn, __entry->end_pfn, __entry->fin_pfn,
end_pfn            33 include/trace/events/page_isolation.h 		__entry->end_pfn <= __entry->fin_pfn ? "success" : "fail")
end_pfn           362 kernel/power/snapshot.c 	unsigned long end_pfn;		/* Zone end page frame + 1     */
end_pfn           516 kernel/power/snapshot.c 	zone->end_pfn = end;
end_pfn           716 kernel/power/snapshot.c 	if (pfn >= zone->start_pfn && pfn < zone->end_pfn)
end_pfn           723 kernel/power/snapshot.c 		if (pfn >= curr->start_pfn && pfn < curr->end_pfn) {
end_pfn           888 kernel/power/snapshot.c 		pages	  = bm->cur.zone->end_pfn - bm->cur.zone->start_pfn;
end_pfn           909 kernel/power/snapshot.c 	unsigned long end_pfn;
end_pfn           949 kernel/power/snapshot.c 				     unsigned long end_pfn, int use_kmalloc)
end_pfn           953 kernel/power/snapshot.c 	if (start_pfn >= end_pfn)
end_pfn           960 kernel/power/snapshot.c 		if (region->end_pfn == start_pfn) {
end_pfn           961 kernel/power/snapshot.c 			region->end_pfn = end_pfn;
end_pfn           978 kernel/power/snapshot.c 	region->end_pfn = end_pfn;
end_pfn           983 kernel/power/snapshot.c 		((unsigned long long) end_pfn << PAGE_SHIFT) - 1);
end_pfn          1055 kernel/power/snapshot.c 			 ((unsigned long long) region->end_pfn << PAGE_SHIFT)
end_pfn          1058 kernel/power/snapshot.c 		for (pfn = region->start_pfn; pfn < region->end_pfn; pfn++)
end_pfn           481 kernel/resource.c 	unsigned long pfn, end_pfn;
end_pfn           491 kernel/resource.c 		end_pfn = PFN_DOWN(res.end + 1);
end_pfn           492 kernel/resource.c 		if (end_pfn > pfn)
end_pfn           493 kernel/resource.c 			ret = (*func)(pfn, end_pfn - pfn, arg);
end_pfn           537 mm/compaction.c 				unsigned long end_pfn,
end_pfn           556 mm/compaction.c 	for (; blockpfn < end_pfn; blockpfn += stride, cursor += stride) {
end_pfn           644 mm/compaction.c 	if (unlikely(blockpfn > end_pfn))
end_pfn           645 mm/compaction.c 		blockpfn = end_pfn;
end_pfn           658 mm/compaction.c 	if (strict && blockpfn < end_pfn)
end_pfn           683 mm/compaction.c 			unsigned long start_pfn, unsigned long end_pfn)
end_pfn           694 mm/compaction.c 	for (; pfn < end_pfn; pfn += isolated,
end_pfn           700 mm/compaction.c 		block_end_pfn = min(block_end_pfn, end_pfn);
end_pfn           710 mm/compaction.c 			block_end_pfn = min(block_end_pfn, end_pfn);
end_pfn           738 mm/compaction.c 	if (pfn < end_pfn) {
end_pfn           783 mm/compaction.c 			unsigned long end_pfn, isolate_mode_t isolate_mode)
end_pfn           820 mm/compaction.c 	for (; low_pfn < end_pfn; low_pfn++) {
end_pfn           870 mm/compaction.c 				low_pfn = end_pfn;
end_pfn          1042 mm/compaction.c 	if (unlikely(low_pfn > end_pfn))
end_pfn          1043 mm/compaction.c 		low_pfn = end_pfn;
end_pfn          1057 mm/compaction.c 	if (low_pfn == end_pfn && (!nr_isolated || cc->rescan)) {
end_pfn          1086 mm/compaction.c 							unsigned long end_pfn)
end_pfn          1097 mm/compaction.c 	for (; pfn < end_pfn; pfn = block_end_pfn,
end_pfn          1101 mm/compaction.c 		block_end_pfn = min(block_end_pfn, end_pfn);
end_pfn          1224 mm/compaction.c 	unsigned long start_pfn, end_pfn;
end_pfn          1237 mm/compaction.c 	end_pfn = min(pageblock_end_pfn(pfn), zone_end_pfn(cc->zone)) - 1;
end_pfn          1248 mm/compaction.c 	if (start_pfn < end_pfn)
end_pfn          1249 mm/compaction.c 		isolate_freepages_block(cc, &start_pfn, end_pfn, &cc->freepages, 1, false);
end_pfn          2076 mm/compaction.c 	unsigned long end_pfn = zone_end_pfn(cc->zone);
end_pfn          2118 mm/compaction.c 		cc->free_pfn = pageblock_start_pfn(end_pfn - 1);
end_pfn          2122 mm/compaction.c 		if (cc->free_pfn < start_pfn || cc->free_pfn >= end_pfn) {
end_pfn          2123 mm/compaction.c 			cc->free_pfn = pageblock_start_pfn(end_pfn - 1);
end_pfn          2126 mm/compaction.c 		if (cc->migrate_pfn < start_pfn || cc->migrate_pfn >= end_pfn) {
end_pfn          2150 mm/compaction.c 				cc->free_pfn, end_pfn, sync);
end_pfn          2283 mm/compaction.c 				cc->free_pfn, end_pfn, sync, ret);
end_pfn          1076 mm/hugetlb.c   	unsigned long end_pfn = start_pfn + nr_pages;
end_pfn          1077 mm/hugetlb.c   	return alloc_contig_range(start_pfn, end_pfn, MIGRATE_MOVABLE,
end_pfn          1084 mm/hugetlb.c   	unsigned long i, end_pfn = start_pfn + nr_pages;
end_pfn          1087 mm/hugetlb.c   	for (i = start_pfn; i < end_pfn; i++) {
end_pfn          1649 mm/hugetlb.c   int dissolve_free_huge_pages(unsigned long start_pfn, unsigned long end_pfn)
end_pfn          1658 mm/hugetlb.c   	for (pfn = start_pfn; pfn < end_pfn; pfn += 1 << minimum_order) {
end_pfn           148 mm/internal.h  				unsigned long end_pfn, struct zone *zone);
end_pfn           151 mm/internal.h  				unsigned long end_pfn, struct zone *zone)
end_pfn           156 mm/internal.h  	return __pageblock_pfn_to_page(start_pfn, end_pfn, zone);
end_pfn           219 mm/internal.h  			unsigned long start_pfn, unsigned long end_pfn);
end_pfn           222 mm/internal.h  			   unsigned long low_pfn, unsigned long end_pfn);
end_pfn           463 mm/internal.h  				unsigned long *end_pfn);
end_pfn           466 mm/internal.h  				unsigned long *end_pfn)
end_pfn          1444 mm/kmemleak.c  		unsigned long end_pfn = node_end_pfn(i);
end_pfn          1447 mm/kmemleak.c  		for (pfn = start_pfn; pfn < end_pfn; pfn++) {
end_pfn          2725 mm/ksm.c       					 unsigned long end_pfn)
end_pfn          2728 mm/ksm.c       	    stable_node->kpfn < end_pfn) {
end_pfn          2741 mm/ksm.c       					   unsigned long end_pfn,
end_pfn          2750 mm/ksm.c       						    end_pfn);
end_pfn          2756 mm/ksm.c       		stable_node_dup_remove_range(dup, start_pfn, end_pfn);
end_pfn          2766 mm/ksm.c       				  unsigned long end_pfn)
end_pfn          2777 mm/ksm.c       							   start_pfn, end_pfn,
end_pfn          2788 mm/ksm.c       		    stable_node->kpfn < end_pfn)
end_pfn          1601 mm/memblock.c  	unsigned long start_pfn, end_pfn;
end_pfn          1605 mm/memblock.c  		end_pfn = memblock_region_memory_end_pfn(r);
end_pfn          1607 mm/memblock.c  		end_pfn = min_t(unsigned long, end_pfn, limit_pfn);
end_pfn          1608 mm/memblock.c  		pages += end_pfn - start_pfn;
end_pfn          1751 mm/memblock.c  			 unsigned long *start_pfn, unsigned long *end_pfn)
end_pfn          1760 mm/memblock.c  	*end_pfn = PFN_DOWN(type->regions[mid].base + type->regions[mid].size);
end_pfn          1912 mm/memblock.c  	unsigned long end_pfn = min_t(unsigned long,
end_pfn          1915 mm/memblock.c  	if (start_pfn >= end_pfn)
end_pfn          1918 mm/memblock.c  	__free_pages_memory(start_pfn, end_pfn);
end_pfn          1920 mm/memblock.c  	return end_pfn - start_pfn;
end_pfn           227 mm/memory_hotplug.c 	unsigned long i, pfn, end_pfn, nr_pages;
end_pfn           238 mm/memory_hotplug.c 	end_pfn = pgdat_end_pfn(pgdat);
end_pfn           241 mm/memory_hotplug.c 	for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
end_pfn           331 mm/memory_hotplug.c 				     unsigned long end_pfn)
end_pfn           333 mm/memory_hotplug.c 	for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) {
end_pfn           352 mm/memory_hotplug.c 				    unsigned long end_pfn)
end_pfn           357 mm/memory_hotplug.c 	pfn = end_pfn - 1;
end_pfn           375 mm/memory_hotplug.c 			     unsigned long end_pfn)
end_pfn           391 mm/memory_hotplug.c 		pfn = find_smallest_section_pfn(nid, zone, end_pfn,
end_pfn           397 mm/memory_hotplug.c 	} else if (zone_end_pfn == end_pfn) {
end_pfn           425 mm/memory_hotplug.c 		if (pfn >= start_pfn && pfn < end_pfn)
end_pfn           619 mm/memory_hotplug.c 	const unsigned long end_pfn = start_pfn + nr_pages;
end_pfn           628 mm/memory_hotplug.c 	for (pfn = start_pfn; pfn < end_pfn; pfn += 1ul << order) {
end_pfn           629 mm/memory_hotplug.c 		order = min(MAX_ORDER - 1, get_order(PFN_PHYS(end_pfn - pfn)));
end_pfn           637 mm/memory_hotplug.c 	online_mem_sections(start_pfn, end_pfn);
end_pfn          1195 mm/memory_hotplug.c 	unsigned long end_pfn, pfn;
end_pfn          1197 mm/memory_hotplug.c 	end_pfn = min(start_pfn + nr_pages,
end_pfn          1201 mm/memory_hotplug.c 	for (pfn = start_pfn; pfn < end_pfn; pfn = next_active_pageblock(pfn)) {
end_pfn          1215 mm/memory_hotplug.c int test_pages_in_a_zone(unsigned long start_pfn, unsigned long end_pfn,
end_pfn          1224 mm/memory_hotplug.c 	     pfn < end_pfn;
end_pfn          1229 mm/memory_hotplug.c 		for (; pfn < sec_end_pfn && pfn < end_pfn;
end_pfn          1236 mm/memory_hotplug.c 			if (i == MAX_ORDER_NR_PAGES || pfn + i >= end_pfn)
end_pfn          1253 mm/memory_hotplug.c 		*valid_end = min(end, end_pfn);
end_pfn          1311 mm/memory_hotplug.c do_migrate_range(unsigned long start_pfn, unsigned long end_pfn)
end_pfn          1318 mm/memory_hotplug.c 	for (pfn = start_pfn; pfn < end_pfn; pfn++) {
end_pfn          1488 mm/memory_hotplug.c 		  unsigned long end_pfn)
end_pfn          1503 mm/memory_hotplug.c 	if (!test_pages_in_a_zone(start_pfn, end_pfn, &valid_start,
end_pfn          1512 mm/memory_hotplug.c 	nr_pages = end_pfn - start_pfn;
end_pfn          1515 mm/memory_hotplug.c 	ret = start_isolate_page_range(start_pfn, end_pfn,
end_pfn          1546 mm/memory_hotplug.c 			pfn = scan_movable_pages(pfn, end_pfn);
end_pfn          1552 mm/memory_hotplug.c 				do_migrate_range(pfn, end_pfn);
end_pfn          1561 mm/memory_hotplug.c 		ret = dissolve_free_huge_pages(start_pfn, end_pfn);
end_pfn          1567 mm/memory_hotplug.c 		ret = walk_system_ram_range(start_pfn, end_pfn - start_pfn,
end_pfn          1573 mm/memory_hotplug.c 	walk_system_ram_range(start_pfn, end_pfn - start_pfn,
end_pfn          1616 mm/memory_hotplug.c 	undo_isolate_page_range(start_pfn, end_pfn, MIGRATE_MOVABLE);
end_pfn          1621 mm/memory_hotplug.c 		 ((unsigned long long) end_pfn << PAGE_SHIFT) - 1,
end_pfn           405 mm/page_alloc.c defer_init(int nid, unsigned long pfn, unsigned long end_pfn)
end_pfn           413 mm/page_alloc.c 	if (prev_end_pfn != end_pfn) {
end_pfn           414 mm/page_alloc.c 		prev_end_pfn = end_pfn;
end_pfn           419 mm/page_alloc.c 	if (end_pfn < pgdat_end_pfn(NODE_DATA(nid)))
end_pfn           442 mm/page_alloc.c static inline bool defer_init(int nid, unsigned long pfn, unsigned long end_pfn)
end_pfn          1393 mm/page_alloc.c 	unsigned long end_pfn = PFN_UP(end);
end_pfn          1395 mm/page_alloc.c 	for (; start_pfn < end_pfn; start_pfn++) {
end_pfn          1516 mm/page_alloc.c 				     unsigned long end_pfn, struct zone *zone)
end_pfn          1522 mm/page_alloc.c 	end_pfn--;
end_pfn          1524 mm/page_alloc.c 	if (!pfn_valid(start_pfn) || !pfn_valid(end_pfn))
end_pfn          1534 mm/page_alloc.c 	end_page = pfn_to_page(end_pfn);
end_pfn          1631 mm/page_alloc.c 				       unsigned long end_pfn)
end_pfn          1636 mm/page_alloc.c 	for (; pfn < end_pfn; pfn++) {
end_pfn          1659 mm/page_alloc.c 						 unsigned long end_pfn)
end_pfn          1667 mm/page_alloc.c 	for (; pfn < end_pfn; pfn++) {
end_pfn          1725 mm/page_alloc.c 		       unsigned long *end_pfn)
end_pfn          1728 mm/page_alloc.c 	unsigned long spfn = *start_pfn, epfn = *end_pfn;
end_pfn          1733 mm/page_alloc.c 	for_each_free_mem_pfn_range_in_zone_from(j, zone, start_pfn, end_pfn) {
end_pfn          1739 mm/page_alloc.c 		t = min(mo_pfn, *end_pfn);
end_pfn          1742 mm/page_alloc.c 		if (mo_pfn < *end_pfn) {
end_pfn          2282 mm/page_alloc.c 	unsigned long start_pfn, end_pfn;
end_pfn          2292 mm/page_alloc.c 	end_pfn = start_pfn + pageblock_nr_pages - 1;
end_pfn          2297 mm/page_alloc.c 	if (!zone_spans_pfn(zone, end_pfn))
end_pfn          5881 mm/page_alloc.c 	unsigned long pfn, end_pfn = start_pfn + size;
end_pfn          5884 mm/page_alloc.c 	if (highest_memmap_pfn < end_pfn - 1)
end_pfn          5885 mm/page_alloc.c 		highest_memmap_pfn = end_pfn - 1;
end_pfn          5901 mm/page_alloc.c 		end_pfn = altmap->base_pfn + vmem_altmap_offset(altmap);
end_pfn          5905 mm/page_alloc.c 	for (pfn = start_pfn; pfn < end_pfn; pfn++) {
end_pfn          5917 mm/page_alloc.c 			if (defer_init(nid, pfn, end_pfn))
end_pfn          5951 mm/page_alloc.c 	unsigned long pfn, end_pfn = start_pfn + size;
end_pfn          5968 mm/page_alloc.c 		size = end_pfn - start_pfn;
end_pfn          5971 mm/page_alloc.c 	for (pfn = start_pfn; pfn < end_pfn; pfn++) {
end_pfn          6238 mm/page_alloc.c 	unsigned long start_pfn, end_pfn;
end_pfn          6244 mm/page_alloc.c 	nid = memblock_search_pfn_nid(pfn, &start_pfn, &end_pfn);
end_pfn          6247 mm/page_alloc.c 		state->last_end = end_pfn;
end_pfn          6266 mm/page_alloc.c 	unsigned long start_pfn, end_pfn;
end_pfn          6269 mm/page_alloc.c 	for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, &this_nid) {
end_pfn          6271 mm/page_alloc.c 		end_pfn = min(end_pfn, max_low_pfn);
end_pfn          6273 mm/page_alloc.c 		if (start_pfn < end_pfn)
end_pfn          6275 mm/page_alloc.c 					(end_pfn - start_pfn) << PAGE_SHIFT,
end_pfn          6289 mm/page_alloc.c 	unsigned long start_pfn, end_pfn;
end_pfn          6292 mm/page_alloc.c 	for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, &this_nid)
end_pfn          6293 mm/page_alloc.c 		memory_present(this_nid, start_pfn, end_pfn);
end_pfn          6308 mm/page_alloc.c 			unsigned long *start_pfn, unsigned long *end_pfn)
end_pfn          6314 mm/page_alloc.c 	*end_pfn = 0;
end_pfn          6318 mm/page_alloc.c 		*end_pfn = max(*end_pfn, this_end_pfn);
end_pfn          6429 mm/page_alloc.c 	unsigned long start_pfn, end_pfn;
end_pfn          6432 mm/page_alloc.c 	for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) {
end_pfn          6434 mm/page_alloc.c 		end_pfn = clamp(end_pfn, range_start_pfn, range_end_pfn);
end_pfn          6435 mm/page_alloc.c 		nr_absent -= end_pfn - start_pfn;
end_pfn          6448 mm/page_alloc.c 							unsigned long end_pfn)
end_pfn          6450 mm/page_alloc.c 	return __absent_pages_in_range(MAX_NUMNODES, start_pfn, end_pfn);
end_pfn          6483 mm/page_alloc.c 		unsigned long start_pfn, end_pfn;
end_pfn          6489 mm/page_alloc.c 			end_pfn = clamp(memblock_region_memory_end_pfn(r),
end_pfn          6494 mm/page_alloc.c 				nr_absent += end_pfn - start_pfn;
end_pfn          6498 mm/page_alloc.c 				nr_absent += end_pfn - start_pfn;
end_pfn          6887 mm/page_alloc.c 	unsigned long end_pfn = 0;
end_pfn          6896 mm/page_alloc.c 	get_pfn_range_for_nid(nid, &start_pfn, &end_pfn);
end_pfn          6899 mm/page_alloc.c 		end_pfn ? ((u64)end_pfn << PAGE_SHIFT) - 1 : 0);
end_pfn          6903 mm/page_alloc.c 	calculate_node_totalpages(pgdat, start_pfn, end_pfn,
end_pfn          7086 mm/page_alloc.c 	unsigned long start_pfn, end_pfn;
end_pfn          7089 mm/page_alloc.c 	for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) {
end_pfn          7090 mm/page_alloc.c 		unsigned long pages = end_pfn - start_pfn;
end_pfn          7217 mm/page_alloc.c 		unsigned long start_pfn, end_pfn;
end_pfn          7235 mm/page_alloc.c 		for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) {
end_pfn          7239 mm/page_alloc.c 			if (start_pfn >= end_pfn)
end_pfn          7245 mm/page_alloc.c 				kernel_pages = min(end_pfn, usable_startpfn)
end_pfn          7254 mm/page_alloc.c 				if (end_pfn <= usable_startpfn) {
end_pfn          7262 mm/page_alloc.c 					zone_movable_pfn[nid] = end_pfn;
end_pfn          7273 mm/page_alloc.c 			size_pages = end_pfn - start_pfn;
end_pfn          7344 mm/page_alloc.c 	unsigned long start_pfn, end_pfn;
end_pfn          7359 mm/page_alloc.c 		end_pfn = max(max_zone_pfn[i], start_pfn);
end_pfn          7361 mm/page_alloc.c 		arch_zone_highest_possible_pfn[i] = end_pfn;
end_pfn          7363 mm/page_alloc.c 		start_pfn = end_pfn;
end_pfn          7401 mm/page_alloc.c 	for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) {
end_pfn          7404 mm/page_alloc.c 			((u64)end_pfn << PAGE_SHIFT) - 1);
end_pfn          7405 mm/page_alloc.c 		subsection_map_init(start_pfn, end_pfn - start_pfn);
end_pfn          8567 mm/page_alloc.c __offline_isolated_pages(unsigned long start_pfn, unsigned long end_pfn)
end_pfn          8577 mm/page_alloc.c 	for (pfn = start_pfn; pfn < end_pfn; pfn++)
end_pfn          8580 mm/page_alloc.c 	if (pfn == end_pfn)
end_pfn          8583 mm/page_alloc.c 	offline_mem_sections(pfn, end_pfn);
end_pfn          8587 mm/page_alloc.c 	while (pfn < end_pfn) {
end_pfn          8610 mm/page_alloc.c 			pfn, 1 << order, end_pfn);
end_pfn           375 mm/page_ext.c  		unsigned long start_pfn, end_pfn;
end_pfn           378 mm/page_ext.c  		end_pfn = node_end_pfn(nid);
end_pfn           384 mm/page_ext.c  		for (pfn = start_pfn; pfn < end_pfn;
end_pfn           127 mm/page_idle.c 	unsigned long pfn, end_pfn;
end_pfn           137 mm/page_idle.c 	end_pfn = pfn + count * BITS_PER_BYTE;
end_pfn           138 mm/page_idle.c 	if (end_pfn > max_pfn)
end_pfn           139 mm/page_idle.c 		end_pfn = max_pfn;
end_pfn           141 mm/page_idle.c 	for (; pfn < end_pfn; pfn++) {
end_pfn           172 mm/page_idle.c 	unsigned long pfn, end_pfn;
end_pfn           182 mm/page_idle.c 	end_pfn = pfn + count * BITS_PER_BYTE;
end_pfn           183 mm/page_idle.c 	if (end_pfn > max_pfn)
end_pfn           184 mm/page_idle.c 		end_pfn = max_pfn;
end_pfn           186 mm/page_idle.c 	for (; pfn < end_pfn; pfn++) {
end_pfn           193 mm/page_isolation.c int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn,
end_pfn           202 mm/page_isolation.c 	BUG_ON(!IS_ALIGNED(end_pfn, pageblock_nr_pages));
end_pfn           205 mm/page_isolation.c 	     pfn < end_pfn;
end_pfn           233 mm/page_isolation.c void undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn,
end_pfn           240 mm/page_isolation.c 	BUG_ON(!IS_ALIGNED(end_pfn, pageblock_nr_pages));
end_pfn           243 mm/page_isolation.c 	     pfn < end_pfn;
end_pfn           259 mm/page_isolation.c __test_page_isolated_in_pageblock(unsigned long pfn, unsigned long end_pfn,
end_pfn           264 mm/page_isolation.c 	while (pfn < end_pfn) {
end_pfn           288 mm/page_isolation.c int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
end_pfn           300 mm/page_isolation.c 	for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) {
end_pfn           305 mm/page_isolation.c 	page = __first_valid_page(start_pfn, end_pfn - start_pfn);
end_pfn           306 mm/page_isolation.c 	if ((pfn < end_pfn) || !page)
end_pfn           311 mm/page_isolation.c 	pfn = __test_page_isolated_in_pageblock(start_pfn, end_pfn,
end_pfn           315 mm/page_isolation.c 	trace_test_pages_isolated(start_pfn, end_pfn, pfn);
end_pfn           317 mm/page_isolation.c 	return pfn < end_pfn ? -EBUSY : 0;
end_pfn           260 mm/page_owner.c 	unsigned long end_pfn = pfn + zone->spanned_pages;
end_pfn           273 mm/page_owner.c 	for (; pfn < end_pfn; ) {
end_pfn           281 mm/page_owner.c 		block_end_pfn = min(block_end_pfn, end_pfn);
end_pfn           549 mm/page_owner.c 	unsigned long end_pfn = zone_end_pfn(zone);
end_pfn           557 mm/page_owner.c 	for (; pfn < end_pfn; ) {
end_pfn           566 mm/page_owner.c 		block_end_pfn = min(block_end_pfn, end_pfn);
end_pfn           109 mm/shuffle.c   	unsigned long end_pfn = zone_end_pfn(z);
end_pfn           115 mm/shuffle.c   	for (i = start_pfn; i < end_pfn; i += order_pages) {
end_pfn           157 mm/sparse.c    						unsigned long *end_pfn)
end_pfn           168 mm/sparse.c    			*start_pfn, *end_pfn, max_sparsemem_pfn);
end_pfn           171 mm/sparse.c    		*end_pfn = max_sparsemem_pfn;
end_pfn           172 mm/sparse.c    	} else if (*end_pfn > max_sparsemem_pfn) {
end_pfn           175 mm/sparse.c    			*start_pfn, *end_pfn, max_sparsemem_pfn);
end_pfn           177 mm/sparse.c    		*end_pfn = max_sparsemem_pfn;
end_pfn           609 mm/sparse.c    void online_mem_sections(unsigned long start_pfn, unsigned long end_pfn)
end_pfn           613 mm/sparse.c    	for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
end_pfn           628 mm/sparse.c    void offline_mem_sections(unsigned long start_pfn, unsigned long end_pfn)
end_pfn           632 mm/sparse.c    	for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
end_pfn          1437 mm/vmstat.c    	unsigned long end_pfn = zone_end_pfn(zone);
end_pfn          1440 mm/vmstat.c    	for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) {