page_zone         401 drivers/base/memory.c 		strcat(buf, page_zone(pfn_to_page(start_pfn))->name);
page_zone         734 drivers/virtio/virtio_balloon.c 	    page_zone(page) != page_zone(newpage)) {
page_zone         196 fs/proc/kcore.c 	if (!memmap_valid_within(pfn, p, page_zone(p)))
page_zone          50 include/linux/migrate.h 	if (PageHighMem(page) || (zone_idx(page_zone(page)) == ZONE_MOVABLE))
page_zone         325 include/linux/vmstat.h 	__inc_zone_state(page_zone(page), item);
page_zone         338 include/linux/vmstat.h 	__dec_zone_state(page_zone(page), item);
page_zone         383 kernel/fork.c  			mod_zone_page_state(page_zone(vm->pages[i]),
page_zone         394 kernel/fork.c  		mod_zone_page_state(page_zone(first_page), NR_KERNEL_STACK_KB,
page_zone        1227 kernel/power/snapshot.c 	if (!page || page_zone(page) != zone)
page_zone        1291 kernel/power/snapshot.c 	if (!page || page_zone(page) != zone)
page_zone         111 mm/cma.c       	zone = page_zone(pfn_to_page(pfn));
page_zone         125 mm/cma.c       			if (page_zone(pfn_to_page(pfn)) != zone)
page_zone         251 mm/compaction.c 	if (zone != page_zone(page))
page_zone        2987 mm/huge_memory.c 			if (zone != page_zone(page))
page_zone        1092 mm/hugetlb.c   		if (page_zone(page) != z)
page_zone         333 mm/internal.h  		__mod_zone_page_state(page_zone(page), NR_MLOCK, -nr_pages);
page_zone         335 mm/internal.h  		__mod_zone_page_state(page_zone(newpage), NR_MLOCK, nr_pages);
page_zone        1707 mm/khugepaged.c 		struct zone *zone = page_zone(new_page);
page_zone         250 mm/memory-failure.c 		drain_all_pages(page_zone(p));
page_zone         340 mm/memory_hotplug.c 		if (zone && zone != page_zone(pfn_to_page(start_pfn)))
page_zone         365 mm/memory_hotplug.c 		if (zone && zone != page_zone(pfn_to_page(pfn)))
page_zone         421 mm/memory_hotplug.c 		if (page_zone(pfn_to_page(pfn)) != zone)
page_zone        1184 mm/memory_hotplug.c 	zone = page_zone(page);
page_zone        1198 mm/memory_hotplug.c 			zone_end_pfn(page_zone(pfn_to_page(start_pfn))));
page_zone        1242 mm/memory_hotplug.c 			if (zone && page_zone(page) != zone)
page_zone        1246 mm/memory_hotplug.c 			zone = page_zone(page);
page_zone        1510 mm/memory_hotplug.c 	zone = page_zone(pfn_to_page(valid_start));
page_zone        2068 mm/mempolicy.c 		__inc_numa_state(page_zone(page), NUMA_INTERLEAVE_HIT);
page_zone         419 mm/migrate.c   	oldzone = page_zone(page);
page_zone         420 mm/migrate.c   	newzone = page_zone(newpage);
page_zone          64 mm/mlock.c     	mod_zone_page_state(page_zone(page), NR_MLOCK,
page_zone          97 mm/mlock.c     		mod_zone_page_state(page_zone(page), NR_MLOCK,
page_zone         206 mm/mlock.c     	__mod_zone_page_state(page_zone(page), NR_MLOCK, -nr_pages);
page_zone         404 mm/mlock.c     		if (!page || page_zone(page) != zone)
page_zone         489 mm/mlock.c     				zone = page_zone(page);
page_zone          82 mm/mmzone.c    	if (page_zone(page) != zone)
page_zone         455 mm/page_alloc.c 	return page_zone(page)->pageblock_flags;
page_zone         465 mm/page_alloc.c 	pfn = pfn - round_down(page_zone(page)->zone_start_pfn, pageblock_nr_pages);
page_zone         535 mm/page_alloc.c 	VM_BUG_ON_PAGE(!zone_spans_pfn(page_zone(page), pfn), page);
page_zone         588 mm/page_alloc.c 	if (zone != page_zone(page))
page_zone        1426 mm/page_alloc.c 	free_one_page(page_zone(page), page, pfn, order, migratetype);
page_zone        1445 mm/page_alloc.c 	atomic_long_add(nr_pages, &page_zone(page)->managed_pages);
page_zone        1531 mm/page_alloc.c 	if (page_zone(start_page) != zone)
page_zone        2268 mm/page_alloc.c 		VM_BUG_ON_PAGE(page_zone(page) != zone, page);
page_zone        2990 mm/page_alloc.c 			if (page_zone(page) != zone)
page_zone        3030 mm/page_alloc.c 	struct zone *zone = page_zone(page);
page_zone        3138 mm/page_alloc.c 	struct free_area *area = &page_zone(page)->free_area[order];
page_zone        3145 mm/page_alloc.c 	zone = page_zone(page);
page_zone        3920 mm/page_alloc.c 		struct zone *zone = page_zone(page);
page_zone        7484 mm/page_alloc.c 	atomic_long_add(count, &page_zone(page)->managed_pages);
page_zone        7530 mm/page_alloc.c 	atomic_long_inc(&page_zone(page)->managed_pages);
page_zone        8393 mm/page_alloc.c 		.zone = page_zone(pfn_to_page(start)),
page_zone        8584 mm/page_alloc.c 	zone = page_zone(pfn_to_page(pfn));
page_zone        8625 mm/page_alloc.c 	struct zone *zone = page_zone(page);
page_zone        8650 mm/page_alloc.c 	struct zone *zone = page_zone(page);
page_zone          26 mm/page_isolation.c 	zone = page_zone(page);
page_zone          99 mm/page_isolation.c 	zone = page_zone(page);
page_zone         309 mm/page_isolation.c 	zone = page_zone(page);
page_zone         292 mm/page_owner.c 			if (page_zone(page) != zone)
page_zone         577 mm/page_owner.c 			if (page_zone(page) != zone)
page_zone         468 mm/swap.c      		__mod_zone_page_state(page_zone(page), NR_MLOCK,
page_zone         413 mm/vmstat.c    	__inc_zone_state(page_zone(page), item);
page_zone         457 mm/vmstat.c    	__dec_zone_state(page_zone(page), item);
page_zone         527 mm/vmstat.c    	mod_zone_state(page_zone(page), item, 1, 1);
page_zone         533 mm/vmstat.c    	mod_zone_state(page_zone(page), item, -1, -1);
page_zone         618 mm/vmstat.c    	zone = page_zone(page);
page_zone        1451 mm/vmstat.c    		if (page_zone(page) != zone)
page_zone        2072 mm/zsmalloc.c  	if (page_zone(newpage) != page_zone(page)) {