nr_pages          193 arch/alpha/mm/init.c 		unsigned long nr_pages = 0;
nr_pages          199 arch/alpha/mm/init.c 			nr_pages += crb->map[i].count;
nr_pages          203 arch/alpha/mm/init.c 		console_remap_vm.size = nr_pages << PAGE_SHIFT;
nr_pages           43 arch/arm/kernel/atags_compat.c 	    unsigned long nr_pages;		/*  4 */
nr_pages          104 arch/arm/kernel/atags_compat.c 	if (params->u1.s.nr_pages != 0x02000 &&
nr_pages          105 arch/arm/kernel/atags_compat.c 	    params->u1.s.nr_pages != 0x04000 &&
nr_pages          106 arch/arm/kernel/atags_compat.c 	    params->u1.s.nr_pages != 0x08000 &&
nr_pages          107 arch/arm/kernel/atags_compat.c 	    params->u1.s.nr_pages != 0x10000) {
nr_pages          110 arch/arm/kernel/atags_compat.c 		params->u1.s.nr_pages = 0x1000;	/* 16MB */
nr_pages          158 arch/arm/kernel/atags_compat.c 	tag = memtag(tag, PHYS_OFFSET, params->u1.s.nr_pages * PAGE_SIZE);
nr_pages           20 arch/arm/mach-rpc/include/mach/uncompress.h 	unsigned long nr_pages;
nr_pages          116 arch/arm/mach-rpc/include/mach/uncompress.h 	unsigned int nr_pages = 0, page_size = PAGE_SIZE;
nr_pages          130 arch/arm/mach-rpc/include/mach/uncompress.h 				nr_pages += (t->u.mem.size / PAGE_SIZE);
nr_pages          134 arch/arm/mach-rpc/include/mach/uncompress.h 		nr_pages = params->nr_pages;
nr_pages          179 arch/arm/mach-rpc/include/mach/uncompress.h 	if (nr_pages * page_size < 4096*1024) error("<4M of mem\n");
nr_pages          793 arch/arm/mm/dma-mapping.c 	unsigned long nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT;
nr_pages          800 arch/arm/mm/dma-mapping.c 	if (off < nr_pages && nr_vma_pages <= (nr_pages - off)) {
nr_pages         1527 arch/arm/mm/dma-mapping.c 	unsigned long nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT;
nr_pages         1533 arch/arm/mm/dma-mapping.c 	if (vma->vm_pgoff >= nr_pages)
nr_pages         1536 arch/arm/mm/dma-mapping.c 	err = vm_map_pages(vma, pages, nr_pages);
nr_pages           24 arch/arm/xen/p2m.c 	unsigned long nr_pages;
nr_pages           73 arch/arm/xen/p2m.c 				entry->pfn + entry->nr_pages > pfn) {
nr_pages          122 arch/arm/xen/p2m.c 		unsigned long mfn, unsigned long nr_pages)
nr_pages          134 arch/arm/xen/p2m.c 					p2m_entry->pfn + p2m_entry->nr_pages > pfn) {
nr_pages          154 arch/arm/xen/p2m.c 	p2m_entry->nr_pages = nr_pages;
nr_pages         1071 arch/arm64/mm/mmu.c 	unsigned long nr_pages = size >> PAGE_SHIFT;
nr_pages         1080 arch/arm64/mm/mmu.c 	__remove_pages(start_pfn, nr_pages, altmap);
nr_pages           96 arch/arm64/mm/pageattr.c 		for (i = 0; i < area->nr_pages; i++) {
nr_pages          676 arch/ia64/mm/init.c 	unsigned long nr_pages = size >> PAGE_SHIFT;
nr_pages          679 arch/ia64/mm/init.c 	ret = __add_pages(nid, start_pfn, nr_pages, restrictions);
nr_pages          691 arch/ia64/mm/init.c 	unsigned long nr_pages = size >> PAGE_SHIFT;
nr_pages          693 arch/ia64/mm/init.c 	__remove_pages(start_pfn, nr_pages, altmap);
nr_pages          277 arch/mips/mm/init.c static int maar_res_walk(unsigned long start_pfn, unsigned long nr_pages,
nr_pages          289 arch/mips/mm/init.c 	cfg->upper = ALIGN_DOWN(PFN_PHYS(start_pfn + nr_pages), maar_align) - 1;
nr_pages          102 arch/mips/mm/ioremap.c static int __ioremap_check_ram(unsigned long start_pfn, unsigned long nr_pages,
nr_pages          107 arch/mips/mm/ioremap.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          199 arch/powerpc/include/asm/kvm_ppc.h extern struct page *kvm_alloc_hpt_cma(unsigned long nr_pages);
nr_pages          200 arch/powerpc/include/asm/kvm_ppc.h extern void kvm_free_hpt_cma(struct page *page, unsigned long nr_pages);
nr_pages           72 arch/powerpc/kvm/book3s_hv_builtin.c struct page *kvm_alloc_hpt_cma(unsigned long nr_pages)
nr_pages           74 arch/powerpc/kvm/book3s_hv_builtin.c 	VM_BUG_ON(order_base_2(nr_pages) < KVM_CMA_CHUNK_ORDER - PAGE_SHIFT);
nr_pages           76 arch/powerpc/kvm/book3s_hv_builtin.c 	return cma_alloc(kvm_cma, nr_pages, order_base_2(HPT_ALIGN_PAGES),
nr_pages           81 arch/powerpc/kvm/book3s_hv_builtin.c void kvm_free_hpt_cma(struct page *page, unsigned long nr_pages)
nr_pages           83 arch/powerpc/kvm/book3s_hv_builtin.c 	cma_release(kvm_cma, page, nr_pages);
nr_pages          842 arch/powerpc/mm/book3s64/radix_tlb.c 	unsigned long nr_pages = (end - start) >> page_shift;
nr_pages          860 arch/powerpc/mm/book3s64/radix_tlb.c 				nr_pages > tlb_single_page_flush_ceiling);
nr_pages          865 arch/powerpc/mm/book3s64/radix_tlb.c 				nr_pages > tlb_local_single_page_flush_ceiling);
nr_pages         1080 arch/powerpc/mm/book3s64/radix_tlb.c 	unsigned long nr_pages = (end - start) >> page_shift;
nr_pages         1098 arch/powerpc/mm/book3s64/radix_tlb.c 				nr_pages > tlb_single_page_flush_ceiling);
nr_pages         1103 arch/powerpc/mm/book3s64/radix_tlb.c 				nr_pages > tlb_local_single_page_flush_ceiling);
nr_pages          289 arch/powerpc/mm/init_64.c 		unsigned long nr_pages, addr;
nr_pages          305 arch/powerpc/mm/init_64.c 		nr_pages = 1 << page_order;
nr_pages          309 arch/powerpc/mm/init_64.c 			vmem_altmap_free(altmap, nr_pages);
nr_pages          319 arch/powerpc/mm/init_64.c 				while (nr_pages--)
nr_pages          132 arch/powerpc/mm/mem.c 	unsigned long nr_pages = size >> PAGE_SHIFT;
nr_pages          147 arch/powerpc/mm/mem.c 	return __add_pages(nid, start_pfn, nr_pages, restrictions);
nr_pages          154 arch/powerpc/mm/mem.c 	unsigned long nr_pages = size >> PAGE_SHIFT;
nr_pages          157 arch/powerpc/mm/mem.c 	__remove_pages(start_pfn, nr_pages, altmap);
nr_pages           71 arch/powerpc/platforms/powernv/memtrace.c static bool memtrace_offline_pages(u32 nid, u64 start_pfn, u64 nr_pages)
nr_pages           74 arch/powerpc/platforms/powernv/memtrace.c 	const unsigned long size = PFN_PHYS(nr_pages);
nr_pages           82 arch/powerpc/platforms/powernv/memtrace.c 	if (offline_pages(start_pfn, nr_pages)) {
nr_pages           97 arch/powerpc/platforms/powernv/memtrace.c 	u64 start_pfn, end_pfn, nr_pages, pfn;
nr_pages          106 arch/powerpc/platforms/powernv/memtrace.c 	nr_pages = size >> PAGE_SHIFT;
nr_pages          109 arch/powerpc/platforms/powernv/memtrace.c 	end_pfn = round_down(end_pfn - nr_pages, nr_pages);
nr_pages          112 arch/powerpc/platforms/powernv/memtrace.c 	for (base_pfn = end_pfn; base_pfn > start_pfn; base_pfn -= nr_pages) {
nr_pages          113 arch/powerpc/platforms/powernv/memtrace.c 		if (memtrace_offline_pages(nid, base_pfn, nr_pages) == true) {
nr_pages          120 arch/powerpc/platforms/powernv/memtrace.c 			end_pfn = base_pfn + nr_pages;
nr_pages          501 arch/powerpc/platforms/pseries/cmm.c 	unsigned long end = start + (marg->nr_pages << PAGE_SHIFT);
nr_pages          554 arch/powerpc/platforms/pseries/cmm.c 	unsigned long end_page = start_page + (marg->nr_pages << PAGE_SHIFT);
nr_pages          560 arch/powerpc/platforms/pseries/cmm.c 			start_page, marg->nr_pages);
nr_pages         1245 arch/powerpc/platforms/pseries/iommu.c 					arg->nr_pages, window->prop);
nr_pages         1255 arch/powerpc/platforms/pseries/iommu.c 					arg->nr_pages, window->prop);
nr_pages          370 arch/s390/include/asm/cio.h struct gen_pool *cio_gp_dma_create(struct device *dma_dev, int nr_pages);
nr_pages         1692 arch/s390/kernel/perf_cpum_sf.c 			      int nr_pages, bool snapshot)
nr_pages         1699 arch/s390/kernel/perf_cpum_sf.c 	if (!nr_pages || !pages)
nr_pages         1702 arch/s390/kernel/perf_cpum_sf.c 	if (nr_pages > CPUM_SF_MAX_SDB * CPUM_SF_SDB_DIAG_FACTOR) {
nr_pages         1705 arch/s390/kernel/perf_cpum_sf.c 		       nr_pages);
nr_pages         1707 arch/s390/kernel/perf_cpum_sf.c 	} else if (nr_pages < CPUM_SF_MIN_SDB * CPUM_SF_SDB_DIAG_FACTOR) {
nr_pages         1710 arch/s390/kernel/perf_cpum_sf.c 		       nr_pages);
nr_pages         1721 arch/s390/kernel/perf_cpum_sf.c 	n_sdbt = (nr_pages + CPUM_SF_SDB_PER_TABLE - 1) / CPUM_SF_SDB_PER_TABLE;
nr_pages         1727 arch/s390/kernel/perf_cpum_sf.c 	aux->sdb_index = kmalloc_array(nr_pages, sizeof(void *), GFP_KERNEL);
nr_pages         1743 arch/s390/kernel/perf_cpum_sf.c 	for (i = 0; i < nr_pages; i++, tail++) {
nr_pages         1758 arch/s390/kernel/perf_cpum_sf.c 	sfb->num_sdb = nr_pages;
nr_pages          623 arch/s390/kernel/setup.c 	if (arg->start_pfn + arg->nr_pages - 1 < PFN_DOWN(crashk_res.start))
nr_pages          798 arch/s390/kvm/gaccess.c 			    unsigned long *pages, unsigned long nr_pages,
nr_pages          806 arch/s390/kvm/gaccess.c 	while (nr_pages) {
nr_pages          825 arch/s390/kvm/gaccess.c 		nr_pages--;
nr_pages          834 arch/s390/kvm/gaccess.c 	unsigned long _len, nr_pages, gpa, idx;
nr_pages          847 arch/s390/kvm/gaccess.c 	nr_pages = (((ga & ~PAGE_MASK) + len - 1) >> PAGE_SHIFT) + 1;
nr_pages          849 arch/s390/kvm/gaccess.c 	if (nr_pages > ARRAY_SIZE(pages_array))
nr_pages          850 arch/s390/kvm/gaccess.c 		pages = vmalloc(array_size(nr_pages, sizeof(unsigned long)));
nr_pages          856 arch/s390/kvm/gaccess.c 	rc = guest_page_range(vcpu, ga, ar, pages, nr_pages, asce, mode);
nr_pages          857 arch/s390/kvm/gaccess.c 	for (idx = 0; idx < nr_pages && !rc; idx++) {
nr_pages          870 arch/s390/kvm/gaccess.c 	if (nr_pages > ARRAY_SIZE(pages_array))
nr_pages          251 arch/s390/mm/init.c 	mem_data.end = mem_data.start + (arg->nr_pages << PAGE_SHIFT);
nr_pages          293 arch/s390/mm/init.c 	unsigned long nr_pages = size >> PAGE_SHIFT;
nr_pages          295 arch/s390/mm/init.c 	__remove_pages(start_pfn, nr_pages, altmap);
nr_pages          136 arch/s390/pci/pci_dma.c 	unsigned int nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT;
nr_pages          142 arch/s390/pci/pci_dma.c 	if (!nr_pages)
nr_pages          151 arch/s390/pci/pci_dma.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          343 arch/s390/pci/pci_dma.c 	unsigned long nr_pages;
nr_pages          348 arch/s390/pci/pci_dma.c 	nr_pages = iommu_num_pages(pa, size, PAGE_SIZE);
nr_pages          349 arch/s390/pci/pci_dma.c 	dma_addr = dma_alloc_address(dev, nr_pages);
nr_pages          356 arch/s390/pci/pci_dma.c 	size = nr_pages * PAGE_SIZE;
nr_pages          365 arch/s390/pci/pci_dma.c 	atomic64_add(nr_pages, &zdev->mapped_pages);
nr_pages          369 arch/s390/pci/pci_dma.c 	dma_free_address(dev, dma_addr, nr_pages);
nr_pages          441 arch/s390/pci/pci_dma.c 	unsigned long nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT;
nr_pages          449 arch/s390/pci/pci_dma.c 	dma_addr_base = dma_alloc_address(dev, nr_pages);
nr_pages          471 arch/s390/pci/pci_dma.c 	atomic64_add(nr_pages, &zdev->mapped_pages);
nr_pages          478 arch/s390/pci/pci_dma.c 	dma_free_address(dev, dma_addr_base, nr_pages);
nr_pages          372 arch/sh/kernel/cpu/sh4/sq.c 	unsigned int nr_pages = 0x04000000 >> PAGE_SHIFT;
nr_pages          373 arch/sh/kernel/cpu/sh4/sq.c 	unsigned int size = (nr_pages + (BITS_PER_LONG - 1)) / BITS_PER_LONG;
nr_pages          412 arch/sh/mm/init.c 	unsigned long nr_pages = size >> PAGE_SHIFT;
nr_pages          416 arch/sh/mm/init.c 	ret = __add_pages(nid, start_pfn, nr_pages, restrictions);
nr_pages          436 arch/sh/mm/init.c 	unsigned long nr_pages = size >> PAGE_SHIFT;
nr_pages          438 arch/sh/mm/init.c 	__remove_pages(start_pfn, nr_pages, altmap);
nr_pages           53 arch/x86/events/intel/bts.c 	unsigned int	nr_pages;
nr_pages           81 arch/x86/events/intel/bts.c 		     int nr_pages, bool overwrite)
nr_pages           88 arch/x86/events/intel/bts.c 	size_t size = nr_pages << PAGE_SHIFT;
nr_pages           92 arch/x86/events/intel/bts.c 	for (pg = 0, nbuf = 0; pg < nr_pages;) {
nr_pages          108 arch/x86/events/intel/bts.c 	buf->nr_pages = nr_pages;
nr_pages          321 arch/x86/events/intel/bts.c 						   buf->nr_pages << PAGE_SHIFT);
nr_pages          379 arch/x86/events/intel/bts.c 	head = handle->head & ((buf->nr_pages << PAGE_SHIFT) - 1);
nr_pages          707 arch/x86/events/intel/pt.c 	p = virt_to_page(buf->data_pages[buf->nr_pages]);
nr_pages          735 arch/x86/events/intel/pt.c 	buf->nr_pages += 1ul << order;
nr_pages          816 arch/x86/events/intel/pt.c 		       ((buf->nr_pages << PAGE_SHIFT) - 1));
nr_pages          818 arch/x86/events/intel/pt.c 			base += buf->nr_pages << PAGE_SHIFT;
nr_pages          930 arch/x86/events/intel/pt.c 	if (WARN_ON_ONCE(pg >= buf->nr_pages))
nr_pages         1061 arch/x86/events/intel/pt.c 	idx &= buf->nr_pages - 1;
nr_pages         1076 arch/x86/events/intel/pt.c 	idx &= buf->nr_pages - 1;
nr_pages         1112 arch/x86/events/intel/pt.c 		head &= (buf->nr_pages << PAGE_SHIFT) - 1;
nr_pages         1114 arch/x86/events/intel/pt.c 	pg = (head >> PAGE_SHIFT) & (buf->nr_pages - 1);
nr_pages         1150 arch/x86/events/intel/pt.c 			       unsigned long nr_pages, gfp_t gfp)
nr_pages         1161 arch/x86/events/intel/pt.c 	while (buf->nr_pages < nr_pages) {
nr_pages         1193 arch/x86/events/intel/pt.c 		    int nr_pages, bool snapshot)
nr_pages         1198 arch/x86/events/intel/pt.c 	if (!nr_pages)
nr_pages         1216 arch/x86/events/intel/pt.c 	ret = pt_buffer_init_topa(buf, cpu, nr_pages, GFP_KERNEL);
nr_pages         1494 arch/x86/events/intel/pt.c 					   buf->nr_pages << PAGE_SHIFT);
nr_pages           79 arch/x86/events/intel/pt.h 	unsigned long		nr_pages;
nr_pages          434 arch/x86/kernel/cpu/microcode/core.c 	unsigned long nr_pages = totalram_pages();
nr_pages          436 arch/x86/kernel/cpu/microcode/core.c 	if ((len >> PAGE_SHIFT) > nr_pages) {
nr_pages          437 arch/x86/kernel/cpu/microcode/core.c 		pr_err("too much data (max %ld pages)\n", nr_pages);
nr_pages          209 arch/x86/kernel/ldt.c 	int i, nr_pages;
nr_pages          225 arch/x86/kernel/ldt.c 	nr_pages = DIV_ROUND_UP(ldt->nr_entries * LDT_ENTRY_SIZE, PAGE_SIZE);
nr_pages          227 arch/x86/kernel/ldt.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          268 arch/x86/kernel/ldt.c 	int i, nr_pages;
nr_pages          277 arch/x86/kernel/ldt.c 	nr_pages = DIV_ROUND_UP(ldt->nr_entries * LDT_ENTRY_SIZE, PAGE_SIZE);
nr_pages          279 arch/x86/kernel/ldt.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          291 arch/x86/kernel/ldt.c 	flush_tlb_mm_range(mm, va, va + nr_pages * PAGE_SIZE, PAGE_SHIFT, false);
nr_pages          616 arch/x86/kernel/machine_kexec_64.c 	unsigned int nr_pages;
nr_pages          626 arch/x86/kernel/machine_kexec_64.c 	nr_pages = (end >> PAGE_SHIFT) - (start >> PAGE_SHIFT) + 1;
nr_pages          628 arch/x86/kernel/machine_kexec_64.c 		return set_pages_ro(page, nr_pages);
nr_pages          630 arch/x86/kernel/machine_kexec_64.c 		return set_pages_rw(page, nr_pages);
nr_pages         6364 arch/x86/kvm/mmu.c 	unsigned long nr_pages = 0;
nr_pages         6373 arch/x86/kvm/mmu.c 			nr_pages += memslot->npages;
nr_pages         6376 arch/x86/kvm/mmu.c 	nr_mmu_pages = nr_pages * KVM_PERMILLE_MMU_PAGES / 1000;
nr_pages          898 arch/x86/mm/init.c 	u64 nr_pages = 0, nr_free_pages = 0;
nr_pages          908 arch/x86/mm/init.c 	nr_pages = 0;
nr_pages          913 arch/x86/mm/init.c 		nr_pages += end_pfn - start_pfn;
nr_pages          930 arch/x86/mm/init.c 	set_dma_reserve(nr_pages - nr_free_pages);
nr_pages          858 arch/x86/mm/init_32.c 	unsigned long nr_pages = size >> PAGE_SHIFT;
nr_pages          860 arch/x86/mm/init_32.c 	return __add_pages(nid, start_pfn, nr_pages, restrictions);
nr_pages          867 arch/x86/mm/init_32.c 	unsigned long nr_pages = size >> PAGE_SHIFT;
nr_pages          869 arch/x86/mm/init_32.c 	__remove_pages(start_pfn, nr_pages, altmap);
nr_pages          846 arch/x86/mm/init_64.c int add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages,
nr_pages          851 arch/x86/mm/init_64.c 	ret = __add_pages(nid, start_pfn, nr_pages, restrictions);
nr_pages          856 arch/x86/mm/init_64.c 				  nr_pages << PAGE_SHIFT);
nr_pages          865 arch/x86/mm/init_64.c 	unsigned long nr_pages = size >> PAGE_SHIFT;
nr_pages          869 arch/x86/mm/init_64.c 	return add_pages(nid, start_pfn, nr_pages, restrictions);
nr_pages          877 arch/x86/mm/init_64.c 	unsigned int nr_pages = 1 << order;
nr_pages          885 arch/x86/mm/init_64.c 			while (nr_pages--)
nr_pages          888 arch/x86/mm/init_64.c 			while (nr_pages--)
nr_pages         1214 arch/x86/mm/init_64.c 	unsigned long nr_pages = size >> PAGE_SHIFT;
nr_pages         1216 arch/x86/mm/init_64.c 	__remove_pages(start_pfn, nr_pages, altmap);
nr_pages         1536 arch/x86/mm/init_64.c 				  struct page *start_page, unsigned long nr_pages)
nr_pages         1539 arch/x86/mm/init_64.c 	unsigned long end = (unsigned long)(start_page + nr_pages);
nr_pages          421 arch/x86/mm/mpx.c 	int nr_pages = 1;
nr_pages          423 arch/x86/mm/mpx.c 	gup_ret = get_user_pages((unsigned long)addr, nr_pages,
nr_pages         1306 arch/x86/xen/enlighten_pv.c 				   xen_start_info->nr_pages);
nr_pages         1227 arch/x86/xen/mmu_pv.c 	size = PAGE_ALIGN(xen_start_info->nr_pages * sizeof(unsigned long));
nr_pages         1248 arch/x86/xen/mmu_pv.c 		size = PAGE_ALIGN(xen_start_info->nr_pages *
nr_pages         2067 arch/x86/xen/mmu_pv.c 	size = PAGE_ALIGN(xen_start_info->nr_pages * sizeof(unsigned long));
nr_pages          300 arch/x86/xen/p2m.c 	xen_p2m_size = ALIGN(xen_start_info->nr_pages, P2M_PER_PAGE);
nr_pages          302 arch/x86/xen/p2m.c 	for (pfn = xen_start_info->nr_pages; pfn < xen_p2m_size; pfn++)
nr_pages          255 arch/x86/xen/setup.c 			unsigned long end_pfn, unsigned long nr_pages)
nr_pages          263 arch/x86/xen/setup.c 	end = min(end_pfn, nr_pages);
nr_pages          388 arch/x86/xen/setup.c 	unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages,
nr_pages          396 arch/x86/xen/setup.c 		remap_pfn = nr_pages;
nr_pages          405 arch/x86/xen/setup.c 		if (cur_pfn >= nr_pages) {
nr_pages          410 arch/x86/xen/setup.c 		if (cur_pfn + size > nr_pages)
nr_pages          411 arch/x86/xen/setup.c 			size = nr_pages - cur_pfn;
nr_pages          417 arch/x86/xen/setup.c 						cur_pfn + left, nr_pages);
nr_pages          444 arch/x86/xen/setup.c 	unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages,
nr_pages          447 arch/x86/xen/setup.c 	if (start_pfn >= nr_pages)
nr_pages          450 arch/x86/xen/setup.c 	return remap_pages + min(end_pfn, nr_pages) - start_pfn;
nr_pages          453 arch/x86/xen/setup.c static unsigned long __init xen_foreach_remap_area(unsigned long nr_pages,
nr_pages          455 arch/x86/xen/setup.c 			      unsigned long nr_pages, unsigned long last_val))
nr_pages          483 arch/x86/xen/setup.c 				ret_val = func(start_pfn, end_pfn, nr_pages,
nr_pages          713 arch/x86/xen/setup.c 		size = PFN_ALIGN(xen_start_info->nr_pages *
nr_pages          754 arch/x86/xen/setup.c 	max_pfn = min(max_pfn, xen_start_info->nr_pages);
nr_pages          206 block/bio-integrity.c 	unsigned int len, nr_pages;
nr_pages          246 block/bio-integrity.c 	nr_pages = end - start;
nr_pages          249 block/bio-integrity.c 	bip = bio_integrity_alloc(bio, GFP_NOIO, nr_pages);
nr_pages          266 block/bio-integrity.c 	for (i = 0 ; i < nr_pages ; i++) {
nr_pages          922 block/bio.c    	unsigned short nr_pages = bio->bi_max_vecs - bio->bi_vcnt;
nr_pages          939 block/bio.c    	size = iov_iter_get_pages(iter, pages, LONG_MAX, nr_pages, &offset);
nr_pages         1282 block/bio.c    	int nr_pages;
nr_pages         1297 block/bio.c    	nr_pages = DIV_ROUND_UP(offset + len, PAGE_SIZE);
nr_pages         1298 block/bio.c    	if (nr_pages > BIO_MAX_PAGES)
nr_pages         1299 block/bio.c    		nr_pages = BIO_MAX_PAGES;
nr_pages         1302 block/bio.c    	bio = bio_kmalloc(gfp_mask, nr_pages);
nr_pages         1309 block/bio.c    		nr_pages = 1 << map_data->page_order;
nr_pages         1321 block/bio.c    			if (i == map_data->nr_entries * nr_pages) {
nr_pages         1326 block/bio.c    			page = map_data->pages[i / nr_pages];
nr_pages         1327 block/bio.c    			page += (i % nr_pages);
nr_pages         1523 block/bio.c    	const int nr_pages = end - start;
nr_pages         1529 block/bio.c    	bio = bio_kmalloc(gfp_mask, nr_pages);
nr_pages         1539 block/bio.c    	for (i = 0; i < nr_pages; i++) {
nr_pages         1607 block/bio.c    	int nr_pages = 0;
nr_pages         1615 block/bio.c    	nr_pages = end - start;
nr_pages         1616 block/bio.c    	bio = bio_kmalloc(gfp_mask, nr_pages);
nr_pages           13 block/blk-lib.c struct bio *blk_next_bio(struct bio *bio, unsigned int nr_pages, gfp_t gfp)
nr_pages           15 block/blk-lib.c 	struct bio *new = bio_alloc(gfp, nr_pages);
nr_pages          352 block/blk.h    struct bio *blk_next_bio(struct bio *bio, unsigned int nr_pages, gfp_t gfp);
nr_pages           70 drivers/base/firmware_loader/firmware.h 	int nr_pages;
nr_pages          278 drivers/base/firmware_loader/main.c 	for (i = 0; i < fw_priv->nr_pages; i++)
nr_pages          283 drivers/base/firmware_loader/main.c 	fw_priv->nr_pages = 0;
nr_pages          307 drivers/base/firmware_loader/main.c 	while (fw_priv->nr_pages < pages_needed) {
nr_pages          308 drivers/base/firmware_loader/main.c 		fw_priv->pages[fw_priv->nr_pages] =
nr_pages          311 drivers/base/firmware_loader/main.c 		if (!fw_priv->pages[fw_priv->nr_pages])
nr_pages          313 drivers/base/firmware_loader/main.c 		fw_priv->nr_pages++;
nr_pages          326 drivers/base/firmware_loader/main.c 	fw_priv->data = vmap(fw_priv->pages, fw_priv->nr_pages, 0,
nr_pages          400 drivers/base/firmware_loader/main.c 		if (fw_grow_paged_buf(fw_priv, fw_priv->nr_pages + 1)) {
nr_pages          406 drivers/base/firmware_loader/main.c 		page = fw_priv->pages[fw_priv->nr_pages - 1];
nr_pages          217 drivers/base/memory.c 	unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block;
nr_pages          227 drivers/base/memory.c 		ret = online_pages(start_pfn, nr_pages, online_type);
nr_pages          230 drivers/base/memory.c 		ret = offline_pages(start_pfn, nr_pages);
nr_pages          366 drivers/base/memory.c 		unsigned long nr_pages, int online_type,
nr_pages          371 drivers/base/memory.c 	zone = zone_for_pfn_range(online_type, nid, start_pfn, nr_pages);
nr_pages          383 drivers/base/memory.c 	unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block;
nr_pages          397 drivers/base/memory.c 		if (!test_pages_in_a_zone(start_pfn, start_pfn + nr_pages,
nr_pages          406 drivers/base/memory.c 	default_zone = zone_for_pfn_range(MMOP_ONLINE_KEEP, nid, start_pfn, nr_pages);
nr_pages          409 drivers/base/memory.c 	print_allowed_zone(buf, nid, start_pfn, nr_pages, MMOP_ONLINE_KERNEL,
nr_pages          411 drivers/base/memory.c 	print_allowed_zone(buf, nid, start_pfn, nr_pages, MMOP_ONLINE_MOVABLE,
nr_pages          136 drivers/block/brd.c 	int nr_pages;
nr_pages          141 drivers/block/brd.c 		nr_pages = radix_tree_gang_lookup(&brd->brd_pages,
nr_pages          144 drivers/block/brd.c 		for (i = 0; i < nr_pages; i++) {
nr_pages          167 drivers/block/brd.c 	} while (nr_pages == FREE_BATCH);
nr_pages          364 drivers/block/drbd/drbd_receiver.c 	unsigned nr_pages = (payload_size + PAGE_SIZE -1) >> PAGE_SHIFT;
nr_pages          376 drivers/block/drbd/drbd_receiver.c 	if (nr_pages) {
nr_pages          377 drivers/block/drbd/drbd_receiver.c 		page = drbd_alloc_pages(peer_device, nr_pages,
nr_pages         1647 drivers/block/drbd/drbd_receiver.c 	unsigned nr_pages = (data_size + PAGE_SIZE -1) >> PAGE_SHIFT;
nr_pages         1689 drivers/block/drbd/drbd_receiver.c 	bio = bio_alloc(GFP_NOIO, nr_pages);
nr_pages         1691 drivers/block/drbd/drbd_receiver.c 		drbd_err(device, "submit_ee: Allocation of a bio failed (nr_pages=%u)\n", nr_pages);
nr_pages         1711 drivers/block/drbd/drbd_receiver.c 		--nr_pages;
nr_pages          735 drivers/block/null_blk_main.c 	int nr_pages;
nr_pages          744 drivers/block/null_blk_main.c 		nr_pages = radix_tree_gang_lookup(root,
nr_pages          747 drivers/block/null_blk_main.c 		for (i = 0; i < nr_pages; i++) {
nr_pages          755 drivers/block/null_blk_main.c 	} while (nr_pages == FREE_BATCH);
nr_pages          880 drivers/block/null_blk_main.c 	int i, err, nr_pages;
nr_pages          889 drivers/block/null_blk_main.c 	nr_pages = radix_tree_gang_lookup(&nullb->dev->cache,
nr_pages          895 drivers/block/null_blk_main.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          908 drivers/block/null_blk_main.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          919 drivers/block/null_blk_main.c 		if (nr_pages == 0)
nr_pages          291 drivers/block/zram/zram_drv.c 	unsigned long nr_pages = zram->disksize >> PAGE_SHIFT;
nr_pages          303 drivers/block/zram/zram_drv.c 	for (index = 0; index < nr_pages; index++) {
nr_pages          452 drivers/block/zram/zram_drv.c 	unsigned long nr_pages, *bitmap = NULL;
nr_pages          497 drivers/block/zram/zram_drv.c 	nr_pages = i_size_read(inode) >> PAGE_SHIFT;
nr_pages          498 drivers/block/zram/zram_drv.c 	bitmap_sz = BITS_TO_LONGS(nr_pages) * sizeof(long);
nr_pages          516 drivers/block/zram/zram_drv.c 	zram->nr_pages = nr_pages;
nr_pages          557 drivers/block/zram/zram_drv.c 	blk_idx = find_next_zero_bit(zram->bitmap, zram->nr_pages, blk_idx);
nr_pages          558 drivers/block/zram/zram_drv.c 	if (blk_idx == zram->nr_pages)
nr_pages          624 drivers/block/zram/zram_drv.c 	unsigned long nr_pages = zram->disksize >> PAGE_SHIFT;
nr_pages          657 drivers/block/zram/zram_drv.c 	for (index = 0; index < nr_pages; index++) {
nr_pages          868 drivers/block/zram/zram_drv.c 	unsigned long nr_pages = zram->disksize >> PAGE_SHIFT;
nr_pages          882 drivers/block/zram/zram_drv.c 	for (index = *ppos; index < nr_pages; index++) {
nr_pages          123 drivers/block/zram/zram_drv.h 	unsigned long nr_pages;
nr_pages          296 drivers/dax/super.c long dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, long nr_pages,
nr_pages          307 drivers/dax/super.c 	if (nr_pages < 0)
nr_pages          308 drivers/dax/super.c 		return nr_pages;
nr_pages          310 drivers/dax/super.c 	avail = dax_dev->ops->direct_access(dax_dev, pgoff, nr_pages,
nr_pages          314 drivers/dax/super.c 	return min(avail, nr_pages);
nr_pages          391 drivers/edac/altera_edac.c 	dimm->nr_pages = ((mem_size - 1) >> PAGE_SHIFT) + 1;
nr_pages         2883 drivers/edac/amd64_edac.c 	u32 cs_mode, nr_pages;
nr_pages         2892 drivers/edac/amd64_edac.c 	nr_pages   = pvt->ops->dbam_to_cs(pvt, dct, cs_mode, csrow_nr);
nr_pages         2893 drivers/edac/amd64_edac.c 	nr_pages <<= 20 - PAGE_SHIFT;
nr_pages         2897 drivers/edac/amd64_edac.c 	edac_dbg(0, "nr_pages/channel: %u\n", nr_pages);
nr_pages         2899 drivers/edac/amd64_edac.c 	return nr_pages;
nr_pages         2935 drivers/edac/amd64_edac.c 			dimm->nr_pages = get_csrow_nr_pages(pvt, umc, cs);
nr_pages         2957 drivers/edac/amd64_edac.c 	int nr_pages = 0;
nr_pages         2991 drivers/edac/amd64_edac.c 			nr_pages = get_csrow_nr_pages(pvt, 0, i);
nr_pages         2992 drivers/edac/amd64_edac.c 			csrow->channels[0]->dimm->nr_pages = nr_pages;
nr_pages         2999 drivers/edac/amd64_edac.c 			csrow->channels[1]->dimm->nr_pages = row_dct1_pages;
nr_pages         3000 drivers/edac/amd64_edac.c 			nr_pages += row_dct1_pages;
nr_pages         3003 drivers/edac/amd64_edac.c 		edac_dbg(1, "Total csrow%d pages: %u\n", i, nr_pages);
nr_pages          210 drivers/edac/amd76x_edac.c 		dimm->nr_pages = (mba_mask + 1) >> PAGE_SHIFT;
nr_pages          211 drivers/edac/amd76x_edac.c 		csrow->last_page = csrow->first_page + dimm->nr_pages - 1;
nr_pages          213 drivers/edac/amd76x_edac.c 		dimm->grain = dimm->nr_pages << PAGE_SHIFT;
nr_pages          251 drivers/edac/armada_xp_edac.c 			dimm->nr_pages = 524288;
nr_pages          254 drivers/edac/armada_xp_edac.c 			dimm->nr_pages = 65536;
nr_pages          257 drivers/edac/armada_xp_edac.c 			dimm->nr_pages = 131072;
nr_pages          260 drivers/edac/armada_xp_edac.c 			dimm->nr_pages = 262144;
nr_pages          263 drivers/edac/armada_xp_edac.c 			dimm->nr_pages = 1048576;
nr_pages          266 drivers/edac/armada_xp_edac.c 			dimm->nr_pages = 2097152;
nr_pages          234 drivers/edac/aspeed_edac.c 	u32 nr_pages, dram_type;
nr_pages          261 drivers/edac/aspeed_edac.c 	nr_pages = resource_size(&r) >> PAGE_SHIFT;
nr_pages          262 drivers/edac/aspeed_edac.c 	csrow->last_page = csrow->first_page + nr_pages - 1;
nr_pages          270 drivers/edac/aspeed_edac.c 	dimm->nr_pages = nr_pages / csrow->nr_channels;
nr_pages          273 drivers/edac/aspeed_edac.c 		csrow->first_page, nr_pages);
nr_pages          212 drivers/edac/bluefield_edac.c 		dimm->nr_pages =
nr_pages          135 drivers/edac/cell_edac.c 	u32				nr_pages;
nr_pages          149 drivers/edac/cell_edac.c 		nr_pages = resource_size(&r) >> PAGE_SHIFT;
nr_pages          150 drivers/edac/cell_edac.c 		csrow->last_page = csrow->first_page + nr_pages - 1;
nr_pages          156 drivers/edac/cell_edac.c 			dimm->nr_pages = nr_pages / csrow->nr_channels;
nr_pages          162 drivers/edac/cell_edac.c 			csrow->first_page, nr_pages);
nr_pages          322 drivers/edac/cpc925_edac.c 	unsigned long row_size, nr_pages, last_nr_pages = 0;
nr_pages          341 drivers/edac/cpc925_edac.c 		nr_pages = row_size >> PAGE_SHIFT;
nr_pages          342 drivers/edac/cpc925_edac.c 		csrow->last_page = csrow->first_page + nr_pages - 1;
nr_pages          370 drivers/edac/cpc925_edac.c 			dimm->nr_pages = nr_pages / csrow->nr_channels;
nr_pages         1076 drivers/edac/e752x_edac.c 	u32 dra, drc, cumul_size, i, nr_pages;
nr_pages         1109 drivers/edac/e752x_edac.c 		nr_pages = cumul_size - last_cumul_size;
nr_pages         1130 drivers/edac/e752x_edac.c 			dimm->nr_pages = nr_pages / csrow->nr_channels;
nr_pages          360 drivers/edac/e7xxx_edac.c 	u32 dra, cumul_size, nr_pages;
nr_pages          391 drivers/edac/e7xxx_edac.c 		nr_pages = cumul_size - last_cumul_size;
nr_pages          412 drivers/edac/e7xxx_edac.c 			dimm->nr_pages = nr_pages / (drc_chan + 1);
nr_pages          159 drivers/edac/edac_mc.c 	edac_dbg(4, "  dimm->nr_pages = 0x%x\n", dimm->nr_pages);
nr_pages          161 drivers/edac/edac_mc.c 	edac_dbg(4, "  dimm->nr_pages = 0x%x\n", dimm->nr_pages);
nr_pages          715 drivers/edac/edac_mc.c 			u32 nr_pages = 0;
nr_pages          719 drivers/edac/edac_mc.c 				nr_pages += csrow->channels[j]->dimm->nr_pages;
nr_pages          720 drivers/edac/edac_mc.c 			if (!nr_pages)
nr_pages          724 drivers/edac/edac_mc.c 				if (csrow->channels[j]->dimm->nr_pages)
nr_pages          728 drivers/edac/edac_mc.c 			if (mci->dimms[i]->nr_pages)
nr_pages          865 drivers/edac/edac_mc.c 			n += dimm->nr_pages;
nr_pages         1167 drivers/edac/edac_mc.c 		if (e->enable_per_layer_report && dimm->nr_pages) {
nr_pages          165 drivers/edac/edac_mc_sysfs.c 	u32 nr_pages = 0;
nr_pages          168 drivers/edac/edac_mc_sysfs.c 		nr_pages += csrow->channels[i]->dimm->nr_pages;
nr_pages          169 drivers/edac/edac_mc_sysfs.c 	return sprintf(data, "%u\n", PAGES_TO_MiB(nr_pages));
nr_pages          367 drivers/edac/edac_mc_sysfs.c 	if (!csrow->channels[idx]->dimm->nr_pages)
nr_pages          392 drivers/edac/edac_mc_sysfs.c 	int chan, nr_pages = 0;
nr_pages          395 drivers/edac/edac_mc_sysfs.c 		nr_pages += csrow->channels[chan]->dimm->nr_pages;
nr_pages          397 drivers/edac/edac_mc_sysfs.c 	return nr_pages;
nr_pages          522 drivers/edac/edac_mc_sysfs.c 	return sprintf(data, "%u\n", PAGES_TO_MiB(dimm->nr_pages));
nr_pages          819 drivers/edac/edac_mc_sysfs.c 			total_pages += dimm->nr_pages;
nr_pages          948 drivers/edac/edac_mc_sysfs.c 		if (!dimm->nr_pages)
nr_pages          968 drivers/edac/edac_mc_sysfs.c 		if (!dimm->nr_pages)
nr_pages          996 drivers/edac/edac_mc_sysfs.c 		if (dimm->nr_pages == 0)
nr_pages          466 drivers/edac/fsl_ddr_edac.c 		dimm->nr_pages = end + 1 - start;
nr_pages          117 drivers/edac/ghes_edac.c 			dimm->nr_pages = MiB_TO_PAGES(32);/* Unknown */
nr_pages          119 drivers/edac/ghes_edac.c 			dimm->nr_pages = MiB_TO_PAGES(entry->extended_size);
nr_pages          122 drivers/edac/ghes_edac.c 				dimm->nr_pages = MiB_TO_PAGES((entry->size & 0x7fff) << 10);
nr_pages          124 drivers/edac/ghes_edac.c 				dimm->nr_pages = MiB_TO_PAGES(entry->size);
nr_pages          188 drivers/edac/ghes_edac.c 		if (dimm->nr_pages) {
nr_pages          191 drivers/edac/ghes_edac.c 				PAGES_TO_MiB(dimm->nr_pages),
nr_pages          556 drivers/edac/ghes_edac.c 		dimm->nr_pages = 1;
nr_pages          222 drivers/edac/highbank_mc_edac.c 	dimm->nr_pages = (~0UL >> PAGE_SHIFT) + 1;
nr_pages          315 drivers/edac/i3000_edac.c 	unsigned long last_cumul_size, nr_pages;
nr_pages          405 drivers/edac/i3000_edac.c 		nr_pages = cumul_size - last_cumul_size;
nr_pages          411 drivers/edac/i3000_edac.c 			dimm->nr_pages = nr_pages / nr_channels;
nr_pages          392 drivers/edac/i3200_edac.c 		unsigned long nr_pages;
nr_pages          398 drivers/edac/i3200_edac.c 			nr_pages = drb_to_nr_pages(drbs, stacked, j, i);
nr_pages          399 drivers/edac/i3200_edac.c 			if (nr_pages == 0)
nr_pages          403 drivers/edac/i3200_edac.c 				 stacked ? " (stacked)" : "", PAGES_TO_MiB(nr_pages));
nr_pages          405 drivers/edac/i3200_edac.c 			dimm->nr_pages = nr_pages;
nr_pages          406 drivers/edac/i3200_edac.c 			dimm->grain = nr_pages << PAGE_SHIFT;
nr_pages         1295 drivers/edac/i5000_edac.c 			dimm->nr_pages = csrow_megs << 8;
nr_pages          866 drivers/edac/i5100_edac.c 		dimm->nr_pages = npages;
nr_pages         1208 drivers/edac/i5400_edac.c 			dimm->nr_pages = size_mb << 8;
nr_pages          649 drivers/edac/i7300_edac.c 	dimm->nr_pages = MiB_TO_PAGES(dinfo->megabytes);
nr_pages          605 drivers/edac/i7core_edac.c 			dimm->nr_pages = npages;
nr_pages          220 drivers/edac/i82443bxgx_edac.c 		dimm->nr_pages = csrow->last_page - csrow->first_page + 1;
nr_pages          176 drivers/edac/i82860_edac.c 		dimm->nr_pages = cumul_size - last_cumul_size;
nr_pages          350 drivers/edac/i82875p_edac.c 	u32 cumul_size, nr_pages;
nr_pages          373 drivers/edac/i82875p_edac.c 		nr_pages = cumul_size - last_cumul_size;
nr_pages          379 drivers/edac/i82875p_edac.c 			dimm->nr_pages = nr_pages / nr_chans;
nr_pages          367 drivers/edac/i82975x_edac.c 	u32 cumul_size, nr_pages;
nr_pages          397 drivers/edac/i82975x_edac.c 		nr_pages = cumul_size - last_cumul_size;
nr_pages          398 drivers/edac/i82975x_edac.c 		if (!nr_pages)
nr_pages          410 drivers/edac/i82975x_edac.c 			dimm->nr_pages = nr_pages / csrow->nr_channels;
nr_pages          485 drivers/edac/ie31200_edac.c 			unsigned long nr_pages;
nr_pages          487 drivers/edac/ie31200_edac.c 			nr_pages = IE31200_PAGES(dimm_info[j][i].size, skl);
nr_pages          488 drivers/edac/ie31200_edac.c 			if (nr_pages == 0)
nr_pages          492 drivers/edac/ie31200_edac.c 				nr_pages = nr_pages / 2;
nr_pages          496 drivers/edac/ie31200_edac.c 				dimm->nr_pages = nr_pages;
nr_pages          497 drivers/edac/ie31200_edac.c 				edac_dbg(0, "set nr pages: 0x%lx\n", nr_pages);
nr_pages          508 drivers/edac/ie31200_edac.c 			dimm->nr_pages = nr_pages;
nr_pages          509 drivers/edac/ie31200_edac.c 			edac_dbg(0, "set nr pages: 0x%lx\n", nr_pages);
nr_pages          672 drivers/edac/mv64x60_edac.c 	dimm->nr_pages = pdata->total_mem >> PAGE_SHIFT;
nr_pages          145 drivers/edac/pasemi_edac.c 			dimm->nr_pages = 128 << (20 - PAGE_SHIFT);
nr_pages          148 drivers/edac/pasemi_edac.c 			dimm->nr_pages = 256 << (20 - PAGE_SHIFT);
nr_pages          152 drivers/edac/pasemi_edac.c 			dimm->nr_pages = 512 << (20 - PAGE_SHIFT);
nr_pages          155 drivers/edac/pasemi_edac.c 			dimm->nr_pages = 1024 << (20 - PAGE_SHIFT);
nr_pages          158 drivers/edac/pasemi_edac.c 			dimm->nr_pages = 2048 << (20 - PAGE_SHIFT);
nr_pages          168 drivers/edac/pasemi_edac.c 		csrow->last_page = csrow->first_page + dimm->nr_pages - 1;
nr_pages          169 drivers/edac/pasemi_edac.c 		last_page_in_mmc += dimm->nr_pages;
nr_pages         1256 drivers/edac/pnd2_edac.c 		dimm->nr_pages = MiB_TO_PAGES(capacity >> (20 - 3));
nr_pages         1322 drivers/edac/pnd2_edac.c 			dimm->nr_pages = MiB_TO_PAGES(capacity >> (20 - 3));
nr_pages          897 drivers/edac/ppc4xx_edac.c 	u32 mbxcf, size, nr_pages;
nr_pages          948 drivers/edac/ppc4xx_edac.c 			nr_pages = SDRAM_MBCF_SZ_TO_PAGES(size);
nr_pages          974 drivers/edac/ppc4xx_edac.c 			dimm->nr_pages  = nr_pages / csi->nr_channels;
nr_pages          255 drivers/edac/r82600_edac.c 		dimm->nr_pages = csrow->last_page - csrow->first_page + 1;
nr_pages         1661 drivers/edac/sb_edac.c 				dimm->nr_pages = npages;
nr_pages          312 drivers/edac/skx_common.c 	dimm->nr_pages = npages;
nr_pages          356 drivers/edac/skx_common.c 	dimm->nr_pages = size >> PAGE_SHIFT;
nr_pages          363 drivers/edac/skx_common.c 		 imc->mc, chan, dimmno, size >> 20, dimm->nr_pages);
nr_pages          787 drivers/edac/synopsys_edac.c 			dimm->nr_pages	= (size >> PAGE_SHIFT) / csi->nr_channels;
nr_pages          181 drivers/edac/ti_edac.c 	dimm->nr_pages = memsize >> PAGE_SHIFT;
nr_pages          373 drivers/edac/x38_edac.c 		unsigned long nr_pages;
nr_pages          376 drivers/edac/x38_edac.c 		nr_pages = drb_to_nr_pages(drbs, stacked,
nr_pages          380 drivers/edac/x38_edac.c 		if (nr_pages == 0)
nr_pages          386 drivers/edac/x38_edac.c 			dimm->nr_pages = nr_pages / x38_channel_num;
nr_pages          387 drivers/edac/x38_edac.c 			dimm->grain = nr_pages << PAGE_SHIFT;
nr_pages           71 drivers/firmware/efi/libstub/arm32-stub.c 	unsigned long nr_pages, map_size, desc_size, buff_size;
nr_pages           94 drivers/firmware/efi/libstub/arm32-stub.c 	nr_pages = MAX_UNCOMP_KERNEL_SIZE / EFI_PAGE_SIZE;
nr_pages           96 drivers/firmware/efi/libstub/arm32-stub.c 				EFI_BOOT_SERVICES_DATA, nr_pages, &alloc_addr);
nr_pages          177 drivers/firmware/efi/libstub/efi-stub-helper.c 	unsigned long nr_pages;
nr_pages          203 drivers/firmware/efi/libstub/efi-stub-helper.c 	nr_pages = size / EFI_PAGE_SIZE;
nr_pages          214 drivers/firmware/efi/libstub/efi-stub-helper.c 		if (desc->num_pages < nr_pages)
nr_pages          247 drivers/firmware/efi/libstub/efi-stub-helper.c 					nr_pages, &max_addr);
nr_pages          272 drivers/firmware/efi/libstub/efi-stub-helper.c 	unsigned long nr_pages;
nr_pages          297 drivers/firmware/efi/libstub/efi-stub-helper.c 	nr_pages = size / EFI_PAGE_SIZE;
nr_pages          308 drivers/firmware/efi/libstub/efi-stub-helper.c 		if (desc->num_pages < nr_pages)
nr_pages          323 drivers/firmware/efi/libstub/efi-stub-helper.c 					nr_pages, &start);
nr_pages          341 drivers/firmware/efi/libstub/efi-stub-helper.c 	unsigned long nr_pages;
nr_pages          346 drivers/firmware/efi/libstub/efi-stub-helper.c 	nr_pages = round_up(size, EFI_ALLOC_ALIGN) / EFI_PAGE_SIZE;
nr_pages          347 drivers/firmware/efi/libstub/efi-stub-helper.c 	efi_call_early(free_pages, addr, nr_pages);
nr_pages          702 drivers/firmware/efi/libstub/efi-stub-helper.c 	unsigned long nr_pages;
nr_pages          720 drivers/firmware/efi/libstub/efi-stub-helper.c 	nr_pages = round_up(alloc_size, EFI_ALLOC_ALIGN) / EFI_PAGE_SIZE;
nr_pages          723 drivers/firmware/efi/libstub/efi-stub-helper.c 				nr_pages, &efi_addr);
nr_pages          254 drivers/fpga/fpga-mgr.c 	int nr_pages;
nr_pages          270 drivers/fpga/fpga-mgr.c 	nr_pages = DIV_ROUND_UP((unsigned long)buf + count, PAGE_SIZE) -
nr_pages          272 drivers/fpga/fpga-mgr.c 	pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_KERNEL);
nr_pages          277 drivers/fpga/fpga-mgr.c 	for (index = 0; index < nr_pages; index++) {
nr_pages          793 drivers/gpu/drm/drm_prime.c struct sg_table *drm_prime_pages_to_sg(struct page **pages, unsigned int nr_pages)
nr_pages          804 drivers/gpu/drm/drm_prime.c 	ret = sg_alloc_table_from_pages(sg, pages, nr_pages, 0,
nr_pages          805 drivers/gpu/drm/drm_prime.c 				nr_pages << PAGE_SHIFT, GFP_KERNEL);
nr_pages           79 drivers/gpu/drm/exynos/exynos_drm_fbdev.c 	unsigned int nr_pages;
nr_pages           93 drivers/gpu/drm/exynos/exynos_drm_fbdev.c 	nr_pages = exynos_gem->size >> PAGE_SHIFT;
nr_pages           95 drivers/gpu/drm/exynos/exynos_drm_fbdev.c 	exynos_gem->kvaddr = (void __iomem *) vmap(exynos_gem->pages, nr_pages,
nr_pages           24 drivers/gpu/drm/exynos/exynos_drm_gem.c 	unsigned int nr_pages;
nr_pages           56 drivers/gpu/drm/exynos/exynos_drm_gem.c 	nr_pages = exynos_gem->size >> PAGE_SHIFT;
nr_pages           58 drivers/gpu/drm/exynos/exynos_drm_gem.c 	exynos_gem->pages = kvmalloc_array(nr_pages, sizeof(struct page *),
nr_pages           82 drivers/gpu/drm/exynos/exynos_drm_gem.c 					     nr_pages)) {
nr_pages          377 drivers/gpu/drm/savage/savage_bci.c 	unsigned int nr_pages = (n - rest + SAVAGE_DMA_PAGE_SIZE - 1) /
nr_pages          383 drivers/gpu/drm/savage/savage_bci.c 		  cur, dev_priv->dma_pages[cur].used, n, rest, nr_pages);
nr_pages          385 drivers/gpu/drm/savage/savage_bci.c 	if (cur + nr_pages < dev_priv->nr_dma_pages) {
nr_pages          395 drivers/gpu/drm/savage/savage_bci.c 		nr_pages =
nr_pages          405 drivers/gpu/drm/savage/savage_bci.c 	for (i = cur; nr_pages > 0; ++i, --nr_pages) {
nr_pages          208 drivers/gpu/drm/virtio/virtgpu_object.c 	int nr_pages = bo->tbo.num_pages;
nr_pages          229 drivers/gpu/drm/virtio/virtgpu_object.c 	ret = __sg_alloc_table_from_pages(bo->pages, pages, nr_pages, 0,
nr_pages          230 drivers/gpu/drm/virtio/virtgpu_object.c 					  nr_pages << PAGE_SHIFT,
nr_pages          591 drivers/hv/hv_balloon.c 					   unsigned long nr_pages)
nr_pages          597 drivers/hv/hv_balloon.c 	while (pfn < start_pfn + nr_pages) {
nr_pages          606 drivers/hv/hv_balloon.c 			       (pfn < start_pfn + nr_pages)) {
nr_pages          644 drivers/hv/hv_balloon.c 						  mem->nr_pages);
nr_pages         1091 drivers/hv/hv_balloon.c 	unsigned long nr_pages = totalram_pages();
nr_pages         1104 drivers/hv/hv_balloon.c 	if (nr_pages < MB2PAGES(128))
nr_pages         1105 drivers/hv/hv_balloon.c 		min_pages = MB2PAGES(8) + (nr_pages >> 1);
nr_pages         1106 drivers/hv/hv_balloon.c 	else if (nr_pages < MB2PAGES(512))
nr_pages         1107 drivers/hv/hv_balloon.c 		min_pages = MB2PAGES(40) + (nr_pages >> 2);
nr_pages         1108 drivers/hv/hv_balloon.c 	else if (nr_pages < MB2PAGES(2048))
nr_pages         1109 drivers/hv/hv_balloon.c 		min_pages = MB2PAGES(104) + (nr_pages >> 3);
nr_pages         1110 drivers/hv/hv_balloon.c 	else if (nr_pages < MB2PAGES(8192))
nr_pages         1111 drivers/hv/hv_balloon.c 		min_pages = MB2PAGES(232) + (nr_pages >> 4);
nr_pages         1113 drivers/hv/hv_balloon.c 		min_pages = MB2PAGES(488) + (nr_pages >> 5);
nr_pages          375 drivers/hwtracing/coresight/coresight-etb10.c 			      int nr_pages, bool overwrite)
nr_pages          387 drivers/hwtracing/coresight/coresight-etb10.c 	buf->nr_pages = nr_pages;
nr_pages          411 drivers/hwtracing/coresight/coresight-etb10.c 	head = handle->head & ((buf->nr_pages << PAGE_SHIFT) - 1);
nr_pages          547 drivers/hwtracing/coresight/coresight-etb10.c 			cur &= buf->nr_pages - 1;
nr_pages          208 drivers/hwtracing/coresight/coresight-etm-perf.c 			   int nr_pages, bool overwrite)
nr_pages          281 drivers/hwtracing/coresight/coresight-etm-perf.c 						     nr_pages, overwrite);
nr_pages           96 drivers/hwtracing/coresight/coresight-priv.h 	unsigned int		nr_pages;
nr_pages          390 drivers/hwtracing/coresight/coresight-tmc-etf.c 				  int nr_pages, bool overwrite)
nr_pages          403 drivers/hwtracing/coresight/coresight-tmc-etf.c 	buf->nr_pages = nr_pages;
nr_pages          427 drivers/hwtracing/coresight/coresight-tmc-etf.c 	head = handle->head & ((buf->nr_pages << PAGE_SHIFT) - 1);
nr_pages          539 drivers/hwtracing/coresight/coresight-tmc-etf.c 			cur &= buf->nr_pages - 1;
nr_pages           45 drivers/hwtracing/coresight/coresight-tmc-etr.c 	int			nr_pages;
nr_pages           50 drivers/hwtracing/coresight/coresight-tmc-etr.c #define PERF_IDX2OFF(idx, buf)	((idx) % ((buf)->nr_pages << PAGE_SHIFT))
nr_pages          122 drivers/hwtracing/coresight/coresight-tmc-etr.c tmc_etr_sg_table_entries(int nr_pages)
nr_pages          124 drivers/hwtracing/coresight/coresight-tmc-etr.c 	unsigned long nr_sgpages = nr_pages * ETR_SG_PAGES_PER_SYSPAGE;
nr_pages          147 drivers/hwtracing/coresight/coresight-tmc-etr.c 	for (i = 0; i < tmc_pages->nr_pages; i++) {
nr_pages          167 drivers/hwtracing/coresight/coresight-tmc-etr.c 	for (i = 0; i < tmc_pages->nr_pages; i++) {
nr_pages          179 drivers/hwtracing/coresight/coresight-tmc-etr.c 	tmc_pages->nr_pages = 0;
nr_pages          194 drivers/hwtracing/coresight/coresight-tmc-etr.c 	int i, nr_pages;
nr_pages          199 drivers/hwtracing/coresight/coresight-tmc-etr.c 	nr_pages = tmc_pages->nr_pages;
nr_pages          200 drivers/hwtracing/coresight/coresight-tmc-etr.c 	tmc_pages->daddrs = kcalloc(nr_pages, sizeof(*tmc_pages->daddrs),
nr_pages          204 drivers/hwtracing/coresight/coresight-tmc-etr.c 	tmc_pages->pages = kcalloc(nr_pages, sizeof(*tmc_pages->pages),
nr_pages          212 drivers/hwtracing/coresight/coresight-tmc-etr.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          275 drivers/hwtracing/coresight/coresight-tmc-etr.c 				     table_pages->nr_pages,
nr_pages          295 drivers/hwtracing/coresight/coresight-tmc-etr.c 					    sg_table->data_pages.nr_pages,
nr_pages          327 drivers/hwtracing/coresight/coresight-tmc-etr.c 	sg_table->data_pages.nr_pages = nr_dpages;
nr_pages          328 drivers/hwtracing/coresight/coresight-tmc-etr.c 	sg_table->table_pages.nr_pages = nr_tpages;
nr_pages          358 drivers/hwtracing/coresight/coresight-tmc-etr.c 		index = i % data->nr_pages;
nr_pages          371 drivers/hwtracing/coresight/coresight-tmc-etr.c 	for (i = 0; i < table_pages->nr_pages; i++)
nr_pages          494 drivers/hwtracing/coresight/coresight-tmc-etr.c 	nr_entries = tmc_etr_sg_table_entries(sg_table->data_pages.nr_pages);
nr_pages         1204 drivers/hwtracing/coresight/coresight-tmc-etr.c 	      int nr_pages, void **pages, bool snapshot)
nr_pages         1215 drivers/hwtracing/coresight/coresight-tmc-etr.c 	if ((nr_pages << PAGE_SHIFT) > drvdata->size) {
nr_pages         1216 drivers/hwtracing/coresight/coresight-tmc-etr.c 		etr_buf = tmc_alloc_etr_buf(drvdata, (nr_pages << PAGE_SHIFT),
nr_pages         1242 drivers/hwtracing/coresight/coresight-tmc-etr.c 			  struct perf_event *event, int nr_pages,
nr_pages         1279 drivers/hwtracing/coresight/coresight-tmc-etr.c 	etr_buf = alloc_etr_buf(drvdata, event, nr_pages, pages, snapshot);
nr_pages         1306 drivers/hwtracing/coresight/coresight-tmc-etr.c 			    struct perf_event *event, int nr_pages,
nr_pages         1313 drivers/hwtracing/coresight/coresight-tmc-etr.c 	return alloc_etr_buf(drvdata, event, nr_pages, pages, snapshot);
nr_pages         1318 drivers/hwtracing/coresight/coresight-tmc-etr.c 		 int nr_pages, void **pages, bool snapshot)
nr_pages         1321 drivers/hwtracing/coresight/coresight-tmc-etr.c 		return get_perf_etr_buf_per_thread(drvdata, event, nr_pages,
nr_pages         1324 drivers/hwtracing/coresight/coresight-tmc-etr.c 	return get_perf_etr_buf_cpu_wide(drvdata, event, nr_pages,
nr_pages         1330 drivers/hwtracing/coresight/coresight-tmc-etr.c 		       int nr_pages, void **pages, bool snapshot)
nr_pages         1342 drivers/hwtracing/coresight/coresight-tmc-etr.c 	etr_buf = get_perf_etr_buf(drvdata, event, nr_pages, pages, snapshot);
nr_pages         1363 drivers/hwtracing/coresight/coresight-tmc-etr.c 				  int nr_pages, bool snapshot)
nr_pages         1369 drivers/hwtracing/coresight/coresight-tmc-etr.c 					  nr_pages, pages, snapshot);
nr_pages         1377 drivers/hwtracing/coresight/coresight-tmc-etr.c 	etr_perf->nr_pages = nr_pages;
nr_pages         1460 drivers/hwtracing/coresight/coresight-tmc-etr.c 			if (++pg_idx == etr_perf->nr_pages)
nr_pages          228 drivers/hwtracing/coresight/coresight-tmc.h 	int nr_pages;
nr_pages          323 drivers/hwtracing/coresight/coresight-tmc.h 	return sg_table->data_pages.nr_pages << PAGE_SHIFT;
nr_pages          141 drivers/hwtracing/intel_th/msu.c 	unsigned long		nr_pages;
nr_pages          780 drivers/hwtracing/intel_th/msu.c 		reg = msc->nr_pages;
nr_pages          835 drivers/hwtracing/intel_th/msu.c 		msc->single_sz = reg & ((msc->nr_pages << PAGE_SHIFT) - 1);
nr_pages          910 drivers/hwtracing/intel_th/msu.c 	unsigned long nr_pages = size >> PAGE_SHIFT;
nr_pages          935 drivers/hwtracing/intel_th/msu.c 	msc->nr_pages = nr_pages;
nr_pages          963 drivers/hwtracing/intel_th/msu.c 	for (off = 0; off < msc->nr_pages << PAGE_SHIFT; off += PAGE_SIZE) {
nr_pages          970 drivers/hwtracing/intel_th/msu.c 	msc->nr_pages = 0;
nr_pages          983 drivers/hwtracing/intel_th/msu.c 	if (pgoff >= msc->nr_pages)
nr_pages         1108 drivers/hwtracing/intel_th/msu.c 	msc->nr_pages += nr_blocks;
nr_pages         1143 drivers/hwtracing/intel_th/msu.c 	msc->nr_pages -= win->nr_blocks;
nr_pages         1230 drivers/hwtracing/intel_th/msu.c static int msc_buffer_multi_alloc(struct msc *msc, unsigned long *nr_pages,
nr_pages         1236 drivers/hwtracing/intel_th/msu.c 		ret = msc_buffer_win_alloc(msc, nr_pages[i]);
nr_pages         1281 drivers/hwtracing/intel_th/msu.c static int msc_buffer_alloc(struct msc *msc, unsigned long *nr_pages,
nr_pages         1294 drivers/hwtracing/intel_th/msu.c 		ret = msc_buffer_contig_alloc(msc, nr_pages[0] << PAGE_SHIFT);
nr_pages         1296 drivers/hwtracing/intel_th/msu.c 		ret = msc_buffer_multi_alloc(msc, nr_pages, nr_wins);
nr_pages         1461 drivers/hwtracing/intel_th/msu.c 	unsigned long size = msc->nr_pages << PAGE_SHIFT, rem = len;
nr_pages         1509 drivers/hwtracing/intel_th/msu.c 		size = msc->nr_pages << PAGE_SHIFT;
nr_pages         1565 drivers/hwtracing/intel_th/msu.c 	for (pg = 0; pg < msc->nr_pages; pg++) {
nr_pages         1623 drivers/hwtracing/intel_th/msu.c 	if (size >> PAGE_SHIFT != msc->nr_pages)
nr_pages         1931 drivers/hwtracing/intel_th/msu.c 		count = scnprintf(buf, PAGE_SIZE, "%ld\n", msc->nr_pages);
nr_pages         2017 drivers/hwtracing/intel_th/msu.c static DEVICE_ATTR_RW(nr_pages);
nr_pages         1032 drivers/iommu/dma-iommu.c 	unsigned long nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT;
nr_pages         1041 drivers/iommu/dma-iommu.c 	if (off >= nr_pages || vma_pages(vma) > nr_pages - off)
nr_pages         2225 drivers/iommu/intel-iommu.c 			    unsigned long nr_pages, int prot)
nr_pages         2233 drivers/iommu/intel-iommu.c 	BUG_ON(!domain_pfn_supported(domain, iov_pfn + nr_pages - 1));
nr_pages         2241 drivers/iommu/intel-iommu.c 		sg_res = nr_pages;
nr_pages         2245 drivers/iommu/intel-iommu.c 	while (nr_pages > 0) {
nr_pages         2304 drivers/iommu/intel-iommu.c 		BUG_ON(nr_pages < lvl_pages);
nr_pages         2307 drivers/iommu/intel-iommu.c 		nr_pages -= lvl_pages;
nr_pages         2325 drivers/iommu/intel-iommu.c 		if (!nr_pages || first_pte_in_page(pte) ||
nr_pages         2332 drivers/iommu/intel-iommu.c 		if (!sg_res && nr_pages)
nr_pages         2340 drivers/iommu/intel-iommu.c 			  unsigned long nr_pages, int prot)
nr_pages         2346 drivers/iommu/intel-iommu.c 	ret = __domain_mapping(domain, iov_pfn, sg, phys_pfn, nr_pages, prot);
nr_pages         2352 drivers/iommu/intel-iommu.c 		__mapping_notify_one(iommu, domain, iov_pfn, nr_pages);
nr_pages         2359 drivers/iommu/intel-iommu.c 				    struct scatterlist *sg, unsigned long nr_pages,
nr_pages         2362 drivers/iommu/intel-iommu.c 	return domain_mapping(domain, iov_pfn, sg, 0, nr_pages, prot);
nr_pages         2366 drivers/iommu/intel-iommu.c 				     unsigned long phys_pfn, unsigned long nr_pages,
nr_pages         2369 drivers/iommu/intel-iommu.c 	return domain_mapping(domain, iov_pfn, NULL, phys_pfn, nr_pages, prot);
nr_pages         4659 drivers/iommu/intel-iommu.c 		end = ((mhp->start_pfn + mhp->nr_pages) << PAGE_SHIFT) - 1;
nr_pages         4670 drivers/iommu/intel-iommu.c 		last_vpfn = mm_to_dma_pfn(mhp->start_pfn + mhp->nr_pages - 1);
nr_pages          216 drivers/iommu/s390-iommu.c 	unsigned long irq_flags, nr_pages, i;
nr_pages          224 drivers/iommu/s390-iommu.c 	nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT;
nr_pages          225 drivers/iommu/s390-iommu.c 	if (!nr_pages)
nr_pages          229 drivers/iommu/s390-iommu.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          243 drivers/iommu/s390-iommu.c 					start_dma_addr, nr_pages * PAGE_SIZE);
nr_pages          324 drivers/lightnvm/pblk-core.c 			 int nr_pages)
nr_pages          340 drivers/lightnvm/pblk-core.c 		       int nr_pages)
nr_pages          346 drivers/lightnvm/pblk-core.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          823 drivers/lightnvm/pblk.h 		       int nr_pages);
nr_pages          825 drivers/lightnvm/pblk.h 			 int nr_pages);
nr_pages          306 drivers/md/dm-kcopyd.c static int client_reserve_pages(struct dm_kcopyd_client *kc, unsigned nr_pages)
nr_pages          311 drivers/md/dm-kcopyd.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          322 drivers/md/dm-kcopyd.c 	kc->nr_reserved_pages += nr_pages;
nr_pages          587 drivers/md/dm-kcopyd.c 	unsigned nr_pages = dm_div_up(job->dests[0].count, PAGE_SIZE >> 9);
nr_pages          589 drivers/md/dm-kcopyd.c 	r = kcopyd_get_pages(job->kc, nr_pages, &job->pages);
nr_pages          167 drivers/md/dm-linear.c 		long nr_pages, void **kaddr, pfn_t *pfn)
nr_pages          176 drivers/md/dm-linear.c 	ret = bdev_dax_pgoff(bdev, dev_sector, nr_pages * PAGE_SIZE, &pgoff);
nr_pages          179 drivers/md/dm-linear.c 	return dax_direct_access(dax_dev, pgoff, nr_pages, kaddr, pfn);
nr_pages          949 drivers/md/dm-log-writes.c 					 long nr_pages, void **kaddr, pfn_t *pfn)
nr_pages          955 drivers/md/dm-log-writes.c 	ret = bdev_dax_pgoff(lc->dev->bdev, sector, nr_pages * PAGE_SIZE, &pgoff);
nr_pages          958 drivers/md/dm-log-writes.c 	return dax_direct_access(lc->dev->dax_dev, pgoff, nr_pages, kaddr, pfn);
nr_pages          318 drivers/md/dm-stripe.c 		long nr_pages, void **kaddr, pfn_t *pfn)
nr_pages          332 drivers/md/dm-stripe.c 	ret = bdev_dax_pgoff(bdev, dev_sector, nr_pages * PAGE_SIZE, &pgoff);
nr_pages          335 drivers/md/dm-stripe.c 	return dax_direct_access(dax_dev, pgoff, nr_pages, kaddr, pfn);
nr_pages          145 drivers/md/dm-target.c 		long nr_pages, void **kaddr, pfn_t *pfn)
nr_pages         1082 drivers/md/dm.c 				 long nr_pages, void **kaddr, pfn_t *pfn)
nr_pages         1099 drivers/md/dm.c 	nr_pages = min(len, nr_pages);
nr_pages         1100 drivers/md/dm.c 	ret = ti->type->direct_access(ti, pgoff, nr_pages, kaddr, pfn);
nr_pages          137 drivers/media/common/saa7146/saa7146_core.c static struct scatterlist* vmalloc_to_sg(unsigned char *virt, int nr_pages)
nr_pages          143 drivers/media/common/saa7146/saa7146_core.c 	sglist = kcalloc(nr_pages, sizeof(struct scatterlist), GFP_KERNEL);
nr_pages          146 drivers/media/common/saa7146/saa7146_core.c 	sg_init_table(sglist, nr_pages);
nr_pages          147 drivers/media/common/saa7146/saa7146_core.c 	for (i = 0; i < nr_pages; i++, virt += PAGE_SIZE) {
nr_pages          239 drivers/media/common/saa7146/saa7146_core.c 	int nr_pages = 0;
nr_pages          258 drivers/media/common/saa7146/saa7146_core.c 			nr_pages++;
nr_pages          265 drivers/media/common/saa7146/saa7146_core.c 	for(i=nr_pages;i<1024;i++) {
nr_pages           71 drivers/media/pci/cx23885/cx23885-alsa.c static int cx23885_alsa_dma_init(struct cx23885_audio_dev *chip, int nr_pages)
nr_pages           77 drivers/media/pci/cx23885/cx23885-alsa.c 	buf->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT);
nr_pages           79 drivers/media/pci/cx23885/cx23885-alsa.c 		dprintk(1, "vmalloc_32(%d pages) failed\n", nr_pages);
nr_pages           84 drivers/media/pci/cx23885/cx23885-alsa.c 		buf->vaddr, nr_pages << PAGE_SHIFT);
nr_pages           86 drivers/media/pci/cx23885/cx23885-alsa.c 	memset(buf->vaddr, 0, nr_pages << PAGE_SHIFT);
nr_pages           87 drivers/media/pci/cx23885/cx23885-alsa.c 	buf->nr_pages = nr_pages;
nr_pages           89 drivers/media/pci/cx23885/cx23885-alsa.c 	buf->sglist = vzalloc(array_size(sizeof(*buf->sglist), buf->nr_pages));
nr_pages           93 drivers/media/pci/cx23885/cx23885-alsa.c 	sg_init_table(buf->sglist, buf->nr_pages);
nr_pages           94 drivers/media/pci/cx23885/cx23885-alsa.c 	for (i = 0; i < buf->nr_pages; i++) {
nr_pages          116 drivers/media/pci/cx23885/cx23885-alsa.c 			buf->nr_pages, PCI_DMA_FROMDEVICE);
nr_pages          329 drivers/media/pci/cx23885/cx23885.h 	int                     nr_pages;
nr_pages           57 drivers/media/pci/cx25821/cx25821-alsa.c 	int                     nr_pages;
nr_pages          134 drivers/media/pci/cx25821/cx25821-alsa.c static int cx25821_alsa_dma_init(struct cx25821_audio_dev *chip, int nr_pages)
nr_pages          140 drivers/media/pci/cx25821/cx25821-alsa.c 	buf->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT);
nr_pages          142 drivers/media/pci/cx25821/cx25821-alsa.c 		dprintk(1, "vmalloc_32(%d pages) failed\n", nr_pages);
nr_pages          148 drivers/media/pci/cx25821/cx25821-alsa.c 				nr_pages << PAGE_SHIFT);
nr_pages          150 drivers/media/pci/cx25821/cx25821-alsa.c 	memset(buf->vaddr, 0, nr_pages << PAGE_SHIFT);
nr_pages          151 drivers/media/pci/cx25821/cx25821-alsa.c 	buf->nr_pages = nr_pages;
nr_pages          153 drivers/media/pci/cx25821/cx25821-alsa.c 	buf->sglist = vzalloc(array_size(sizeof(*buf->sglist), buf->nr_pages));
nr_pages          157 drivers/media/pci/cx25821/cx25821-alsa.c 	sg_init_table(buf->sglist, buf->nr_pages);
nr_pages          158 drivers/media/pci/cx25821/cx25821-alsa.c 	for (i = 0; i < buf->nr_pages; i++) {
nr_pages          180 drivers/media/pci/cx25821/cx25821-alsa.c 			buf->nr_pages, PCI_DMA_FROMDEVICE);
nr_pages           50 drivers/media/pci/cx88/cx88-alsa.c 	int                     nr_pages;
nr_pages          274 drivers/media/pci/cx88/cx88-alsa.c static int cx88_alsa_dma_init(struct cx88_audio_dev *chip, int nr_pages)
nr_pages          280 drivers/media/pci/cx88/cx88-alsa.c 	buf->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT);
nr_pages          282 drivers/media/pci/cx88/cx88-alsa.c 		dprintk(1, "vmalloc_32(%d pages) failed\n", nr_pages);
nr_pages          287 drivers/media/pci/cx88/cx88-alsa.c 		buf->vaddr, nr_pages << PAGE_SHIFT);
nr_pages          289 drivers/media/pci/cx88/cx88-alsa.c 	memset(buf->vaddr, 0, nr_pages << PAGE_SHIFT);
nr_pages          290 drivers/media/pci/cx88/cx88-alsa.c 	buf->nr_pages = nr_pages;
nr_pages          292 drivers/media/pci/cx88/cx88-alsa.c 	buf->sglist = vzalloc(array_size(sizeof(*buf->sglist), buf->nr_pages));
nr_pages          296 drivers/media/pci/cx88/cx88-alsa.c 	sg_init_table(buf->sglist, buf->nr_pages);
nr_pages          297 drivers/media/pci/cx88/cx88-alsa.c 	for (i = 0; i < buf->nr_pages; i++) {
nr_pages          319 drivers/media/pci/cx88/cx88-alsa.c 			buf->nr_pages, PCI_DMA_FROMDEVICE);
nr_pages          255 drivers/media/pci/saa7134/saa7134-alsa.c static int saa7134_alsa_dma_init(struct saa7134_dev *dev, int nr_pages)
nr_pages          261 drivers/media/pci/saa7134/saa7134-alsa.c 	dma->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT);
nr_pages          263 drivers/media/pci/saa7134/saa7134-alsa.c 		pr_debug("vmalloc_32(%d pages) failed\n", nr_pages);
nr_pages          268 drivers/media/pci/saa7134/saa7134-alsa.c 		 dma->vaddr, nr_pages << PAGE_SHIFT);
nr_pages          270 drivers/media/pci/saa7134/saa7134-alsa.c 	memset(dma->vaddr, 0, nr_pages << PAGE_SHIFT);
nr_pages          271 drivers/media/pci/saa7134/saa7134-alsa.c 	dma->nr_pages = nr_pages;
nr_pages          273 drivers/media/pci/saa7134/saa7134-alsa.c 	dma->sglist = vzalloc(array_size(sizeof(*dma->sglist), dma->nr_pages));
nr_pages          277 drivers/media/pci/saa7134/saa7134-alsa.c 	sg_init_table(dma->sglist, dma->nr_pages);
nr_pages          278 drivers/media/pci/saa7134/saa7134-alsa.c 	for (i = 0; i < dma->nr_pages; i++) {
nr_pages          300 drivers/media/pci/saa7134/saa7134-alsa.c 			dma->nr_pages, PCI_DMA_FROMDEVICE);
nr_pages          512 drivers/media/pci/saa7134/saa7134.h 	int                        nr_pages;
nr_pages           63 drivers/media/v4l2-core/videobuf-dma-sg.c 						  int nr_pages)
nr_pages           69 drivers/media/v4l2-core/videobuf-dma-sg.c 	sglist = vzalloc(array_size(nr_pages, sizeof(*sglist)));
nr_pages           72 drivers/media/v4l2-core/videobuf-dma-sg.c 	sg_init_table(sglist, nr_pages);
nr_pages           73 drivers/media/v4l2-core/videobuf-dma-sg.c 	for (i = 0; i < nr_pages; i++, virt += PAGE_SIZE) {
nr_pages           93 drivers/media/v4l2-core/videobuf-dma-sg.c 					int nr_pages, int offset, size_t size)
nr_pages          100 drivers/media/v4l2-core/videobuf-dma-sg.c 	sglist = vmalloc(array_size(nr_pages, sizeof(*sglist)));
nr_pages          103 drivers/media/v4l2-core/videobuf-dma-sg.c 	sg_init_table(sglist, nr_pages);
nr_pages          111 drivers/media/v4l2-core/videobuf-dma-sg.c 	for (i = 1; i < nr_pages; i++) {
nr_pages          174 drivers/media/v4l2-core/videobuf-dma-sg.c 	dma->nr_pages = last-first+1;
nr_pages          175 drivers/media/v4l2-core/videobuf-dma-sg.c 	dma->pages = kmalloc_array(dma->nr_pages, sizeof(struct page *),
nr_pages          184 drivers/media/v4l2-core/videobuf-dma-sg.c 		data, size, dma->nr_pages);
nr_pages          186 drivers/media/v4l2-core/videobuf-dma-sg.c 	err = get_user_pages(data & PAGE_MASK, dma->nr_pages,
nr_pages          189 drivers/media/v4l2-core/videobuf-dma-sg.c 	if (err != dma->nr_pages) {
nr_pages          190 drivers/media/v4l2-core/videobuf-dma-sg.c 		dma->nr_pages = (err >= 0) ? err : 0;
nr_pages          192 drivers/media/v4l2-core/videobuf-dma-sg.c 			dma->nr_pages);
nr_pages          211 drivers/media/v4l2-core/videobuf-dma-sg.c 			     int nr_pages)
nr_pages          215 drivers/media/v4l2-core/videobuf-dma-sg.c 	dprintk(1, "init kernel [%d pages]\n", nr_pages);
nr_pages          218 drivers/media/v4l2-core/videobuf-dma-sg.c 	dma->vaddr_pages = kcalloc(nr_pages, sizeof(*dma->vaddr_pages),
nr_pages          223 drivers/media/v4l2-core/videobuf-dma-sg.c 	dma->dma_addr = kcalloc(nr_pages, sizeof(*dma->dma_addr), GFP_KERNEL);
nr_pages          228 drivers/media/v4l2-core/videobuf-dma-sg.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          238 drivers/media/v4l2-core/videobuf-dma-sg.c 	dma->vaddr = vmap(dma->vaddr_pages, nr_pages, VM_MAP | VM_IOREMAP,
nr_pages          241 drivers/media/v4l2-core/videobuf-dma-sg.c 		dprintk(1, "vmalloc_32(%d pages) failed\n", nr_pages);
nr_pages          246 drivers/media/v4l2-core/videobuf-dma-sg.c 		dma->vaddr, nr_pages << PAGE_SHIFT);
nr_pages          248 drivers/media/v4l2-core/videobuf-dma-sg.c 	memset(dma->vaddr, 0, nr_pages << PAGE_SHIFT);
nr_pages          249 drivers/media/v4l2-core/videobuf-dma-sg.c 	dma->nr_pages = nr_pages;
nr_pages          270 drivers/media/v4l2-core/videobuf-dma-sg.c 			      dma_addr_t addr, int nr_pages)
nr_pages          273 drivers/media/v4l2-core/videobuf-dma-sg.c 		nr_pages, (unsigned long)addr);
nr_pages          280 drivers/media/v4l2-core/videobuf-dma-sg.c 	dma->nr_pages = nr_pages;
nr_pages          288 drivers/media/v4l2-core/videobuf-dma-sg.c 	BUG_ON(0 == dma->nr_pages);
nr_pages          291 drivers/media/v4l2-core/videobuf-dma-sg.c 		dma->sglist = videobuf_pages_to_sg(dma->pages, dma->nr_pages,
nr_pages          296 drivers/media/v4l2-core/videobuf-dma-sg.c 						     dma->nr_pages);
nr_pages          305 drivers/media/v4l2-core/videobuf-dma-sg.c 			sg_dma_len(&dma->sglist[0]) = dma->nr_pages * PAGE_SIZE;
nr_pages          314 drivers/media/v4l2-core/videobuf-dma-sg.c 					dma->nr_pages, dma->direction);
nr_pages          335 drivers/media/v4l2-core/videobuf-dma-sg.c 	dma_unmap_sg(dev, dma->sglist, dma->nr_pages, dma->direction);
nr_pages          352 drivers/media/v4l2-core/videobuf-dma-sg.c 		for (i = 0; i < dma->nr_pages; i++) {
nr_pages          362 drivers/media/v4l2-core/videobuf-dma-sg.c 		for (i = 0; i < dma->nr_pages; i++) {
nr_pages          585 drivers/media/v4l2-core/videobuf-dma-sg.c 			    mem->dma.nr_pages, mem->dma.direction);
nr_pages          172 drivers/misc/genwqe/card_base.h 	unsigned int nr_pages;		/* number of pages */
nr_pages          356 drivers/misc/genwqe/card_base.h 	unsigned long nr_pages;
nr_pages          450 drivers/misc/genwqe/card_dev.c 	dma_map->nr_pages = DIV_ROUND_UP(vsize, PAGE_SIZE);
nr_pages          296 drivers/misc/genwqe/card_utils.c 	sgl->nr_pages = DIV_ROUND_UP(sgl->fpage_offs + user_size, PAGE_SIZE);
nr_pages          300 drivers/misc/genwqe/card_utils.c 		__func__, user_addr, user_size, sgl->nr_pages,
nr_pages          306 drivers/misc/genwqe/card_utils.c 	sgl->sgl_size = genwqe_sgl_size(sgl->nr_pages);
nr_pages          385 drivers/misc/genwqe/card_utils.c 	while (p < sgl->nr_pages) {
nr_pages          403 drivers/misc/genwqe/card_utils.c 			} else if ((p == sgl->nr_pages - 1) &&
nr_pages          424 drivers/misc/genwqe/card_utils.c 				if (p == sgl->nr_pages)
nr_pages          440 drivers/misc/genwqe/card_utils.c 			if (p == sgl->nr_pages)
nr_pages          527 drivers/misc/genwqe/card_utils.c 			unsigned int nr_pages, int dirty)
nr_pages          531 drivers/misc/genwqe/card_utils.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          585 drivers/misc/genwqe/card_utils.c 	m->nr_pages = DIV_ROUND_UP(offs + size, PAGE_SIZE);
nr_pages          587 drivers/misc/genwqe/card_utils.c 	m->page_list = kcalloc(m->nr_pages,
nr_pages          592 drivers/misc/genwqe/card_utils.c 		m->nr_pages = 0;
nr_pages          597 drivers/misc/genwqe/card_utils.c 	m->dma_list = (dma_addr_t *)(m->page_list + m->nr_pages);
nr_pages          601 drivers/misc/genwqe/card_utils.c 				 m->nr_pages,
nr_pages          608 drivers/misc/genwqe/card_utils.c 	if (rc < m->nr_pages) {
nr_pages          614 drivers/misc/genwqe/card_utils.c 	rc = genwqe_map_pages(cd, m->page_list, m->nr_pages, m->dma_list);
nr_pages          621 drivers/misc/genwqe/card_utils.c 	genwqe_free_user_pages(m->page_list, m->nr_pages, m->write);
nr_pages          627 drivers/misc/genwqe/card_utils.c 	m->nr_pages = 0;
nr_pages          650 drivers/misc/genwqe/card_utils.c 		genwqe_unmap_pages(cd, m->dma_list, m->nr_pages);
nr_pages          653 drivers/misc/genwqe/card_utils.c 		genwqe_free_user_pages(m->page_list, m->nr_pages, m->write);
nr_pages          658 drivers/misc/genwqe/card_utils.c 		m->nr_pages = 0;
nr_pages           49 drivers/misc/mic/scif/scif_debugfs.c 		   window->nr_pages, window->nr_contig_chunks, window->prot);
nr_pages           60 drivers/misc/mic/scif/scif_debugfs.c 		for (j = 0; j < window->nr_pages; j++)
nr_pages          124 drivers/misc/mic/scif/scif_dma.c 		end_va = start_va + (window->nr_pages << PAGE_SHIFT);
nr_pages          376 drivers/misc/mic/scif/scif_dma.c 	(*out_window)->nr_pages = pinned_pages->nr_pages;
nr_pages          631 drivers/misc/mic/scif/scif_dma.c 			atomic_sub(window->nr_pages,
nr_pages          754 drivers/misc/mic/scif/scif_dma.c 	if (window->nr_pages == window->nr_contig_chunks) {
nr_pages          818 drivers/misc/mic/scif/scif_dma.c 		(window->nr_pages << PAGE_SHIFT);
nr_pages          823 drivers/misc/mic/scif/scif_dma.c 				(window->nr_pages << PAGE_SHIFT);
nr_pages          921 drivers/misc/mic/scif/scif_dma.c 		(window->nr_pages << PAGE_SHIFT);
nr_pages          929 drivers/misc/mic/scif/scif_dma.c 				(window->nr_pages << PAGE_SHIFT);
nr_pages         1034 drivers/misc/mic/scif/scif_dma.c 				(window->nr_pages << PAGE_SHIFT);
nr_pages         1110 drivers/misc/mic/scif/scif_dma.c 		(src_window->nr_pages << PAGE_SHIFT);
nr_pages         1112 drivers/misc/mic/scif/scif_dma.c 		(dst_window->nr_pages << PAGE_SHIFT);
nr_pages         1117 drivers/misc/mic/scif/scif_dma.c 				(src_window->nr_pages << PAGE_SHIFT);
nr_pages         1123 drivers/misc/mic/scif/scif_dma.c 				(dst_window->nr_pages << PAGE_SHIFT);
nr_pages         1275 drivers/misc/mic/scif/scif_dma.c 		(src_window->nr_pages << PAGE_SHIFT);
nr_pages         1277 drivers/misc/mic/scif/scif_dma.c 		(dst_window->nr_pages << PAGE_SHIFT);
nr_pages         1284 drivers/misc/mic/scif/scif_dma.c 				(src_window->nr_pages << PAGE_SHIFT);
nr_pages         1290 drivers/misc/mic/scif/scif_dma.c 				(dst_window->nr_pages << PAGE_SHIFT);
nr_pages         1514 drivers/misc/mic/scif/scif_dma.c 				(src_window->nr_pages << PAGE_SHIFT);
nr_pages         1516 drivers/misc/mic/scif/scif_dma.c 				(dst_window->nr_pages << PAGE_SHIFT);
nr_pages         1727 drivers/misc/mic/scif/scif_dma.c 			atomic_add_return(local_window->nr_pages,
nr_pages           34 drivers/misc/mic/scif/scif_mmap.c 	req.nr_bytes = recv_window->nr_pages << PAGE_SHIFT;
nr_pages           48 drivers/misc/mic/scif/scif_mmap.c 	scif_put_window(window, window->nr_pages);
nr_pages          130 drivers/misc/mic/scif/scif_mmap.c 			scif_put_window(window, window->nr_pages);
nr_pages          218 drivers/misc/mic/scif/scif_mmap.c 	int nr_pages, err, i;
nr_pages          233 drivers/misc/mic/scif/scif_mmap.c 	nr_pages = len >> PAGE_SHIFT;
nr_pages          259 drivers/misc/mic/scif/scif_mmap.c 	(*pages)->phys_addr = scif_zalloc(nr_pages * sizeof(dma_addr_t));
nr_pages          267 drivers/misc/mic/scif/scif_mmap.c 		((*pages)->va = scif_zalloc(nr_pages * sizeof(void *)));
nr_pages          275 drivers/misc/mic/scif/scif_mmap.c 	(*pages)->nr_pages = nr_pages;
nr_pages          278 drivers/misc/mic/scif/scif_mmap.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          291 drivers/misc/mic/scif/scif_mmap.c 	scif_get_window(window, nr_pages);
nr_pages          297 drivers/misc/mic/scif/scif_mmap.c 				  nr_pages * sizeof(dma_addr_t));
nr_pages          299 drivers/misc/mic/scif/scif_mmap.c 				  nr_pages * sizeof(void *));
nr_pages          336 drivers/misc/mic/scif/scif_mmap.c 	scif_put_window(window, pages->nr_pages);
nr_pages          356 drivers/misc/mic/scif/scif_mmap.c 	scif_free(pages->phys_addr, pages->nr_pages * sizeof(dma_addr_t));
nr_pages          357 drivers/misc/mic/scif/scif_mmap.c 	scif_free(pages->va, pages->nr_pages * sizeof(void *));
nr_pages          373 drivers/misc/mic/scif/scif_mmap.c 			      int nr_pages, struct vm_area_struct *vma)
nr_pages          377 drivers/misc/mic/scif/scif_mmap.c 	int loop_nr_pages, nr_pages_left = nr_pages;
nr_pages          388 drivers/misc/mic/scif/scif_mmap.c 			(window->nr_pages << PAGE_SHIFT);
nr_pages          419 drivers/misc/mic/scif/scif_mmap.c 	nr_pages_left = nr_pages;
nr_pages          424 drivers/misc/mic/scif/scif_mmap.c 			(window->nr_pages << PAGE_SHIFT);
nr_pages          450 drivers/misc/mic/scif/scif_mmap.c 				 s64 offset, int nr_pages)
nr_pages          454 drivers/misc/mic/scif/scif_mmap.c 	int loop_nr_pages, nr_pages_left = nr_pages;
nr_pages          462 drivers/misc/mic/scif/scif_mmap.c 	nr_pages_left = nr_pages;
nr_pages          465 drivers/misc/mic/scif/scif_mmap.c 			(window->nr_pages << PAGE_SHIFT);
nr_pages          546 drivers/misc/mic/scif/scif_mmap.c 	int nr_pages = vma_pages(vma);
nr_pages          561 drivers/misc/mic/scif/scif_mmap.c 		ep, nr_pages, offset);
nr_pages          576 drivers/misc/mic/scif/scif_mmap.c 		scif_rma_list_munmap(window, offset, nr_pages);
nr_pages          608 drivers/misc/mic/scif/scif_mmap.c 	int nr_pages = vma_pages(vma);
nr_pages          614 drivers/misc/mic/scif/scif_mmap.c 		ep, start_offset, nr_pages);
nr_pages          672 drivers/misc/mic/scif/scif_mmap.c 	err = scif_rma_list_mmap(window, start_offset, nr_pages, vma);
nr_pages           81 drivers/misc/mic/scif/scif_rma.c scif_create_pinned_pages(int nr_pages, int prot)
nr_pages           90 drivers/misc/mic/scif/scif_rma.c 	pin->pages = scif_zalloc(nr_pages * sizeof(*pin->pages));
nr_pages          116 drivers/misc/mic/scif/scif_rma.c 	for (j = 0; j < pin->nr_pages; j++) {
nr_pages          125 drivers/misc/mic/scif/scif_rma.c 		  pin->nr_pages * sizeof(*pin->pages));
nr_pages          139 drivers/misc/mic/scif/scif_rma.c struct scif_window *scif_create_window(struct scif_endpt *ep, int nr_pages,
nr_pages          149 drivers/misc/mic/scif/scif_rma.c 	window->dma_addr = scif_zalloc(nr_pages * sizeof(*window->dma_addr));
nr_pages          153 drivers/misc/mic/scif/scif_rma.c 	window->num_pages = scif_zalloc(nr_pages * sizeof(*window->num_pages));
nr_pages          171 drivers/misc/mic/scif/scif_rma.c 		  nr_pages * sizeof(*window->dma_addr));
nr_pages          188 drivers/misc/mic/scif/scif_rma.c 	int nr_pages = window->nr_pages;
nr_pages          213 drivers/misc/mic/scif/scif_rma.c 	scif_free(window->dma_addr, nr_pages * sizeof(*window->dma_addr));
nr_pages          214 drivers/misc/mic/scif/scif_rma.c 	scif_free(window->num_pages, nr_pages * sizeof(*window->num_pages));
nr_pages          266 drivers/misc/mic/scif/scif_rma.c 			  int nr_pages)
nr_pages          268 drivers/misc/mic/scif/scif_rma.c 	if (!mm || !nr_pages || !scif_ulimit_check)
nr_pages          271 drivers/misc/mic/scif/scif_rma.c 	atomic64_sub(nr_pages, &mm->pinned_vm);
nr_pages          276 drivers/misc/mic/scif/scif_rma.c 					     int nr_pages)
nr_pages          280 drivers/misc/mic/scif/scif_rma.c 	if (!mm || !nr_pages || !scif_ulimit_check)
nr_pages          284 drivers/misc/mic/scif/scif_rma.c 	locked = atomic64_add_return(nr_pages, &mm->pinned_vm);
nr_pages          287 drivers/misc/mic/scif/scif_rma.c 		atomic64_sub(nr_pages, &mm->pinned_vm);
nr_pages          307 drivers/misc/mic/scif/scif_rma.c 	int nr_pages = window->nr_pages;
nr_pages          311 drivers/misc/mic/scif/scif_rma.c 		__scif_dec_pinned_vm_lock(window->mm, window->nr_pages);
nr_pages          334 drivers/misc/mic/scif/scif_rma.c 	scif_free(window->dma_addr, nr_pages * sizeof(*window->dma_addr));
nr_pages          335 drivers/misc/mic/scif/scif_rma.c 	scif_free(window->num_pages, nr_pages * sizeof(*window->num_pages));
nr_pages          354 drivers/misc/mic/scif/scif_rma.c 	int nr_pages = window->nr_pages;
nr_pages          365 drivers/misc/mic/scif/scif_rma.c 	window->nr_lookup = ALIGN(nr_pages * PAGE_SIZE,
nr_pages          392 drivers/misc/mic/scif/scif_rma.c 	for (i = 0, j = 0; i < nr_pages; i += SCIF_NR_ADDR_IN_PAGE, j++) {
nr_pages          430 drivers/misc/mic/scif/scif_rma.c 		for (i = 0, j = 0; i < window->nr_pages;
nr_pages          464 drivers/misc/mic/scif/scif_rma.c scif_create_remote_window(struct scif_dev *scifdev, int nr_pages)
nr_pages          474 drivers/misc/mic/scif/scif_rma.c 	window->nr_pages = nr_pages;
nr_pages          476 drivers/misc/mic/scif/scif_rma.c 	window->dma_addr = scif_zalloc(nr_pages * sizeof(*window->dma_addr));
nr_pages          480 drivers/misc/mic/scif/scif_rma.c 	window->num_pages = scif_zalloc(nr_pages *
nr_pages          508 drivers/misc/mic/scif/scif_rma.c 	scif_free(window->dma_addr, window->nr_pages *
nr_pages          510 drivers/misc/mic/scif/scif_rma.c 	scif_free(window->num_pages, window->nr_pages *
nr_pages          535 drivers/misc/mic/scif/scif_rma.c 	err = sg_alloc_table(window->st, window->nr_pages, GFP_KERNEL);
nr_pages          587 drivers/misc/mic/scif/scif_rma.c 	for (i = 0, j = 0; i < window->nr_pages; i += nr_contig_pages, j++) {
nr_pages          592 drivers/misc/mic/scif/scif_rma.c 		for (k = i + 1; k < window->nr_pages; k++) {
nr_pages          709 drivers/misc/mic/scif/scif_rma.c 		scif_put_window(window, window->nr_pages);
nr_pages          722 drivers/misc/mic/scif/scif_rma.c 						       window->nr_pages)) {
nr_pages          752 drivers/misc/mic/scif/scif_rma.c 	msg.payload[1] = window->nr_pages;
nr_pages         1045 drivers/misc/mic/scif/scif_rma.c 	int nr_pages = msg->payload[1];
nr_pages         1047 drivers/misc/mic/scif/scif_rma.c 	window = scif_create_remote_window(scifdev, nr_pages);
nr_pages         1065 drivers/misc/mic/scif/scif_rma.c 		__func__, __LINE__, err, window, nr_pages);
nr_pages         1187 drivers/misc/mic/scif/scif_rma.c 	req.nr_bytes = recv_window->nr_pages << PAGE_SHIFT;
nr_pages         1202 drivers/misc/mic/scif/scif_rma.c 			scif_put_window(window, window->nr_pages);
nr_pages         1311 drivers/misc/mic/scif/scif_rma.c 	int nr_pages, err = 0, i;
nr_pages         1335 drivers/misc/mic/scif/scif_rma.c 	nr_pages = len >> PAGE_SHIFT;
nr_pages         1338 drivers/misc/mic/scif/scif_rma.c 	pinned_pages = scif_create_pinned_pages(nr_pages, prot);
nr_pages         1346 drivers/misc/mic/scif/scif_rma.c 		for (i = 0; i < nr_pages; i++) {
nr_pages         1354 drivers/misc/mic/scif/scif_rma.c 		pinned_pages->nr_pages = nr_pages;
nr_pages         1371 drivers/misc/mic/scif/scif_rma.c 			err = __scif_check_inc_pinned_vm(mm, nr_pages);
nr_pages         1373 drivers/misc/mic/scif/scif_rma.c 				pinned_pages->nr_pages = 0;
nr_pages         1378 drivers/misc/mic/scif/scif_rma.c 		pinned_pages->nr_pages = get_user_pages_fast(
nr_pages         1380 drivers/misc/mic/scif/scif_rma.c 				nr_pages,
nr_pages         1383 drivers/misc/mic/scif/scif_rma.c 		if (nr_pages != pinned_pages->nr_pages) {
nr_pages         1386 drivers/misc/mic/scif/scif_rma.c 					__scif_dec_pinned_vm_lock(mm, nr_pages);
nr_pages         1388 drivers/misc/mic/scif/scif_rma.c 				for (i = 0; i < pinned_pages->nr_pages; i++) {
nr_pages         1401 drivers/misc/mic/scif/scif_rma.c 	if (pinned_pages->nr_pages < nr_pages) {
nr_pages         1403 drivers/misc/mic/scif/scif_rma.c 		pinned_pages->nr_pages = nr_pages;
nr_pages         1413 drivers/misc/mic/scif/scif_rma.c 		__scif_dec_pinned_vm_lock(mm, nr_pages);
nr_pages         1416 drivers/misc/mic/scif/scif_rma.c 	pinned_pages->nr_pages = nr_pages;
nr_pages         1481 drivers/misc/mic/scif/scif_rma.c 	len = pinned_pages->nr_pages << PAGE_SHIFT;
nr_pages         1514 drivers/misc/mic/scif/scif_rma.c 	window = scif_create_window(ep, pinned_pages->nr_pages,
nr_pages         1523 drivers/misc/mic/scif/scif_rma.c 	window->nr_pages = pinned_pages->nr_pages;
nr_pages         1632 drivers/misc/mic/scif/scif_rma.c 	window->nr_pages = len >> PAGE_SHIFT;
nr_pages         1698 drivers/misc/mic/scif/scif_rma.c 	int nr_pages, err;
nr_pages         1720 drivers/misc/mic/scif/scif_rma.c 	nr_pages = len >> PAGE_SHIFT;
nr_pages         1743 drivers/misc/mic/scif/scif_rma.c 	err = scif_rma_list_unregister(window, offset, nr_pages);
nr_pages          187 drivers/misc/mic/scif/scif_rma.h 	s64 nr_pages;
nr_pages          256 drivers/misc/mic/scif/scif_rma.h 	s64 nr_pages;
nr_pages          323 drivers/misc/mic/scif/scif_rma.h 			   s64 offset, int nr_pages, s64 *out_offset);
nr_pages          328 drivers/misc/mic/scif/scif_rma.h struct scif_window *scif_create_window(struct scif_endpt *ep, int nr_pages,
nr_pages          439 drivers/misc/mic/scif/scif_rma.h static inline void scif_get_window(struct scif_window *window, int nr_pages)
nr_pages          441 drivers/misc/mic/scif/scif_rma.h 	window->ref_count += nr_pages;
nr_pages          444 drivers/misc/mic/scif/scif_rma.h static inline void scif_put_window(struct scif_window *window, int nr_pages)
nr_pages          446 drivers/misc/mic/scif/scif_rma.h 	window->ref_count -= nr_pages;
nr_pages          449 drivers/misc/mic/scif/scif_rma.h static inline void scif_set_window_ref(struct scif_window *window, int nr_pages)
nr_pages          451 drivers/misc/mic/scif/scif_rma.h 	window->ref_count = nr_pages;
nr_pages           65 drivers/misc/mic/scif/scif_rma_list.c 	scif_set_window_ref(window, window->nr_pages);
nr_pages           92 drivers/misc/mic/scif/scif_rma_list.c 			(window->nr_pages << PAGE_SHIFT);
nr_pages          100 drivers/misc/mic/scif/scif_rma_list.c 			(window->nr_pages << PAGE_SHIFT);
nr_pages          148 drivers/misc/mic/scif/scif_rma_list.c 			(window->nr_pages << PAGE_SHIFT);
nr_pages          195 drivers/misc/mic/scif/scif_rma_list.c 			     s64 offset, int nr_pages)
nr_pages          205 drivers/misc/mic/scif/scif_rma_list.c 		end_offset = window->offset + (window->nr_pages << PAGE_SHIFT);
nr_pages          207 drivers/misc/mic/scif/scif_rma_list.c 				    nr_pages);
nr_pages          211 drivers/misc/mic/scif/scif_rma_list.c 		nr_pages -= loop_nr_pages;
nr_pages          213 drivers/misc/mic/scif/scif_rma_list.c 		if (!nr_pages)
nr_pages           47 drivers/misc/mic/scif/scif_rma_list.h 			     int nr_pages);
nr_pages          615 drivers/mtd/devices/mtd_dataflash.c static int add_dataflash_otp(struct spi_device *spi, char *name, int nr_pages,
nr_pages          640 drivers/mtd/devices/mtd_dataflash.c 	device->size = nr_pages * pagesize;
nr_pages          673 drivers/mtd/devices/mtd_dataflash.c 				int nr_pages, int pagesize, int pageoffset)
nr_pages          675 drivers/mtd/devices/mtd_dataflash.c 	return add_dataflash_otp(spi, name, nr_pages, pagesize,
nr_pages          688 drivers/mtd/devices/mtd_dataflash.c 	unsigned	nr_pages;
nr_pages          856 drivers/mtd/devices/mtd_dataflash.c 		return add_dataflash_otp(spi, info->name, info->nr_pages,
nr_pages           56 drivers/mtd/devices/sst25l.c 	unsigned		nr_pages;
nr_pages          377 drivers/mtd/devices/sst25l.c 	flash->mtd.size		= flash_info->page_size * flash_info->nr_pages;
nr_pages         1143 drivers/mtd/mtdswap.c 			unsigned nr_pages)
nr_pages         1152 drivers/mtd/mtdswap.c 	for (page = first; page < first + nr_pages; page++) {
nr_pages         2641 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	for (i = 0; i < rmem->nr_pages; i++) {
nr_pages         2651 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		size_t pg_tbl_size = rmem->nr_pages * 8;
nr_pages         2673 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	if ((rmem->nr_pages > 1 || rmem->depth > 0) && !rmem->pg_tbl) {
nr_pages         2674 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		size_t pg_tbl_size = rmem->nr_pages * 8;
nr_pages         2685 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	for (i = 0; i < rmem->nr_pages; i++) {
nr_pages         2695 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		if (rmem->nr_pages > 1 || rmem->depth > 0) {
nr_pages         2696 drivers/net/ethernet/broadcom/bnxt/bnxt.c 			if (i == rmem->nr_pages - 2 &&
nr_pages         2699 drivers/net/ethernet/broadcom/bnxt/bnxt.c 			else if (i == rmem->nr_pages - 1 &&
nr_pages         3011 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	rmem->nr_pages = bp->cp_nr_pages;
nr_pages         3097 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		rmem->nr_pages = bp->cp_nr_pages;
nr_pages         3109 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		rmem->nr_pages = bp->rx_nr_pages;
nr_pages         3118 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		rmem->nr_pages = bp->rx_agg_nr_pages;
nr_pages         3132 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		rmem->nr_pages = bp->tx_nr_pages;
nr_pages         3148 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	for (i = 0, prod = 0; i < ring->ring_mem.nr_pages; i++) {
nr_pages         5231 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	if (rmem->nr_pages > 1) {
nr_pages         6654 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	ctx_pg->nr_pages = DIV_ROUND_UP(mem_size, BNXT_PAGE_SIZE);
nr_pages         6655 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	if (ctx_pg->nr_pages > MAX_CTX_TOTAL_PAGES) {
nr_pages         6656 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		ctx_pg->nr_pages = 0;
nr_pages         6659 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	if (ctx_pg->nr_pages > MAX_CTX_PAGES || depth > 1) {
nr_pages         6667 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		nr_tbls = DIV_ROUND_UP(ctx_pg->nr_pages, MAX_CTX_PAGES);
nr_pages         6668 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		rmem->nr_pages = nr_tbls;
nr_pages         6683 drivers/net/ethernet/broadcom/bnxt/bnxt.c 			rmem->nr_pages = MAX_CTX_PAGES;
nr_pages         6685 drivers/net/ethernet/broadcom/bnxt/bnxt.c 				int rem = ctx_pg->nr_pages % MAX_CTX_PAGES;
nr_pages         6688 drivers/net/ethernet/broadcom/bnxt/bnxt.c 					rmem->nr_pages = rem;
nr_pages         6695 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		rmem->nr_pages = DIV_ROUND_UP(mem_size, BNXT_PAGE_SIZE);
nr_pages         6696 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		if (rmem->nr_pages > 1 || depth)
nr_pages         6708 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	if (rmem->depth > 1 || ctx_pg->nr_pages > MAX_CTX_PAGES ||
nr_pages         6710 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		int i, nr_tbls = rmem->nr_pages;
nr_pages         6729 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	ctx_pg->nr_pages = 0;
nr_pages          711 drivers/net/ethernet/broadcom/bnxt/bnxt.h 	int			nr_pages;
nr_pages         1309 drivers/net/ethernet/broadcom/bnxt/bnxt.h 	u32		nr_pages;
nr_pages          408 drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c 	u32 nr_pages, size, i, j, k = 0;
nr_pages          417 drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c 	nr_pages = size / BNXT_PAGE_SIZE;
nr_pages          419 drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c 		nr_pages++;
nr_pages          421 drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          447 drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c 	bp->pf.hwrm_cmd_req_pages = nr_pages;
nr_pages          247 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c int cxgbi_ppm_ppods_reserve(struct cxgbi_ppm *ppm, unsigned short nr_pages,
nr_pages          257 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 	npods = (nr_pages + PPOD_PAGES_MAX - 1) >> PPOD_PAGES_SHIFT;
nr_pages          260 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 			__func__, nr_pages, npods);
nr_pages          271 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 			 nr_pages, npods, ppm->next, caller_data);
nr_pages          287 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c 		 nr_pages, tag, idx, npods, caller_data);
nr_pages           94 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h 	int nr_pages;
nr_pages          323 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h int cxgbi_ppm_ppods_reserve(struct cxgbi_ppm *, unsigned short nr_pages,
nr_pages          246 drivers/net/ethernet/ibm/ehea/ehea.h 	u32 nr_pages;
nr_pages          291 drivers/net/ethernet/ibm/ehea/ehea.h 	u32 nr_pages;
nr_pages         3255 drivers/net/ethernet/ibm/ehea/ehea_main.c 		if (ehea_add_sect_bmap(arg->start_pfn, arg->nr_pages))
nr_pages         3263 drivers/net/ethernet/ibm/ehea/ehea_main.c 		if (ehea_rem_sect_bmap(arg->start_pfn, arg->nr_pages))
nr_pages          314 drivers/net/ethernet/ibm/ehea/ehea_phyp.c 	cq_attr->nr_pages = outs[4];
nr_pages          381 drivers/net/ethernet/ibm/ehea/ehea_phyp.c 	eq_attr->nr_pages = outs[4];
nr_pages           93 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	int i, nr_pages;
nr_pages          100 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	nr_pages = queue->queue_length / queue->pagesize;
nr_pages          102 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	for (i = 0; i < nr_pages; i += pages_per_kpage)
nr_pages          134 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	ret = hw_queue_ctor(&cq->hw_queue, cq->attr.nr_pages,
nr_pages          139 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	for (counter = 0; counter < cq->attr.nr_pages; counter++) {
nr_pages          152 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 			       cq, hret, counter, cq->attr.nr_pages);
nr_pages          156 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 		if (counter == (cq->attr.nr_pages - 1)) {
nr_pages          255 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	ret = hw_queue_ctor(&eq->hw_queue, eq->attr.nr_pages,
nr_pages          262 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	for (i = 0; i < eq->attr.nr_pages; i++) {
nr_pages          276 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 		if (i == (eq->attr.nr_pages - 1)) {
nr_pages          358 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 			   int nr_pages, int wqe_size, int act_nr_sges,
nr_pages          365 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	ret = hw_queue_ctor(hw_queue, nr_pages, EHEA_PAGESIZE, wqe_size);
nr_pages          369 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	for (cnt = 0; cnt < nr_pages; cnt++) {
nr_pages          608 drivers/net/ethernet/ibm/ehea/ehea_qmr.c static int ehea_update_busmap(unsigned long pfn, unsigned long nr_pages, int add)
nr_pages          612 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	if (!nr_pages)
nr_pages          622 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	end_section = start_section + ((nr_pages * PAGE_SIZE) / EHEA_SECTSIZE);
nr_pages          651 drivers/net/ethernet/ibm/ehea/ehea_qmr.c int ehea_add_sect_bmap(unsigned long pfn, unsigned long nr_pages)
nr_pages          656 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	ret = ehea_update_busmap(pfn, nr_pages, EHEA_BUSMAP_ADD_SECT);
nr_pages          661 drivers/net/ethernet/ibm/ehea/ehea_qmr.c int ehea_rem_sect_bmap(unsigned long pfn, unsigned long nr_pages)
nr_pages          666 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	ret = ehea_update_busmap(pfn, nr_pages, EHEA_BUSMAP_REM_SECT);
nr_pages          689 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	unsigned long pfn, start_pfn, end_pfn, nr_pages;
nr_pages          703 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 			nr_pages = pfn - start_pfn;
nr_pages          704 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 			ret = ehea_update_busmap(start_pfn, nr_pages,
nr_pages          717 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	nr_pages = pfn - start_pfn;
nr_pages          718 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	return ehea_update_busmap(start_pfn, nr_pages, EHEA_BUSMAP_ADD_SECT);
nr_pages          384 drivers/net/ethernet/ibm/ehea/ehea_qmr.h int ehea_add_sect_bmap(unsigned long pfn, unsigned long nr_pages);
nr_pages          385 drivers/net/ethernet/ibm/ehea/ehea_qmr.h int ehea_rem_sect_bmap(unsigned long pfn, unsigned long nr_pages);
nr_pages          244 drivers/nvdimm/pmem.c 		long nr_pages, void **kaddr, pfn_t *pfn)
nr_pages          249 drivers/nvdimm/pmem.c 					PFN_PHYS(nr_pages))))
nr_pages          262 drivers/nvdimm/pmem.c 		return nr_pages;
nr_pages          273 drivers/nvdimm/pmem.c 		pgoff_t pgoff, long nr_pages, void **kaddr, pfn_t *pfn)
nr_pages          277 drivers/nvdimm/pmem.c 	return __pmem_direct_access(pmem, pgoff, nr_pages, kaddr, pfn);
nr_pages           30 drivers/nvdimm/pmem.h 		long nr_pages, void **kaddr, pfn_t *pfn);
nr_pages           45 drivers/perf/arm_spe_pmu.c 	int					nr_pages;
nr_pages           78 drivers/perf/arm_spe_pmu.c #define PERF_IDX2OFF(idx, buf)	((idx) % ((buf)->nr_pages << PAGE_SHIFT))
nr_pages          362 drivers/perf/arm_spe_pmu.c 	u64 limit = buf->nr_pages * PAGE_SIZE;
nr_pages          380 drivers/perf/arm_spe_pmu.c 		limit = ((buf->nr_pages * PAGE_SIZE) >> 1) + handle->head;
nr_pages          390 drivers/perf/arm_spe_pmu.c 	const u64 bufsize = buf->nr_pages * PAGE_SIZE;
nr_pages          818 drivers/perf/arm_spe_pmu.c 				   int nr_pages, bool snapshot)
nr_pages          825 drivers/perf/arm_spe_pmu.c 	if (nr_pages < 2)
nr_pages          834 drivers/perf/arm_spe_pmu.c 	if (!nr_pages || (snapshot && (nr_pages & 1)))
nr_pages          844 drivers/perf/arm_spe_pmu.c 	pglist = kcalloc(nr_pages, sizeof(*pglist), GFP_KERNEL);
nr_pages          848 drivers/perf/arm_spe_pmu.c 	for (i = 0; i < nr_pages; ++i)
nr_pages          851 drivers/perf/arm_spe_pmu.c 	buf->base = vmap(pglist, nr_pages, VM_MAP, PAGE_KERNEL);
nr_pages          855 drivers/perf/arm_spe_pmu.c 	buf->nr_pages	= nr_pages;
nr_pages          542 drivers/rapidio/devices/rio_mport_cdev.c 	unsigned int nr_pages;
nr_pages          581 drivers/rapidio/devices/rio_mport_cdev.c 		for (i = 0; i < req->nr_pages; i++)
nr_pages          813 drivers/rapidio/devices/rio_mport_cdev.c 	unsigned long nr_pages = 0;
nr_pages          856 drivers/rapidio/devices/rio_mport_cdev.c 		nr_pages = PAGE_ALIGN(xfer->length + offset) >> PAGE_SHIFT;
nr_pages          858 drivers/rapidio/devices/rio_mport_cdev.c 		page_list = kmalloc_array(nr_pages,
nr_pages          867 drivers/rapidio/devices/rio_mport_cdev.c 				nr_pages,
nr_pages          871 drivers/rapidio/devices/rio_mport_cdev.c 		if (pinned != nr_pages) {
nr_pages          875 drivers/rapidio/devices/rio_mport_cdev.c 				nr_pages = 0;
nr_pages          878 drivers/rapidio/devices/rio_mport_cdev.c 					   pinned, nr_pages);
nr_pages          884 drivers/rapidio/devices/rio_mport_cdev.c 			nr_pages = pinned;
nr_pages          888 drivers/rapidio/devices/rio_mport_cdev.c 		ret = sg_alloc_table_from_pages(&req->sgt, page_list, nr_pages,
nr_pages          896 drivers/rapidio/devices/rio_mport_cdev.c 		req->nr_pages = nr_pages;
nr_pages          954 drivers/rapidio/devices/rio_mport_cdev.c 		for (i = 0; i < nr_pages; i++)
nr_pages           37 drivers/s390/block/dcssblk.c 		long nr_pages, void **kaddr, pfn_t *pfn);
nr_pages          920 drivers/s390/block/dcssblk.c 		long nr_pages, void **kaddr, pfn_t *pfn)
nr_pages          937 drivers/s390/block/dcssblk.c 		long nr_pages, void **kaddr, pfn_t *pfn)
nr_pages          941 drivers/s390/block/dcssblk.c 	return __dcssblk_direct_access(dev_info, pgoff, nr_pages, kaddr, pfn);
nr_pages          328 drivers/s390/char/sclp_cmd.c 	size = arg->nr_pages << PAGE_SHIFT;
nr_pages           63 drivers/s390/char/vmcp.c 	int nr_pages, order;
nr_pages           66 drivers/s390/char/vmcp.c 	nr_pages = ALIGN(session->bufsize, PAGE_SIZE) >> PAGE_SHIFT;
nr_pages           73 drivers/s390/char/vmcp.c 		page = cma_alloc(vmcp_cma, nr_pages, 0, false);
nr_pages           84 drivers/s390/char/vmcp.c 	int nr_pages, order;
nr_pages           90 drivers/s390/char/vmcp.c 	nr_pages = ALIGN(session->bufsize, PAGE_SIZE) >> PAGE_SHIFT;
nr_pages           93 drivers/s390/char/vmcp.c 		cma_release(vmcp_cma, page, nr_pages);
nr_pages         1099 drivers/s390/cio/css.c struct gen_pool *cio_gp_dma_create(struct device *dma_dev, int nr_pages)
nr_pages         1109 drivers/s390/cio/css.c 	for (i = 0; i < nr_pages; ++i) {
nr_pages         1368 drivers/scsi/cxgbi/libcxgbi.c 	ttinfo->nr_pages = (xferlen + sgl->offset + (1 << PAGE_SHIFT) - 1) >>
nr_pages         1376 drivers/scsi/cxgbi/libcxgbi.c 	err = cxgbi_ppm_ppods_reserve(ppm, ttinfo->nr_pages, 0, &ttinfo->idx,
nr_pages         1393 drivers/scsi/cxgbi/libcxgbi.c 	if (err != ttinfo->nr_pages) {
nr_pages         4903 drivers/scsi/st.c 	const int nr_pages = end - start;
nr_pages         4913 drivers/scsi/st.c         if (nr_pages > max_pages)
nr_pages         4926 drivers/scsi/st.c 	res = get_user_pages_fast(uaddr, nr_pages, rw == READ ? FOLL_WRITE : 0,
nr_pages         4930 drivers/scsi/st.c 	if (res < nr_pages)
nr_pages         4933 drivers/scsi/st.c         for (i=0; i < nr_pages; i++) {
nr_pages         4943 drivers/scsi/st.c 	return nr_pages;
nr_pages         4957 drivers/scsi/st.c 				const unsigned int nr_pages, int dirtied)
nr_pages         4961 drivers/scsi/st.c 	for (i=0; i < nr_pages; i++) {
nr_pages           35 drivers/staging/android/ion/ion_cma_heap.c 	unsigned long nr_pages = size >> PAGE_SHIFT;
nr_pages           42 drivers/staging/android/ion/ion_cma_heap.c 	pages = cma_alloc(cma_heap->cma, nr_pages, align, false);
nr_pages           47 drivers/staging/android/ion/ion_cma_heap.c 		unsigned long nr_clear_pages = nr_pages;
nr_pages           79 drivers/staging/android/ion/ion_cma_heap.c 	cma_release(cma_heap->cma, pages, nr_pages);
nr_pages           87 drivers/staging/android/ion/ion_cma_heap.c 	unsigned long nr_pages = PAGE_ALIGN(buffer->size) >> PAGE_SHIFT;
nr_pages           90 drivers/staging/android/ion/ion_cma_heap.c 	cma_release(cma_heap->cma, pages, nr_pages);
nr_pages         3169 drivers/staging/exfat/exfat_super.c 			   struct list_head *pages, unsigned int nr_pages)
nr_pages         3171 drivers/staging/exfat/exfat_super.c 	return  mpage_readpages(mapping, pages, nr_pages, exfat_get_block);
nr_pages          191 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 	ttinfo->nr_pages = (xferlen + sgl->offset +
nr_pages          197 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 	ret = cxgbi_ppm_ppods_reserve(ppm, ttinfo->nr_pages, 0, &ttinfo->idx,
nr_pages          854 drivers/target/target_core_pscsi.c 	int nr_pages = (cmd->data_length + sgl[0].offset +
nr_pages          861 drivers/target/target_core_pscsi.c 	pr_debug("PSCSI: nr_pages: %d\n", nr_pages);
nr_pages          884 drivers/target/target_core_pscsi.c 				nr_vecs = min_t(int, BIO_MAX_PAGES, nr_pages);
nr_pages          885 drivers/target/target_core_pscsi.c 				nr_pages -= nr_vecs;
nr_pages           31 drivers/tee/optee/shm_pool.c 		unsigned int nr_pages = 1 << order, i;
nr_pages           34 drivers/tee/optee/shm_pool.c 		pages = kcalloc(nr_pages, sizeof(pages), GFP_KERNEL);
nr_pages           38 drivers/tee/optee/shm_pool.c 		for (i = 0; i < nr_pages; i++) {
nr_pages           44 drivers/tee/optee/shm_pool.c 		rc = optee_shm_register(shm->ctx, shm, pages, nr_pages,
nr_pages          645 drivers/video/fbdev/pvr2fb.c 	unsigned int nr_pages;
nr_pages          649 drivers/video/fbdev/pvr2fb.c 	nr_pages = (count + PAGE_SIZE - 1) >> PAGE_SHIFT;
nr_pages          651 drivers/video/fbdev/pvr2fb.c 	pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_KERNEL);
nr_pages          655 drivers/video/fbdev/pvr2fb.c 	ret = get_user_pages_fast((unsigned long)buf, nr_pages, FOLL_WRITE, pages);
nr_pages          656 drivers/video/fbdev/pvr2fb.c 	if (ret < nr_pages) {
nr_pages          657 drivers/video/fbdev/pvr2fb.c 		nr_pages = ret;
nr_pages          666 drivers/video/fbdev/pvr2fb.c 	end   = (unsigned long)page_address(pages[nr_pages]);
nr_pages          667 drivers/video/fbdev/pvr2fb.c 	len   = nr_pages << PAGE_SHIFT;
nr_pages          685 drivers/video/fbdev/pvr2fb.c 	for (i = 0; i < nr_pages; i++, dst += PAGE_SIZE) {
nr_pages          701 drivers/video/fbdev/pvr2fb.c 	for (i = 0; i < nr_pages; i++)
nr_pages          447 drivers/video/fbdev/sh_mobile_lcdcfb.c 	int nr_pages = 0;
nr_pages          452 drivers/video/fbdev/sh_mobile_lcdcfb.c 		sg_set_page(&ch->sglist[nr_pages++], page, PAGE_SIZE, 0);
nr_pages          454 drivers/video/fbdev/sh_mobile_lcdcfb.c 	return nr_pages;
nr_pages          482 drivers/video/fbdev/sh_mobile_lcdcfb.c 		unsigned int nr_pages = sh_mobile_lcdc_sginit(info, pagelist);
nr_pages          485 drivers/video/fbdev/sh_mobile_lcdcfb.c 		dma_map_sg(ch->lcdc->dev, ch->sglist, nr_pages, DMA_TO_DEVICE);
nr_pages          489 drivers/video/fbdev/sh_mobile_lcdcfb.c 		dma_unmap_sg(ch->lcdc->dev, ch->sglist, nr_pages,
nr_pages           48 drivers/video/fbdev/xen-fbfront.c 	int			nr_pages;
nr_pages          413 drivers/video/fbdev/xen-fbfront.c 	info->nr_pages = (fb_size + PAGE_SIZE - 1) >> PAGE_SHIFT;
nr_pages          415 drivers/video/fbdev/xen-fbfront.c 	info->gfns = vmalloc(array_size(sizeof(unsigned long), info->nr_pages));
nr_pages          561 drivers/video/fbdev/xen-fbfront.c 	for (i = 0; i < info->nr_pages; i++)
nr_pages          564 drivers/video/fbdev/xen-fbfront.c 	for (i = 0; i * epd < info->nr_pages; i++)
nr_pages          414 drivers/xen/balloon.c static enum bp_state increase_reservation(unsigned long nr_pages)
nr_pages          420 drivers/xen/balloon.c 	if (nr_pages > ARRAY_SIZE(frame_list))
nr_pages          421 drivers/xen/balloon.c 		nr_pages = ARRAY_SIZE(frame_list);
nr_pages          424 drivers/xen/balloon.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          426 drivers/xen/balloon.c 			nr_pages = i;
nr_pages          434 drivers/xen/balloon.c 	rc = xenmem_reservation_increase(nr_pages, frame_list);
nr_pages          453 drivers/xen/balloon.c static enum bp_state decrease_reservation(unsigned long nr_pages, gfp_t gfp)
nr_pages          461 drivers/xen/balloon.c 	if (nr_pages > ARRAY_SIZE(frame_list))
nr_pages          462 drivers/xen/balloon.c 		nr_pages = ARRAY_SIZE(frame_list);
nr_pages          464 drivers/xen/balloon.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          467 drivers/xen/balloon.c 			nr_pages = i;
nr_pages          502 drivers/xen/balloon.c 	ret = xenmem_reservation_decrease(nr_pages, frame_list);
nr_pages          503 drivers/xen/balloon.c 	BUG_ON(ret != nr_pages);
nr_pages          505 drivers/xen/balloon.c 	balloon_stats.current_pages -= nr_pages;
nr_pages          566 drivers/xen/balloon.c static int add_ballooned_pages(int nr_pages)
nr_pages          581 drivers/xen/balloon.c 	if (si_mem_available() < nr_pages)
nr_pages          584 drivers/xen/balloon.c 	st = decrease_reservation(nr_pages, GFP_USER);
nr_pages          597 drivers/xen/balloon.c int alloc_xenballooned_pages(int nr_pages, struct page **pages)
nr_pages          605 drivers/xen/balloon.c 	balloon_stats.target_unpopulated += nr_pages;
nr_pages          607 drivers/xen/balloon.c 	while (pgno < nr_pages) {
nr_pages          625 drivers/xen/balloon.c 			ret = add_ballooned_pages(nr_pages - pgno);
nr_pages          644 drivers/xen/balloon.c void free_xenballooned_pages(int nr_pages, struct page **pages)
nr_pages          650 drivers/xen/balloon.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          655 drivers/xen/balloon.c 	balloon_stats.target_unpopulated -= nr_pages;
nr_pages          698 drivers/xen/balloon.c 		? min(xen_start_info->nr_pages - xen_released_pages, max_pfn)
nr_pages           58 drivers/xen/gntdev-dmabuf.c 	int nr_pages;
nr_pages          201 drivers/xen/gntdev-dmabuf.c dmabuf_pages_to_sgt(struct page **pages, unsigned int nr_pages)
nr_pages          212 drivers/xen/gntdev-dmabuf.c 	ret = sg_alloc_table_from_pages(sgt, pages, nr_pages, 0,
nr_pages          213 drivers/xen/gntdev-dmabuf.c 					nr_pages << PAGE_SHIFT,
nr_pages          271 drivers/xen/gntdev-dmabuf.c 	pr_debug("Mapping %d pages for dev %p\n", gntdev_dmabuf->nr_pages,
nr_pages          289 drivers/xen/gntdev-dmabuf.c 				  gntdev_dmabuf->nr_pages);
nr_pages          399 drivers/xen/gntdev-dmabuf.c 	gntdev_dmabuf->nr_pages = args->count;
nr_pages          593 drivers/xen/gntdev-dmabuf.c 	gntdev_dmabuf->nr_pages = count;
nr_pages          645 drivers/xen/gntdev-dmabuf.c 	if (attach->dmabuf->size != gntdev_dmabuf->nr_pages << PAGE_SHIFT) {
nr_pages          648 drivers/xen/gntdev-dmabuf.c 			 attach->dmabuf->size, gntdev_dmabuf->nr_pages);
nr_pages          733 drivers/xen/gntdev-dmabuf.c 				      gntdev_dmabuf->nr_pages);
nr_pages          105 drivers/xen/gntdev.c 		args.nr_pages = map->count;
nr_pages          173 drivers/xen/gntdev.c 		args.nr_pages = count;
nr_pages          833 drivers/xen/gntdev.c 	unsigned int nr_pages;
nr_pages          848 drivers/xen/gntdev.c 	batch->pages[batch->nr_pages++] = page;
nr_pages          860 drivers/xen/gntdev.c 	for (i = 0; i < batch->nr_pages; i++)
nr_pages          862 drivers/xen/gntdev.c 	batch->nr_pages = 0;
nr_pages         1001 drivers/xen/gntdev.c 	batch.nr_pages = 0;
nr_pages          776 drivers/xen/grant-table.c int gnttab_pages_set_private(int nr_pages, struct page **pages)
nr_pages          780 drivers/xen/grant-table.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          802 drivers/xen/grant-table.c int gnttab_alloc_pages(int nr_pages, struct page **pages)
nr_pages          806 drivers/xen/grant-table.c 	ret = alloc_xenballooned_pages(nr_pages, pages);
nr_pages          810 drivers/xen/grant-table.c 	ret = gnttab_pages_set_private(nr_pages, pages);
nr_pages          812 drivers/xen/grant-table.c 		gnttab_free_pages(nr_pages, pages);
nr_pages          818 drivers/xen/grant-table.c void gnttab_pages_clear_private(int nr_pages, struct page **pages)
nr_pages          822 drivers/xen/grant-table.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          838 drivers/xen/grant-table.c void gnttab_free_pages(int nr_pages, struct page **pages)
nr_pages          840 drivers/xen/grant-table.c 	gnttab_pages_clear_private(nr_pages, pages);
nr_pages          841 drivers/xen/grant-table.c 	free_xenballooned_pages(nr_pages, pages);
nr_pages          856 drivers/xen/grant-table.c 	size = args->nr_pages << PAGE_SHIFT;
nr_pages          871 drivers/xen/grant-table.c 	for (pfn = start_pfn, i = 0; pfn < start_pfn + args->nr_pages;
nr_pages          880 drivers/xen/grant-table.c 	xenmem_reservation_va_mapping_reset(args->nr_pages, args->pages);
nr_pages          882 drivers/xen/grant-table.c 	ret = xenmem_reservation_decrease(args->nr_pages, args->frames);
nr_pages          883 drivers/xen/grant-table.c 	if (ret != args->nr_pages) {
nr_pages          889 drivers/xen/grant-table.c 	ret = gnttab_pages_set_private(args->nr_pages, args->pages);
nr_pages          910 drivers/xen/grant-table.c 	gnttab_pages_clear_private(args->nr_pages, args->pages);
nr_pages          912 drivers/xen/grant-table.c 	for (i = 0; i < args->nr_pages; i++)
nr_pages          915 drivers/xen/grant-table.c 	ret = xenmem_reservation_increase(args->nr_pages, args->frames);
nr_pages          916 drivers/xen/grant-table.c 	if (ret != args->nr_pages) {
nr_pages          923 drivers/xen/grant-table.c 	xenmem_reservation_va_mapping_update(args->nr_pages, args->pages,
nr_pages          926 drivers/xen/grant-table.c 	size = args->nr_pages << PAGE_SHIFT;
nr_pages           68 drivers/xen/privcmd.c                unsigned long nr_pages);
nr_pages          453 drivers/xen/privcmd.c 	unsigned long nr_pages;
nr_pages          481 drivers/xen/privcmd.c 	nr_pages = DIV_ROUND_UP(m.num, XEN_PFN_PER_PAGE);
nr_pages          482 drivers/xen/privcmd.c 	if ((m.num <= 0) || (nr_pages > (LONG_MAX >> PAGE_SHIFT)))
nr_pages          524 drivers/xen/privcmd.c 		    m.addr + (nr_pages << PAGE_SHIFT) != vma->vm_end) {
nr_pages          529 drivers/xen/privcmd.c 			ret = alloc_empty_pages(vma, nr_pages);
nr_pages          536 drivers/xen/privcmd.c 		    m.addr + (nr_pages << PAGE_SHIFT) > vma->vm_end) {
nr_pages          540 drivers/xen/privcmd.c 		if (privcmd_vma_range_is_mapped(vma, m.addr, nr_pages)) {
nr_pages          585 drivers/xen/privcmd.c 	struct page *pages[], unsigned int nr_pages)
nr_pages          596 drivers/xen/privcmd.c 		if (requested > nr_pages)
nr_pages          605 drivers/xen/privcmd.c 		nr_pages -= pinned;
nr_pages          612 drivers/xen/privcmd.c static void unlock_pages(struct page *pages[], unsigned int nr_pages)
nr_pages          619 drivers/xen/privcmd.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          630 drivers/xen/privcmd.c 	unsigned int nr_pages = 0;
nr_pages          671 drivers/xen/privcmd.c 		nr_pages += DIV_ROUND_UP(
nr_pages          676 drivers/xen/privcmd.c 	pages = kcalloc(nr_pages, sizeof(*pages), GFP_KERNEL);
nr_pages          688 drivers/xen/privcmd.c 	rc = lock_pages(kbufs, kdata.num, pages, nr_pages);
nr_pages          702 drivers/xen/privcmd.c 	unlock_pages(pages, nr_pages);
nr_pages          951 drivers/xen/privcmd.c 	           unsigned long nr_pages)
nr_pages          953 drivers/xen/privcmd.c 	return apply_to_page_range(vma->vm_mm, addr, nr_pages << PAGE_SHIFT,
nr_pages           91 drivers/xen/swiotlb-xen.c 	unsigned int i, nr_pages = XEN_PFN_UP(xen_offset_in_page(p) + size);
nr_pages           95 drivers/xen/swiotlb-xen.c 	for (i = 1; i < nr_pages; i++)
nr_pages          360 drivers/xen/xenbus/xenbus_client.c 		      unsigned int nr_pages, grant_ref_t *grefs)
nr_pages          365 drivers/xen/xenbus/xenbus_client.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          562 drivers/xen/xenbus/xenbus_client.c 	unsigned int nr_pages = XENBUS_PAGES(nr_grefs);
nr_pages          573 drivers/xen/xenbus/xenbus_client.c 	err = alloc_xenballooned_pages(nr_pages, node->hvm.pages);
nr_pages          588 drivers/xen/xenbus/xenbus_client.c 	addr = vmap(node->hvm.pages, nr_pages, VM_MAP | VM_IOREMAP,
nr_pages          609 drivers/xen/xenbus/xenbus_client.c 			 addr, nr_pages);
nr_pages          612 drivers/xen/xenbus/xenbus_client.c 		free_xenballooned_pages(nr_pages, node->hvm.pages);
nr_pages          829 drivers/xen/xenbus/xenbus_client.c 	unsigned int nr_pages;
nr_pages          849 drivers/xen/xenbus/xenbus_client.c 	nr_pages = XENBUS_PAGES(node->nr_handles);
nr_pages          859 drivers/xen/xenbus/xenbus_client.c 		free_xenballooned_pages(nr_pages, node->hvm.pages);
nr_pages          862 drivers/xen/xenbus/xenbus_client.c 		WARN(1, "Leaking %p, size %u page(s)\n", vaddr, nr_pages);
nr_pages          222 drivers/xen/xlate_mmu.c 	unsigned long nr_pages;
nr_pages          225 drivers/xen/xlate_mmu.c 	nr_pages = DIV_ROUND_UP(nr_grant_frames, XEN_PFN_PER_PAGE);
nr_pages          226 drivers/xen/xlate_mmu.c 	pages = kcalloc(nr_pages, sizeof(pages[0]), GFP_KERNEL);
nr_pages          235 drivers/xen/xlate_mmu.c 	rc = alloc_xenballooned_pages(nr_pages, pages);
nr_pages          238 drivers/xen/xlate_mmu.c 			nr_pages, rc);
nr_pages          248 drivers/xen/xlate_mmu.c 	vaddr = vmap(pages, nr_pages, 0, PAGE_KERNEL);
nr_pages          251 drivers/xen/xlate_mmu.c 			nr_pages, rc);
nr_pages          252 drivers/xen/xlate_mmu.c 		free_xenballooned_pages(nr_pages, pages);
nr_pages          281 fs/9p/cache.c  				  unsigned *nr_pages)
nr_pages          286 fs/9p/cache.c  	p9_debug(P9_DEBUG_FSC, "inode %p pages %u\n", inode, *nr_pages);
nr_pages          291 fs/9p/cache.c  					  mapping, pages, nr_pages,
nr_pages          302 fs/9p/cache.c  		BUG_ON(*nr_pages != 0);
nr_pages           37 fs/9p/cache.h  					 unsigned *nr_pages);
nr_pages           62 fs/9p/cache.h  					      unsigned *nr_pages)
nr_pages           65 fs/9p/cache.h  					     nr_pages);
nr_pages          118 fs/9p/cache.h  					      unsigned *nr_pages)
nr_pages           99 fs/9p/vfs_addr.c 			     struct list_head *pages, unsigned nr_pages)
nr_pages          107 fs/9p/vfs_addr.c 	ret = v9fs_readpages_from_fscache(inode, mapping, pages, &nr_pages);
nr_pages          169 fs/afs/dir.c   	for (i = 0; i < req->nr_pages; i++)
nr_pages          179 fs/afs/dir.c   		req->pos, req->index, req->nr_pages, req->offset);
nr_pages          181 fs/afs/dir.c   	for (i = 0; i < req->nr_pages; i++) {
nr_pages          219 fs/afs/dir.c   	int nr_pages, nr_inline, i, n;
nr_pages          236 fs/afs/dir.c   	nr_pages = (i_size + PAGE_SIZE - 1) / PAGE_SIZE;
nr_pages          237 fs/afs/dir.c   	nr_inline = nr_pages;
nr_pages          246 fs/afs/dir.c   	req->nr_pages = nr_pages;
nr_pages          248 fs/afs/dir.c   	req->len = nr_pages * PAGE_SIZE; /* We can ask for more than there is */
nr_pages          253 fs/afs/dir.c   		req->pages = kcalloc(nr_pages, sizeof(struct page *),
nr_pages          267 fs/afs/dir.c   					  req->nr_pages - i,
nr_pages          269 fs/afs/dir.c   		_debug("find %u at %u/%u", n, i, req->nr_pages);
nr_pages          293 fs/afs/dir.c   	} while (i < req->nr_pages);
nr_pages          315 fs/afs/dir.c   		task_io_account_read(PAGE_SIZE * req->nr_pages);
nr_pages           26 fs/afs/file.c  			 struct list_head *pages, unsigned nr_pages);
nr_pages          195 fs/afs/file.c  			for (i = 0; i < req->nr_pages; i++)
nr_pages          326 fs/afs/file.c  		req->nr_pages = 1;
nr_pages          498 fs/afs/file.c  		req->pages[req->nr_pages++] = page;
nr_pages          500 fs/afs/file.c  	} while (req->nr_pages < n);
nr_pages          502 fs/afs/file.c  	if (req->nr_pages == 0) {
nr_pages          511 fs/afs/file.c  	task_io_account_read(PAGE_SIZE * req->nr_pages);
nr_pages          523 fs/afs/file.c  	for (i = 0; i < req->nr_pages; i++) {
nr_pages          542 fs/afs/file.c  			 struct list_head *pages, unsigned nr_pages)
nr_pages          549 fs/afs/file.c  	       key_serial(key), mapping->host->i_ino, nr_pages);
nr_pages          564 fs/afs/file.c  					  &nr_pages,
nr_pages          576 fs/afs/file.c  		BUG_ON(nr_pages != 0);
nr_pages          363 fs/afs/fsclient.c 		ASSERTCMP(req->index, <, req->nr_pages);
nr_pages          437 fs/afs/fsclient.c 	for (; req->index < req->nr_pages; req->index++) {
nr_pages          445 fs/afs/fsclient.c 		for (req->index = 0; req->index < req->nr_pages; req->index++)
nr_pages          233 fs/afs/internal.h 	unsigned int		nr_pages;
nr_pages           54 fs/afs/write.c 	req->nr_pages = 1;
nr_pages          475 fs/afs/yfsclient.c 		ASSERTCMP(req->index, <, req->nr_pages);
nr_pages          553 fs/afs/yfsclient.c 	for (; req->index < req->nr_pages; req->index++) {
nr_pages          561 fs/afs/yfsclient.c 		for (req->index = 0; req->index < req->nr_pages; req->index++)
nr_pages          128 fs/aio.c       	long			nr_pages;
nr_pages          236 fs/aio.c       static struct file *aio_private_file(struct kioctx *ctx, loff_t nr_pages)
nr_pages          245 fs/aio.c       	inode->i_size = PAGE_SIZE * nr_pages;
nr_pages          311 fs/aio.c       	for (i = 0; i < ctx->nr_pages; i++) {
nr_pages          413 fs/aio.c       	if (idx < (pgoff_t)ctx->nr_pages) {
nr_pages          466 fs/aio.c       	int nr_pages;
nr_pages          476 fs/aio.c       	nr_pages = PFN_UP(size);
nr_pages          477 fs/aio.c       	if (nr_pages < 0)
nr_pages          480 fs/aio.c       	file = aio_private_file(ctx, nr_pages);
nr_pages          487 fs/aio.c       	nr_events = (PAGE_SIZE * nr_pages - sizeof(struct aio_ring))
nr_pages          491 fs/aio.c       	if (nr_pages > AIO_RING_PAGES) {
nr_pages          492 fs/aio.c       		ctx->ring_pages = kcalloc(nr_pages, sizeof(struct page *),
nr_pages          500 fs/aio.c       	for (i = 0; i < nr_pages; i++) {
nr_pages          513 fs/aio.c       	ctx->nr_pages = i;
nr_pages          515 fs/aio.c       	if (unlikely(i != nr_pages)) {
nr_pages          520 fs/aio.c       	ctx->mmap_size = nr_pages * PAGE_SIZE;
nr_pages          204 fs/block_dev.c 		int nr_pages)
nr_pages          219 fs/block_dev.c 	if (nr_pages <= DIO_INLINE_BIO_VECS)
nr_pages          222 fs/block_dev.c 		vecs = kmalloc_array(nr_pages, sizeof(struct bio_vec),
nr_pages          228 fs/block_dev.c 	bio_init(&bio, vecs, nr_pages);
nr_pages          339 fs/block_dev.c __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages)
nr_pages          357 fs/block_dev.c 	bio = bio_alloc_bioset(GFP_KERNEL, nr_pages, &blkdev_dio_pool);
nr_pages          406 fs/block_dev.c 		nr_pages = iov_iter_npages(iter, BIO_MAX_PAGES);
nr_pages          407 fs/block_dev.c 		if (!nr_pages) {
nr_pages          437 fs/block_dev.c 		bio = bio_alloc(GFP_KERNEL, nr_pages);
nr_pages          469 fs/block_dev.c 	int nr_pages;
nr_pages          471 fs/block_dev.c 	nr_pages = iov_iter_npages(iter, BIO_MAX_PAGES + 1);
nr_pages          472 fs/block_dev.c 	if (!nr_pages)
nr_pages          474 fs/block_dev.c 	if (is_sync_kiocb(iocb) && nr_pages <= BIO_MAX_PAGES)
nr_pages          475 fs/block_dev.c 		return __blkdev_direct_IO_simple(iocb, iter, nr_pages);
nr_pages          477 fs/block_dev.c 	return __blkdev_direct_IO(iocb, iter, min(nr_pages, BIO_MAX_PAGES));
nr_pages          619 fs/block_dev.c 			struct list_head *pages, unsigned nr_pages)
nr_pages          621 fs/block_dev.c 	return mpage_readpages(mapping, pages, nr_pages, blkdev_get_block);
nr_pages           93 fs/btrfs/compression.c 	for (i = 0; i < cb->nr_pages; i++) {
nr_pages          176 fs/btrfs/compression.c 	for (index = 0; index < cb->nr_pages; index++) {
nr_pages          217 fs/btrfs/compression.c 	unsigned long nr_pages = end_index - index + 1;
nr_pages          224 fs/btrfs/compression.c 	while (nr_pages > 0) {
nr_pages          227 fs/btrfs/compression.c 				     nr_pages, ARRAY_SIZE(pages)), pages);
nr_pages          229 fs/btrfs/compression.c 			nr_pages -= 1;
nr_pages          239 fs/btrfs/compression.c 		nr_pages -= ret;
nr_pages          287 fs/btrfs/compression.c 	for (index = 0; index < cb->nr_pages; index++) {
nr_pages          313 fs/btrfs/compression.c 				 unsigned long nr_pages,
nr_pages          340 fs/btrfs/compression.c 	cb->nr_pages = nr_pages;
nr_pages          353 fs/btrfs/compression.c 	for (pg_index = 0; pg_index < cb->nr_pages; pg_index++) {
nr_pages          397 fs/btrfs/compression.c 			       bytes_left, cb->compressed_len, cb->nr_pages);
nr_pages          438 fs/btrfs/compression.c 	unsigned long nr_pages = 0;
nr_pages          521 fs/btrfs/compression.c 			nr_pages++;
nr_pages          553 fs/btrfs/compression.c 	unsigned long nr_pages;
nr_pages          601 fs/btrfs/compression.c 	nr_pages = DIV_ROUND_UP(compressed_len, PAGE_SIZE);
nr_pages          602 fs/btrfs/compression.c 	cb->compressed_pages = kcalloc(nr_pages, sizeof(struct page *),
nr_pages          609 fs/btrfs/compression.c 	for (pg_index = 0; pg_index < nr_pages; pg_index++) {
nr_pages          618 fs/btrfs/compression.c 	faili = nr_pages - 1;
nr_pages          619 fs/btrfs/compression.c 	cb->nr_pages = nr_pages;
nr_pages          633 fs/btrfs/compression.c 	for (pg_index = 0; pg_index < nr_pages; pg_index++) {
nr_pages           51 fs/btrfs/compression.h 	unsigned long nr_pages;
nr_pages           95 fs/btrfs/compression.h 				  unsigned long nr_pages,
nr_pages         1872 fs/btrfs/extent_io.c 	unsigned long nr_pages = end_index - start_index + 1;
nr_pages         1885 fs/btrfs/extent_io.c 	if ((page_ops & PAGE_SET_ERROR) && nr_pages > 0)
nr_pages         1888 fs/btrfs/extent_io.c 	while (nr_pages > 0) {
nr_pages         1891 fs/btrfs/extent_io.c 				     nr_pages, ARRAY_SIZE(pages)), pages);
nr_pages         1934 fs/btrfs/extent_io.c 		nr_pages -= ret;
nr_pages         3265 fs/btrfs/extent_io.c 					     struct page *pages[], int nr_pages,
nr_pages         3277 fs/btrfs/extent_io.c 	for (index = 0; index < nr_pages; index++) {
nr_pages         3936 fs/btrfs/extent_io.c 	int nr_pages;
nr_pages         3964 fs/btrfs/extent_io.c 	       (nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end,
nr_pages         3968 fs/btrfs/extent_io.c 		for (i = 0; i < nr_pages; i++) {
nr_pages         4104 fs/btrfs/extent_io.c 	int nr_pages;
nr_pages         4162 fs/btrfs/extent_io.c 			(nr_pages = pagevec_lookup_range_tag(&pvec, mapping,
nr_pages         4166 fs/btrfs/extent_io.c 		for (i = 0; i < nr_pages; i++) {
nr_pages         4273 fs/btrfs/extent_io.c 	unsigned long nr_pages = (end - start + PAGE_SIZE) >>
nr_pages         4284 fs/btrfs/extent_io.c 		.nr_to_write	= nr_pages * 2,
nr_pages         4333 fs/btrfs/extent_io.c 		     unsigned nr_pages)
nr_pages          416 fs/btrfs/extent_io.h 		     unsigned nr_pages);
nr_pages          359 fs/btrfs/inode.c 	unsigned long nr_pages;
nr_pages          385 fs/btrfs/inode.c 				     unsigned long nr_pages,
nr_pages          396 fs/btrfs/inode.c 	async_extent->nr_pages = nr_pages;
nr_pages          480 fs/btrfs/inode.c 	unsigned long nr_pages;
nr_pages          507 fs/btrfs/inode.c 	nr_pages = (end >> PAGE_SHIFT) - (start >> PAGE_SHIFT) + 1;
nr_pages          509 fs/btrfs/inode.c 	nr_pages = min_t(unsigned long, nr_pages,
nr_pages          547 fs/btrfs/inode.c 		pages = kcalloc(nr_pages, sizeof(struct page *), GFP_NOFS);
nr_pages          550 fs/btrfs/inode.c 			nr_pages = 0;
nr_pages          581 fs/btrfs/inode.c 					   &nr_pages,
nr_pages          587 fs/btrfs/inode.c 			struct page *page = pages[nr_pages - 1];
nr_pages          643 fs/btrfs/inode.c 			for (i = 0; i < nr_pages; i++) {
nr_pages          676 fs/btrfs/inode.c 					total_compressed, pages, nr_pages,
nr_pages          693 fs/btrfs/inode.c 		for (i = 0; i < nr_pages; i++) {
nr_pages          700 fs/btrfs/inode.c 		nr_pages = 0;
nr_pages          738 fs/btrfs/inode.c 	for (i = 0; i < async_extent->nr_pages; i++) {
nr_pages          743 fs/btrfs/inode.c 	async_extent->nr_pages = 0;
nr_pages          882 fs/btrfs/inode.c 				    async_extent->nr_pages,
nr_pages         1174 fs/btrfs/inode.c 	unsigned long nr_pages;
nr_pages         1176 fs/btrfs/inode.c 	nr_pages = (async_chunk->end - async_chunk->start + PAGE_SIZE) >>
nr_pages         1180 fs/btrfs/inode.c 	if (atomic_sub_return(nr_pages, &fs_info->async_delalloc_pages) <
nr_pages         1217 fs/btrfs/inode.c 	unsigned long nr_pages;
nr_pages         1291 fs/btrfs/inode.c 		nr_pages = DIV_ROUND_UP(cur_end - start, PAGE_SIZE);
nr_pages         1292 fs/btrfs/inode.c 		atomic_add(nr_pages, &fs_info->async_delalloc_pages);
nr_pages         1296 fs/btrfs/inode.c 		*nr_written += nr_pages;
nr_pages         8895 fs/btrfs/inode.c 		struct list_head *pages, unsigned nr_pages)
nr_pages         8897 fs/btrfs/inode.c 	return extent_readpages(mapping, pages, nr_pages);
nr_pages         10801 fs/btrfs/inode.c 	unsigned long nr_pages;
nr_pages         10808 fs/btrfs/inode.c 	unsigned long nr_pages;
nr_pages         10818 fs/btrfs/inode.c 	nr_pages = next_ppage - first_ppage;
nr_pages         10828 fs/btrfs/inode.c 	ret = add_swap_extent(sis, bsi->nr_pages, nr_pages, first_ppage);
nr_pages         10832 fs/btrfs/inode.c 	bsi->nr_pages += nr_pages;
nr_pages         11048 fs/btrfs/inode.c 	sis->max = bsi.nr_pages;
nr_pages         11049 fs/btrfs/inode.c 	sis->pages = bsi.nr_pages - 1;
nr_pages         11050 fs/btrfs/inode.c 	sis->highest_bit = bsi.nr_pages - 1;
nr_pages          146 fs/btrfs/lzo.c 	int nr_pages = 0;
nr_pages          182 fs/btrfs/lzo.c 	nr_pages = 1;
nr_pages          237 fs/btrfs/lzo.c 				if (nr_pages == nr_dest_pages) {
nr_pages          249 fs/btrfs/lzo.c 				pages[nr_pages++] = out_page;
nr_pages          294 fs/btrfs/lzo.c 	*out_pages = nr_pages;
nr_pages          132 fs/btrfs/raid56.c 	int nr_pages;
nr_pages          263 fs/btrfs/raid56.c 	for (i = 0; i < rbio->nr_pages; i++) {
nr_pages          310 fs/btrfs/raid56.c 	for (i = 0; i < dest->nr_pages; i++) {
nr_pages          846 fs/btrfs/raid56.c 	for (i = 0; i < rbio->nr_pages; i++) {
nr_pages         1005 fs/btrfs/raid56.c 	rbio->nr_pages = num_pages;
nr_pages         1047 fs/btrfs/raid56.c 	for (i = 0; i < rbio->nr_pages; i++) {
nr_pages         1066 fs/btrfs/raid56.c 	for (; i < rbio->nr_pages; i++) {
nr_pages          316 fs/btrfs/space-info.c 					 unsigned long nr_pages, int nr_items)
nr_pages          321 fs/btrfs/space-info.c 		writeback_inodes_sb_nr(sb, nr_pages, WB_REASON_FS_FREE_SPACE);
nr_pages          365 fs/btrfs/space-info.c 	unsigned long nr_pages;
nr_pages          396 fs/btrfs/space-info.c 		nr_pages = min(delalloc_bytes, to_reclaim) >> PAGE_SHIFT;
nr_pages          403 fs/btrfs/space-info.c 		btrfs_writeback_inodes_sb_nr(fs_info, nr_pages, items);
nr_pages          418 fs/btrfs/space-info.c 		if (async_pages <= nr_pages)
nr_pages          421 fs/btrfs/space-info.c 			async_pages -= nr_pages;
nr_pages           26 fs/btrfs/tests/extent-io-tests.c 	unsigned long nr_pages = end_index - index + 1;
nr_pages           31 fs/btrfs/tests/extent-io-tests.c 	while (nr_pages > 0) {
nr_pages           33 fs/btrfs/tests/extent-io-tests.c 				     min_t(unsigned long, nr_pages,
nr_pages           45 fs/btrfs/tests/extent-io-tests.c 		nr_pages -= ret;
nr_pages           52 fs/btrfs/tests/extent-io-tests.c 				start, end, nr_pages, ret);
nr_pages          103 fs/btrfs/zlib.c 	int nr_pages = 0;
nr_pages          134 fs/btrfs/zlib.c 	nr_pages = 1;
nr_pages          164 fs/btrfs/zlib.c 			if (nr_pages == nr_dest_pages) {
nr_pages          175 fs/btrfs/zlib.c 			pages[nr_pages] = out_page;
nr_pages          176 fs/btrfs/zlib.c 			nr_pages++;
nr_pages          220 fs/btrfs/zlib.c 	*out_pages = nr_pages;
nr_pages          381 fs/btrfs/zstd.c 	int nr_pages = 0;
nr_pages          418 fs/btrfs/zstd.c 	pages[nr_pages++] = out_page;
nr_pages          455 fs/btrfs/zstd.c 			if (nr_pages == nr_dest_pages) {
nr_pages          465 fs/btrfs/zstd.c 			pages[nr_pages++] = out_page;
nr_pages          515 fs/btrfs/zstd.c 		if (nr_pages == nr_dest_pages) {
nr_pages          525 fs/btrfs/zstd.c 		pages[nr_pages++] = out_page;
nr_pages          540 fs/btrfs/zstd.c 	*out_pages = nr_pages;
nr_pages          684 fs/cachefiles/rdwr.c 				   unsigned *nr_pages,
nr_pages          703 fs/cachefiles/rdwr.c 	       *nr_pages);
nr_pages          709 fs/cachefiles/rdwr.c 	if (cachefiles_has_space(cache, 0, *nr_pages) < 0)
nr_pages          752 fs/cachefiles/rdwr.c 			(*nr_pages)--;
nr_pages          778 fs/cachefiles/rdwr.c 	       ret, *nr_pages, list_empty(pages) ? " empty" : "");
nr_pages          782 fs/cachefiles/rdwr.c 	fscache_retrieval_complete(op, *nr_pages);
nr_pages          836 fs/cachefiles/rdwr.c 			      unsigned *nr_pages,
nr_pages          850 fs/cachefiles/rdwr.c 	_enter("%p,,,%d,", object, *nr_pages);
nr_pages          852 fs/cachefiles/rdwr.c 	ret = cachefiles_has_space(cache, 0, *nr_pages);
nr_pages          868 fs/cachefiles/rdwr.c 	fscache_retrieval_complete(op, *nr_pages);
nr_pages          321 fs/ceph/addr.c 	int nr_pages = 0;
nr_pages          356 fs/ceph/addr.c 		nr_pages++;
nr_pages          358 fs/ceph/addr.c 		if (max && nr_pages == max)
nr_pages          361 fs/ceph/addr.c 	len = nr_pages << PAGE_SHIFT;
nr_pages          362 fs/ceph/addr.c 	dout("start_read %p nr_pages %d is %lld~%lld\n", inode, nr_pages,
nr_pages          376 fs/ceph/addr.c 	nr_pages = calc_pages_for(0, len);
nr_pages          377 fs/ceph/addr.c 	pages = kmalloc_array(nr_pages, sizeof(*pages), GFP_KERNEL);
nr_pages          382 fs/ceph/addr.c 	for (i = 0; i < nr_pages; ++i) {
nr_pages          395 fs/ceph/addr.c 			nr_pages = i;
nr_pages          396 fs/ceph/addr.c 			if (nr_pages > 0) {
nr_pages          397 fs/ceph/addr.c 				len = nr_pages << PAGE_SHIFT;
nr_pages          420 fs/ceph/addr.c 	return nr_pages;
nr_pages          423 fs/ceph/addr.c 	for (i = 0; i < nr_pages; ++i) {
nr_pages          427 fs/ceph/addr.c 	ceph_put_page_vector(pages, nr_pages, false);
nr_pages          442 fs/ceph/addr.c 			  struct list_head *page_list, unsigned nr_pages)
nr_pages          455 fs/ceph/addr.c 					 &nr_pages);
nr_pages          463 fs/ceph/addr.c 	     inode, file, rw_ctx, nr_pages, max);
nr_pages          255 fs/ceph/cache.c 				  unsigned *nr_pages)
nr_pages          263 fs/ceph/cache.c 	ret = fscache_read_or_alloc_pages(ci->fscache, mapping, pages, nr_pages,
nr_pages           31 fs/ceph/cache.h 				unsigned *nr_pages);
nr_pages          135 fs/ceph/cache.h 					      unsigned *nr_pages)
nr_pages         1334 fs/cifs/cifsglob.h 	unsigned int			nr_pages;
nr_pages         1360 fs/cifs/cifsglob.h 	unsigned int			nr_pages;
nr_pages          559 fs/cifs/cifsproto.h struct cifs_writedata *cifs_writedata_alloc(unsigned int nr_pages,
nr_pages         1702 fs/cifs/cifssmb.c 				 .rq_npages = rdata->nr_pages,
nr_pages         2080 fs/cifs/cifssmb.c 		unsigned int j, nr_pages, wsize, tailsz, cur_len;
nr_pages         2084 fs/cifs/cifssmb.c 			nr_pages = wsize / PAGE_SIZE;
nr_pages         2085 fs/cifs/cifssmb.c 			if (!nr_pages) {
nr_pages         2089 fs/cifs/cifssmb.c 			cur_len = nr_pages * PAGE_SIZE;
nr_pages         2092 fs/cifs/cifssmb.c 			nr_pages = DIV_ROUND_UP(rest_len, PAGE_SIZE);
nr_pages         2094 fs/cifs/cifssmb.c 			tailsz = rest_len - (nr_pages - 1) * PAGE_SIZE;
nr_pages         2097 fs/cifs/cifssmb.c 		wdata2 = cifs_writedata_alloc(nr_pages, cifs_writev_complete);
nr_pages         2103 fs/cifs/cifssmb.c 		for (j = 0; j < nr_pages; j++) {
nr_pages         2110 fs/cifs/cifssmb.c 		wdata2->nr_pages = nr_pages;
nr_pages         2129 fs/cifs/cifssmb.c 		for (j = 0; j < nr_pages; j++) {
nr_pages         2142 fs/cifs/cifssmb.c 			i += nr_pages;
nr_pages         2147 fs/cifs/cifssmb.c 		i += nr_pages;
nr_pages         2148 fs/cifs/cifssmb.c 	} while (i < wdata->nr_pages);
nr_pages         2151 fs/cifs/cifssmb.c 	for (; i < wdata->nr_pages; i++) {
nr_pages         2179 fs/cifs/cifssmb.c 	for (i = 0; i < wdata->nr_pages; i++) {
nr_pages         2194 fs/cifs/cifssmb.c cifs_writedata_alloc(unsigned int nr_pages, work_func_t complete)
nr_pages         2197 fs/cifs/cifssmb.c 		kcalloc(nr_pages, sizeof(struct page *), GFP_NOFS);
nr_pages         2321 fs/cifs/cifssmb.c 	rqst.rq_npages = wdata->nr_pages;
nr_pages         2175 fs/cifs/file.c 	unsigned int nr_pages = 0, i;
nr_pages         2187 fs/cifs/file.c 		if (nr_pages == 0)
nr_pages         2232 fs/cifs/file.c 		++nr_pages;
nr_pages         2236 fs/cifs/file.c 	if (nr_pages == 0)
nr_pages         2240 fs/cifs/file.c 	for (i = nr_pages; i < found_pages; i++) {
nr_pages         2245 fs/cifs/file.c 	return nr_pages;
nr_pages         2249 fs/cifs/file.c wdata_send_pages(struct cifs_writedata *wdata, unsigned int nr_pages,
nr_pages         2257 fs/cifs/file.c 	wdata->nr_pages = nr_pages;
nr_pages         2261 fs/cifs/file.c 			page_offset(wdata->pages[nr_pages - 1]),
nr_pages         2263 fs/cifs/file.c 	wdata->bytes = ((nr_pages - 1) * PAGE_SIZE) + wdata->tailsz;
nr_pages         2313 fs/cifs/file.c 		unsigned int i, nr_pages, found_pages, wsize;
nr_pages         2352 fs/cifs/file.c 		nr_pages = wdata_prepare_pages(wdata, found_pages, mapping, wbc,
nr_pages         2356 fs/cifs/file.c 		if (nr_pages == 0) {
nr_pages         2374 fs/cifs/file.c 			rc = wdata_send_pages(wdata, nr_pages, mapping, wbc);
nr_pages         2376 fs/cifs/file.c 		for (i = 0; i < nr_pages; ++i)
nr_pages         2382 fs/cifs/file.c 			for (i = 0; i < nr_pages; ++i) {
nr_pages         2410 fs/cifs/file.c 		wbc->nr_to_write -= nr_pages;
nr_pages         2686 fs/cifs/file.c 	for (i = 0; i < wdata->nr_pages; i++)
nr_pages         2718 fs/cifs/file.c 	unsigned long i, nr_pages = *num_pages;
nr_pages         2721 fs/cifs/file.c 	for (i = 0; i < nr_pages; i++) {
nr_pages         2826 fs/cifs/file.c 	unsigned long nr_pages, num_pages, i;
nr_pages         2885 fs/cifs/file.c 			nr_pages =
nr_pages         2899 fs/cifs/file.c 				nr_pages > 1 ?
nr_pages         2901 fs/cifs/file.c 					(nr_pages - 2) * PAGE_SIZE :
nr_pages         2904 fs/cifs/file.c 			nr_pages = get_numpages(wsize, len, &cur_len);
nr_pages         2905 fs/cifs/file.c 			wdata = cifs_writedata_alloc(nr_pages,
nr_pages         2913 fs/cifs/file.c 			rc = cifs_write_allocate_pages(wdata->pages, nr_pages);
nr_pages         2921 fs/cifs/file.c 			num_pages = nr_pages;
nr_pages         2925 fs/cifs/file.c 				for (i = 0; i < nr_pages; i++)
nr_pages         2937 fs/cifs/file.c 			for ( ; nr_pages > num_pages; nr_pages--)
nr_pages         2938 fs/cifs/file.c 				put_page(wdata->pages[nr_pages - 1]);
nr_pages         2940 fs/cifs/file.c 			wdata->tailsz = cur_len - ((nr_pages - 1) * PAGE_SIZE);
nr_pages         2944 fs/cifs/file.c 		wdata->nr_pages = nr_pages;
nr_pages         3285 fs/cifs/file.c cifs_readdata_alloc(unsigned int nr_pages, work_func_t complete)
nr_pages         3288 fs/cifs/file.c 		kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL);
nr_pages         3319 fs/cifs/file.c cifs_read_allocate_pages(struct cifs_readdata *rdata, unsigned int nr_pages)
nr_pages         3325 fs/cifs/file.c 	for (i = 0; i < nr_pages; i++) {
nr_pages         3353 fs/cifs/file.c 	for (i = 0; i < rdata->nr_pages; i++) {
nr_pages         3374 fs/cifs/file.c 	for (i = 0; i < rdata->nr_pages; i++) {
nr_pages         3414 fs/cifs/file.c 	unsigned int nr_pages = rdata->nr_pages;
nr_pages         3419 fs/cifs/file.c 	for (i = 0; i < nr_pages; i++) {
nr_pages         3433 fs/cifs/file.c 			rdata->nr_pages--;
nr_pages         3647 fs/cifs/file.c 		rdata->nr_pages = npages;
nr_pages         4099 fs/cifs/file.c 	for (i = 0; i < rdata->nr_pages; i++) {
nr_pages         4133 fs/cifs/file.c 	unsigned int nr_pages = rdata->nr_pages;
nr_pages         4143 fs/cifs/file.c 	for (i = 0; i < nr_pages; i++) {
nr_pages         4178 fs/cifs/file.c 			rdata->nr_pages--;
nr_pages         4186 fs/cifs/file.c 			rdata->nr_pages--;
nr_pages         4228 fs/cifs/file.c 		    unsigned int *nr_pages, loff_t *offset, unsigned int *bytes)
nr_pages         4257 fs/cifs/file.c 	*nr_pages = 1;
nr_pages         4279 fs/cifs/file.c 		(*nr_pages)++;
nr_pages         4333 fs/cifs/file.c 		unsigned int i, nr_pages, bytes, rsize;
nr_pages         4366 fs/cifs/file.c 					 &nr_pages, &offset, &bytes);
nr_pages         4372 fs/cifs/file.c 		rdata = cifs_readdata_alloc(nr_pages, cifs_readv_complete);
nr_pages         4399 fs/cifs/file.c 			rdata->pages[rdata->nr_pages++] = page;
nr_pages         4413 fs/cifs/file.c 			for (i = 0; i < rdata->nr_pages; i++) {
nr_pages          278 fs/cifs/fscache.c 				unsigned *nr_pages)
nr_pages          283 fs/cifs/fscache.c 		 __func__, CIFS_I(inode)->fscache, *nr_pages, inode);
nr_pages          285 fs/cifs/fscache.c 					  pages, nr_pages,
nr_pages           95 fs/cifs/fscache.h 					      unsigned *nr_pages)
nr_pages           99 fs/cifs/fscache.h 						     nr_pages);
nr_pages          149 fs/cifs/fscache.h 					      unsigned *nr_pages)
nr_pages         3578 fs/cifs/smb2pdu.c 				rdata->nr_pages, rdata->page_offset,
nr_pages         3639 fs/cifs/smb2pdu.c 				 .rq_npages = rdata->nr_pages,
nr_pages         3986 fs/cifs/smb2pdu.c 				wdata->nr_pages, wdata->page_offset,
nr_pages         3994 fs/cifs/smb2pdu.c 		if (wdata->nr_pages > 1)
nr_pages         3997 fs/cifs/smb2pdu.c 					(wdata->nr_pages - 1) * wdata->pagesz -
nr_pages         4022 fs/cifs/smb2pdu.c 	rqst.rq_npages = wdata->nr_pages;
nr_pages          728 fs/direct-io.c 	int ret, nr_pages;
nr_pages          734 fs/direct-io.c 	nr_pages = min(sdio->pages_in_io, BIO_MAX_PAGES);
nr_pages          735 fs/direct-io.c 	BUG_ON(nr_pages <= 0);
nr_pages          736 fs/direct-io.c 	dio_bio_alloc(dio, sdio, map_bh->b_bdev, sector, nr_pages);
nr_pages          286 fs/erofs/data.c 				      unsigned int nr_pages)
nr_pages          293 fs/erofs/data.c 	trace_erofs_readpages(mapping->host, page, nr_pages, true);
nr_pages          295 fs/erofs/data.c 	for (; nr_pages; --nr_pages) {
nr_pages          303 fs/erofs/data.c 						  &last_block, nr_pages, true);
nr_pages           64 fs/erofs/zdata.c 	cl->nr_pages = 0;
nr_pages           76 fs/erofs/zdata.c 	DBG_BUGON(cl->nr_pages);
nr_pages          679 fs/erofs/zdata.c 	clt->cl->nr_pages = max_t(pgoff_t, clt->cl->nr_pages, index + 1);
nr_pages          762 fs/erofs/zdata.c 	unsigned int i, outputsize, llen, nr_pages;
nr_pages          773 fs/erofs/zdata.c 	DBG_BUGON(!READ_ONCE(cl->nr_pages));
nr_pages          776 fs/erofs/zdata.c 	nr_pages = cl->nr_pages;
nr_pages          778 fs/erofs/zdata.c 	if (nr_pages <= Z_EROFS_VMAP_ONSTACK_PAGES) {
nr_pages          780 fs/erofs/zdata.c 	} else if (nr_pages <= Z_EROFS_VMAP_GLOBAL_PAGES &&
nr_pages          786 fs/erofs/zdata.c 		if (nr_pages > Z_EROFS_VMAP_GLOBAL_PAGES)
nr_pages          789 fs/erofs/zdata.c 		pages = kvmalloc_array(nr_pages, sizeof(struct page *),
nr_pages          799 fs/erofs/zdata.c 	for (i = 0; i < nr_pages; ++i)
nr_pages          823 fs/erofs/zdata.c 		DBG_BUGON(pagenr >= nr_pages);
nr_pages          864 fs/erofs/zdata.c 			DBG_BUGON(pagenr >= nr_pages);
nr_pages          887 fs/erofs/zdata.c 	if (nr_pages << PAGE_SHIFT >= cl->pageofs + llen) {
nr_pages          891 fs/erofs/zdata.c 		outputsize = (nr_pages << PAGE_SHIFT) - cl->pageofs;
nr_pages          921 fs/erofs/zdata.c 	for (i = 0; i < nr_pages; ++i) {
nr_pages          943 fs/erofs/zdata.c 	cl->nr_pages = 0;
nr_pages         1370 fs/erofs/zdata.c 					      unsigned int nr_pages)
nr_pages         1375 fs/erofs/zdata.c 	bool sync = should_decompress_synchronously(sbi, nr_pages);
nr_pages         1382 fs/erofs/zdata.c 			      nr_pages, false);
nr_pages         1386 fs/erofs/zdata.c 	for (; nr_pages; --nr_pages) {
nr_pages           32 fs/erofs/zdata.h 	unsigned short nr_pages;
nr_pages          882 fs/ext2/inode.c 		struct list_head *pages, unsigned nr_pages)
nr_pages          884 fs/ext2/inode.c 	return mpage_readpages(mapping, pages, nr_pages, ext2_get_block);
nr_pages         3241 fs/ext4/ext4.h 				unsigned nr_pages, bool is_readahead);
nr_pages         1692 fs/ext4/inode.c 	int nr_pages, i;
nr_pages         1713 fs/ext4/inode.c 		nr_pages = pagevec_lookup_range(&pvec, mapping, &index, end);
nr_pages         1714 fs/ext4/inode.c 		if (nr_pages == 0)
nr_pages         1716 fs/ext4/inode.c 		for (i = 0; i < nr_pages; i++) {
nr_pages         2369 fs/ext4/inode.c 	int nr_pages, i;
nr_pages         2385 fs/ext4/inode.c 		nr_pages = pagevec_lookup_range(&pvec, inode->i_mapping,
nr_pages         2387 fs/ext4/inode.c 		if (nr_pages == 0)
nr_pages         2389 fs/ext4/inode.c 		for (i = 0; i < nr_pages; i++) {
nr_pages         2635 fs/ext4/inode.c 	unsigned int nr_pages;
nr_pages         2654 fs/ext4/inode.c 		nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end,
nr_pages         2656 fs/ext4/inode.c 		if (nr_pages == 0)
nr_pages         2659 fs/ext4/inode.c 		for (i = 0; i < nr_pages; i++) {
nr_pages         3343 fs/ext4/inode.c 		struct list_head *pages, unsigned nr_pages)
nr_pages         3351 fs/ext4/inode.c 	return ext4_mpage_readpages(mapping, pages, NULL, nr_pages, true);
nr_pages          227 fs/ext4/readpage.c 			 unsigned nr_pages, bool is_readahead)
nr_pages          251 fs/ext4/readpage.c 	for (; nr_pages; nr_pages--) {
nr_pages          269 fs/ext4/readpage.c 		last_block = block_in_file + nr_pages * blocks_per_page;
nr_pages          377 fs/ext4/readpage.c 				min_t(int, nr_pages, BIO_MAX_PAGES));
nr_pages          369 fs/f2fs/checkpoint.c 	int nr_pages;
nr_pages          379 fs/f2fs/checkpoint.c 	while ((nr_pages = pagevec_lookup_tag(&pvec, mapping, &index,
nr_pages          383 fs/f2fs/checkpoint.c 		for (i = 0; i < nr_pages; i++) {
nr_pages          677 fs/f2fs/data.c 				      unsigned nr_pages, unsigned op_flag,
nr_pages          685 fs/f2fs/data.c 	bio = f2fs_bio_alloc(sbi, min_t(int, nr_pages, BIO_MAX_PAGES), false);
nr_pages         1639 fs/f2fs/data.c 					unsigned nr_pages,
nr_pages         1655 fs/f2fs/data.c 	last_block = block_in_file + nr_pages;
nr_pages         1723 fs/f2fs/data.c 		bio = f2fs_grab_read_bio(inode, block_nr, nr_pages,
nr_pages         1767 fs/f2fs/data.c 			unsigned nr_pages, bool is_readahead)
nr_pages         1784 fs/f2fs/data.c 	for (; nr_pages; nr_pages--) {
nr_pages         1796 fs/f2fs/data.c 		ret = f2fs_read_single_page(inode, page, nr_pages, &map, &bio,
nr_pages         1831 fs/f2fs/data.c 			struct list_head *pages, unsigned nr_pages)
nr_pages         1836 fs/f2fs/data.c 	trace_f2fs_readpages(inode, page, nr_pages);
nr_pages         1842 fs/f2fs/data.c 	return f2fs_mpage_readpages(mapping, pages, NULL, nr_pages, true);
nr_pages         2256 fs/f2fs/data.c 	int nr_pages;
nr_pages         2300 fs/f2fs/data.c 		nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end,
nr_pages         2302 fs/f2fs/data.c 		if (nr_pages == 0)
nr_pages         2305 fs/f2fs/data.c 		for (i = 0; i < nr_pages; i++) {
nr_pages         1274 fs/f2fs/f2fs.h 	atomic_t nr_pages[NR_COUNT_TYPE];
nr_pages         1870 fs/f2fs/f2fs.h 	atomic_inc(&sbi->nr_pages[count_type]);
nr_pages         1891 fs/f2fs/f2fs.h 	atomic_dec(&sbi->nr_pages[count_type]);
nr_pages         1909 fs/f2fs/f2fs.h 	return atomic_read(&sbi->nr_pages[count_type]);
nr_pages          347 fs/f2fs/file.c 	int nr_pages;
nr_pages          353 fs/f2fs/file.c 	nr_pages = find_get_pages_tag(mapping, &pgofs, PAGECACHE_TAG_DIRTY,
nr_pages          355 fs/f2fs/file.c 	if (!nr_pages)
nr_pages         1447 fs/f2fs/node.c 	int nr_pages;
nr_pages         1452 fs/f2fs/node.c 	while ((nr_pages = pagevec_lookup_tag(&pvec, NODE_MAPPING(sbi), &index,
nr_pages         1456 fs/f2fs/node.c 		for (i = 0; i < nr_pages; i++) {
nr_pages         1659 fs/f2fs/node.c 	int nr_pages;
nr_pages         1671 fs/f2fs/node.c 	while ((nr_pages = pagevec_lookup_tag(&pvec, NODE_MAPPING(sbi), &index,
nr_pages         1675 fs/f2fs/node.c 		for (i = 0; i < nr_pages; i++) {
nr_pages         1817 fs/f2fs/node.c 	int nr_pages, done = 0;
nr_pages         1824 fs/f2fs/node.c 	while (!done && (nr_pages = pagevec_lookup_tag(&pvec,
nr_pages         1828 fs/f2fs/node.c 		for (i = 0; i < nr_pages; i++) {
nr_pages         2832 fs/f2fs/super.c 		atomic_set(&sbi->nr_pages[i], 0);
nr_pages          213 fs/fat/inode.c 			 struct list_head *pages, unsigned nr_pages)
nr_pages          215 fs/fat/inode.c 	return mpage_readpages(mapping, pages, nr_pages, fat_get_block);
nr_pages          393 fs/file_table.c 	unsigned long nr_pages = totalram_pages();
nr_pages          394 fs/file_table.c 	unsigned long memreserve = (nr_pages - nr_free_pages()) * 3/2;
nr_pages          396 fs/file_table.c 	memreserve = min(memreserve, nr_pages - 1);
nr_pages          397 fs/file_table.c 	n = ((nr_pages - memreserve) * (PAGE_SIZE / 1024)) / 10;
nr_pages           43 fs/fs-writeback.c 	long nr_pages;
nr_pages          809 fs/fs-writeback.c static long wb_split_bdi_pages(struct bdi_writeback *wb, long nr_pages)
nr_pages          814 fs/fs-writeback.c 	if (nr_pages == LONG_MAX)
nr_pages          823 fs/fs-writeback.c 		return nr_pages;
nr_pages          825 fs/fs-writeback.c 		return DIV_ROUND_UP_ULL((u64)nr_pages * this_bw, tot_bw);
nr_pages          854 fs/fs-writeback.c 		long nr_pages;
nr_pages          869 fs/fs-writeback.c 		nr_pages = wb_split_bdi_pages(wb, base_work->nr_pages);
nr_pages          874 fs/fs-writeback.c 			work->nr_pages = nr_pages;
nr_pages          883 fs/fs-writeback.c 		work->nr_pages = nr_pages;
nr_pages          970 fs/fs-writeback.c 		work->nr_pages = nr;
nr_pages         1047 fs/fs-writeback.c static long wb_split_bdi_pages(struct bdi_writeback *wb, long nr_pages)
nr_pages         1049 fs/fs-writeback.c 	return nr_pages;
nr_pages         1607 fs/fs-writeback.c 		pages = min(pages, work->nr_pages);
nr_pages         1719 fs/fs-writeback.c 		work->nr_pages -= write_chunk - wbc.nr_to_write;
nr_pages         1759 fs/fs-writeback.c 			if (work->nr_pages <= 0)
nr_pages         1792 fs/fs-writeback.c 			if (work->nr_pages <= 0)
nr_pages         1800 fs/fs-writeback.c static long writeback_inodes_wb(struct bdi_writeback *wb, long nr_pages,
nr_pages         1804 fs/fs-writeback.c 		.nr_pages	= nr_pages,
nr_pages         1819 fs/fs-writeback.c 	return nr_pages - work.nr_pages;
nr_pages         1841 fs/fs-writeback.c 	long nr_pages = work->nr_pages;
nr_pages         1856 fs/fs-writeback.c 		if (work->nr_pages <= 0)
nr_pages         1930 fs/fs-writeback.c 	return nr_pages - work->nr_pages;
nr_pages         1955 fs/fs-writeback.c 			.nr_pages	= LONG_MAX,
nr_pages         1971 fs/fs-writeback.c 	long nr_pages;
nr_pages         1985 fs/fs-writeback.c 	nr_pages = get_nr_dirty_pages();
nr_pages         1987 fs/fs-writeback.c 	if (nr_pages) {
nr_pages         1989 fs/fs-writeback.c 			.nr_pages	= nr_pages,
nr_pages         2004 fs/fs-writeback.c 	long nr_pages;
nr_pages         2009 fs/fs-writeback.c 	nr_pages = get_nr_dirty_pages();
nr_pages         2010 fs/fs-writeback.c 	if (nr_pages) {
nr_pages         2012 fs/fs-writeback.c 			.nr_pages	= wb_split_bdi_pages(wb, nr_pages),
nr_pages         2018 fs/fs-writeback.c 		nr_pages = wb_writeback(wb, &work);
nr_pages         2022 fs/fs-writeback.c 	return nr_pages;
nr_pages         2460 fs/fs-writeback.c 		.nr_pages		= nr,
nr_pages         2536 fs/fs-writeback.c 		.nr_pages	= LONG_MAX,
nr_pages          561 fs/fscache/page.c 				  unsigned *nr_pages,
nr_pages          571 fs/fscache/page.c 	_enter("%p,,%d,,,", cookie, *nr_pages);
nr_pages          584 fs/fscache/page.c 	ASSERTCMP(*nr_pages, >, 0);
nr_pages          593 fs/fscache/page.c 	atomic_set(&op->n_pages, *nr_pages);
nr_pages          627 fs/fscache/page.c 			op, pages, nr_pages, gfp);
nr_pages          632 fs/fscache/page.c 			op, pages, nr_pages, gfp);
nr_pages          922 fs/fuse/file.c 	unsigned int nr_pages;
nr_pages          940 fs/fuse/file.c 		data->max_pages = min_t(unsigned int, data->nr_pages,
nr_pages          961 fs/fuse/file.c 	data->nr_pages--;
nr_pages          966 fs/fuse/file.c 			  struct list_head *pages, unsigned nr_pages)
nr_pages          979 fs/fuse/file.c 	data.nr_pages = nr_pages;
nr_pages          980 fs/fuse/file.c 	data.max_pages = min_t(unsigned int, nr_pages, fc->max_pages);
nr_pages         1230 fs/fuse/file.c 		unsigned int nr_pages = fuse_wr_pages(pos, iov_iter_count(ii),
nr_pages         1233 fs/fuse/file.c 		ap->pages = fuse_pages_alloc(nr_pages, GFP_KERNEL, &ap->descs);
nr_pages         1239 fs/fuse/file.c 		count = fuse_fill_write_pages(ap, mapping, ii, pos, nr_pages);
nr_pages         1346 fs/fuse/file.c 					       unsigned int nr_pages)
nr_pages         1350 fs/fuse/file.c 	for (i = index; i < index + nr_pages; i++)
nr_pages          241 fs/gfs2/aops.c 				    int nr_pages,
nr_pages          246 fs/gfs2/aops.c 	unsigned nrblocks = nr_pages * (PAGE_SIZE >> inode->i_blkbits);
nr_pages          254 fs/gfs2/aops.c 	for(i = 0; i < nr_pages; i++) {
nr_pages          339 fs/gfs2/aops.c 	int nr_pages;
nr_pages          374 fs/gfs2/aops.c 		nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end,
nr_pages          376 fs/gfs2/aops.c 		if (nr_pages == 0)
nr_pages          379 fs/gfs2/aops.c 		ret = gfs2_write_jdata_pagevec(mapping, wbc, &pvec, nr_pages, &done_index);
nr_pages          600 fs/gfs2/aops.c 			  struct list_head *pages, unsigned nr_pages)
nr_pages          613 fs/gfs2/aops.c 		ret = mpage_readpages(mapping, pages, nr_pages, gfs2_block_map);
nr_pages          129 fs/hpfs/file.c 			  struct list_head *pages, unsigned nr_pages)
nr_pages          131 fs/hpfs/file.c 	return mpage_readpages(mapping, pages, nr_pages, hpfs_get_block);
nr_pages         3312 fs/io_uring.c  static void io_unaccount_mem(struct user_struct *user, unsigned long nr_pages)
nr_pages         3314 fs/io_uring.c  	atomic_long_sub(nr_pages, &user->locked_vm);
nr_pages         3317 fs/io_uring.c  static int io_account_mem(struct user_struct *user, unsigned long nr_pages)
nr_pages         3326 fs/io_uring.c  		new_pages = cur_pages + nr_pages;
nr_pages         3467 fs/io_uring.c  		int pret, nr_pages;
nr_pages         3491 fs/io_uring.c  		nr_pages = end - start;
nr_pages         3494 fs/io_uring.c  			ret = io_account_mem(ctx->user, nr_pages);
nr_pages         3500 fs/io_uring.c  		if (!pages || nr_pages > got_pages) {
nr_pages         3503 fs/io_uring.c  			pages = kvmalloc_array(nr_pages, sizeof(struct page *),
nr_pages         3505 fs/io_uring.c  			vmas = kvmalloc_array(nr_pages,
nr_pages         3511 fs/io_uring.c  					io_unaccount_mem(ctx->user, nr_pages);
nr_pages         3514 fs/io_uring.c  			got_pages = nr_pages;
nr_pages         3517 fs/io_uring.c  		imu->bvec = kvmalloc_array(nr_pages, sizeof(struct bio_vec),
nr_pages         3522 fs/io_uring.c  				io_unaccount_mem(ctx->user, nr_pages);
nr_pages         3528 fs/io_uring.c  		pret = get_user_pages(ubuf, nr_pages,
nr_pages         3531 fs/io_uring.c  		if (pret == nr_pages) {
nr_pages         3533 fs/io_uring.c  			for (j = 0; j < nr_pages; j++) {
nr_pages         3554 fs/io_uring.c  				io_unaccount_mem(ctx->user, nr_pages);
nr_pages         3561 fs/io_uring.c  		for (j = 0; j < nr_pages; j++) {
nr_pages         3574 fs/io_uring.c  		imu->nr_bvecs = nr_pages;
nr_pages          382 fs/iomap/buffered-io.c 		unsigned nr_pages, const struct iomap_ops *ops)
nr_pages          208 fs/iomap/direct-io.c 	int nr_pages, ret = 0;
nr_pages          245 fs/iomap/direct-io.c 	nr_pages = iov_iter_npages(&iter, BIO_MAX_PAGES);
nr_pages          246 fs/iomap/direct-io.c 	if (nr_pages <= 0)
nr_pages          247 fs/iomap/direct-io.c 		return nr_pages;
nr_pages          263 fs/iomap/direct-io.c 		bio = bio_alloc(GFP_KERNEL, nr_pages);
nr_pages          303 fs/iomap/direct-io.c 		nr_pages = iov_iter_npages(&iter, BIO_MAX_PAGES);
nr_pages          305 fs/iomap/direct-io.c 	} while (nr_pages);
nr_pages           88 fs/iomap/seek.c 		unsigned nr_pages, i;
nr_pages           90 fs/iomap/seek.c 		nr_pages = pagevec_lookup_range(&pvec, inode->i_mapping, &index,
nr_pages           92 fs/iomap/seek.c 		if (nr_pages == 0)
nr_pages           95 fs/iomap/seek.c 		for (i = 0; i < nr_pages; i++) {
nr_pages           19 fs/iomap/swapfile.c 	unsigned long nr_pages;		/* number of pages collected */
nr_pages           32 fs/iomap/swapfile.c 	unsigned long nr_pages;
nr_pages           49 fs/iomap/swapfile.c 	nr_pages = next_ppage - first_ppage;
nr_pages           65 fs/iomap/swapfile.c 	error = add_swap_extent(isi->sis, isi->nr_pages, nr_pages, first_ppage);
nr_pages           69 fs/iomap/swapfile.c 	isi->nr_pages += nr_pages;
nr_pages          173 fs/iomap/swapfile.c 	sis->max = isi.nr_pages;
nr_pages          174 fs/iomap/swapfile.c 	sis->pages = isi.nr_pages - 1;
nr_pages          175 fs/iomap/swapfile.c 	sis->highest_bit = isi.nr_pages - 1;
nr_pages         1189 fs/isofs/inode.c 			struct list_head *pages, unsigned nr_pages)
nr_pages         1191 fs/isofs/inode.c 	return mpage_readpages(mapping, pages, nr_pages, isofs_get_block);
nr_pages          300 fs/jfs/inode.c 		struct list_head *pages, unsigned nr_pages)
nr_pages          302 fs/jfs/inode.c 	return mpage_readpages(mapping, pages, nr_pages, jfs_get_block);
nr_pages          139 fs/mpage.c     	unsigned int nr_pages;
nr_pages          190 fs/mpage.c     	last_block = block_in_file + args->nr_pages * blocks_per_page;
nr_pages          307 fs/mpage.c     					min_t(int, args->nr_pages,
nr_pages          386 fs/mpage.c     				unsigned nr_pages, get_block_t get_block)
nr_pages          394 fs/mpage.c     	for (page_idx = 0; page_idx < nr_pages; page_idx++) {
nr_pages          403 fs/mpage.c     			args.nr_pages = nr_pages - page_idx;
nr_pages          422 fs/mpage.c     		.nr_pages = 1,
nr_pages          495 fs/nfs/blocklayout/extent_tree.c 		int nr_pages = DIV_ROUND_UP(buffer_size, PAGE_SIZE), i;
nr_pages          497 fs/nfs/blocklayout/extent_tree.c 		for (i = 0; i < nr_pages; i++)
nr_pages          448 fs/nfs/fscache.c 				 unsigned *nr_pages)
nr_pages          450 fs/nfs/fscache.c 	unsigned npages = *nr_pages;
nr_pages          457 fs/nfs/fscache.c 					  mapping, pages, nr_pages,
nr_pages          461 fs/nfs/fscache.c 	if (*nr_pages < npages)
nr_pages          464 fs/nfs/fscache.c 	if (*nr_pages > 0)
nr_pages          466 fs/nfs/fscache.c 				      *nr_pages);
nr_pages          471 fs/nfs/fscache.c 		BUG_ON(*nr_pages != 0);
nr_pages          146 fs/nfs/fscache.h 					     unsigned *nr_pages)
nr_pages          150 fs/nfs/fscache.h 						    nr_pages);
nr_pages          225 fs/nfs/fscache.h 					     unsigned *nr_pages)
nr_pages          405 fs/nfs/read.c  		struct list_head *pages, unsigned nr_pages)
nr_pages          419 fs/nfs/read.c  			nr_pages);
nr_pages          436 fs/nfs/read.c  					 pages, &nr_pages);
nr_pages          157 fs/nilfs2/inode.c 			   struct list_head *pages, unsigned int nr_pages)
nr_pages          159 fs/nilfs2/inode.c 	return mpage_readpages(mapping, pages, nr_pages, nilfs_get_block);
nr_pages          497 fs/ntfs/compress.c 	unsigned int nr_pages = (end_vcn - start_vcn) <<
nr_pages          507 fs/ntfs/compress.c 			"%i.", index, cb_size, nr_pages);
nr_pages          515 fs/ntfs/compress.c 	pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_NOFS);
nr_pages          516 fs/ntfs/compress.c 	completed_pages = kmalloc_array(nr_pages + 1, sizeof(int), GFP_NOFS);
nr_pages          559 fs/ntfs/compress.c 	if (nr_pages < max_page)
nr_pages          560 fs/ntfs/compress.c 		max_page = nr_pages;
nr_pages          495 fs/ntfs/file.c 		pgoff_t index, const unsigned nr_pages, struct page **pages,
nr_pages          500 fs/ntfs/file.c 	BUG_ON(!nr_pages);
nr_pages          526 fs/ntfs/file.c 	} while (nr < nr_pages);
nr_pages          571 fs/ntfs/file.c 		unsigned nr_pages, s64 pos, size_t bytes)
nr_pages          599 fs/ntfs/file.c 	BUG_ON(!nr_pages);
nr_pages          607 fs/ntfs/file.c 			vi->i_ino, ni->type, pages[0]->index, nr_pages,
nr_pages          624 fs/ntfs/file.c 	} while (++u < nr_pages);
nr_pages         1176 fs/ntfs/file.c 	if (likely(!err && ++u < nr_pages))
nr_pages         1222 fs/ntfs/file.c 		} while (++u < nr_pages);
nr_pages         1330 fs/ntfs/file.c 	nr_pages = u;
nr_pages         1337 fs/ntfs/file.c 			if (u == nr_pages &&
nr_pages         1355 fs/ntfs/file.c 	} while (++u <= nr_pages);
nr_pages         1361 fs/ntfs/file.c 		unsigned nr_pages)
nr_pages         1363 fs/ntfs/file.c 	BUG_ON(!nr_pages);
nr_pages         1370 fs/ntfs/file.c 		--nr_pages;
nr_pages         1371 fs/ntfs/file.c 		flush_dcache_page(pages[nr_pages]);
nr_pages         1372 fs/ntfs/file.c 	} while (nr_pages > 0);
nr_pages         1385 fs/ntfs/file.c 		struct page **pages, const unsigned nr_pages,
nr_pages         1431 fs/ntfs/file.c 	} while (++u < nr_pages);
nr_pages         1540 fs/ntfs/file.c 		const unsigned nr_pages, s64 pos, size_t bytes)
nr_pages         1555 fs/ntfs/file.c 	BUG_ON(!nr_pages);
nr_pages         1563 fs/ntfs/file.c 			vi->i_ino, ni->type, page->index, nr_pages,
nr_pages         1567 fs/ntfs/file.c 				nr_pages, pos, bytes);
nr_pages         1568 fs/ntfs/file.c 	BUG_ON(nr_pages > 1);
nr_pages         1685 fs/ntfs/file.c static size_t ntfs_copy_from_user_iter(struct page **pages, unsigned nr_pages,
nr_pages         1688 fs/ntfs/file.c 	struct page **last_page = pages + nr_pages;
nr_pages         1743 fs/ntfs/file.c 	unsigned nr_pages;
nr_pages         1774 fs/ntfs/file.c 	nr_pages = 1;
nr_pages         1776 fs/ntfs/file.c 		nr_pages = vol->cluster_size >> PAGE_SHIFT;
nr_pages         1788 fs/ntfs/file.c 		if (nr_pages > 1) {
nr_pages         1823 fs/ntfs/file.c 					do_pages = nr_pages;
nr_pages          355 fs/ocfs2/aops.c 			   struct list_head *pages, unsigned nr_pages)
nr_pages          392 fs/ocfs2/aops.c 	err = mpage_readpages(mapping, pages, nr_pages, ocfs2_get_block);
nr_pages          293 fs/omfs/file.c 		struct list_head *pages, unsigned nr_pages)
nr_pages          295 fs/omfs/file.c 	return mpage_readpages(mapping, pages, nr_pages, omfs_get_block);
nr_pages         1057 fs/pipe.c      	unsigned int size, nr_pages;
nr_pages         1062 fs/pipe.c      	nr_pages = size >> PAGE_SHIFT;
nr_pages         1064 fs/pipe.c      	if (!nr_pages)
nr_pages         1074 fs/pipe.c      	if (nr_pages > pipe->buffers &&
nr_pages         1078 fs/pipe.c      	user_bufs = account_pipe_buffers(pipe->user, pipe->buffers, nr_pages);
nr_pages         1080 fs/pipe.c      	if (nr_pages > pipe->buffers &&
nr_pages         1094 fs/pipe.c      	if (nr_pages < pipe->nrbufs) {
nr_pages         1099 fs/pipe.c      	bufs = kcalloc(nr_pages, sizeof(*bufs),
nr_pages         1130 fs/pipe.c      	pipe->buffers = nr_pages;
nr_pages         1131 fs/pipe.c      	return nr_pages * PAGE_SIZE;
nr_pages         1134 fs/pipe.c      	(void) account_pipe_buffers(pipe->user, nr_pages, pipe->buffers);
nr_pages          148 fs/proc/kcore.c 	unsigned long nr_pages = ent->size >> PAGE_SHIFT;
nr_pages          154 fs/proc/kcore.c 	end = ((unsigned long)pfn_to_page(pfn + nr_pages)) - 1;
nr_pages          186 fs/proc/kcore.c kclist_add_private(unsigned long pfn, unsigned long nr_pages, void *arg)
nr_pages          203 fs/proc/kcore.c 	ent->size = nr_pages << PAGE_SHIFT;
nr_pages         1668 fs/proc/task_mmu.c 			unsigned long nr_pages)
nr_pages         1672 fs/proc/task_mmu.c 	md->pages += nr_pages;
nr_pages         1674 fs/proc/task_mmu.c 		md->dirty += nr_pages;
nr_pages         1677 fs/proc/task_mmu.c 		md->swapcache += nr_pages;
nr_pages         1680 fs/proc/task_mmu.c 		md->active += nr_pages;
nr_pages         1683 fs/proc/task_mmu.c 		md->writeback += nr_pages;
nr_pages         1686 fs/proc/task_mmu.c 		md->anon += nr_pages;
nr_pages         1691 fs/proc/task_mmu.c 	md->node[page_to_nid(page)] += nr_pages;
nr_pages          103 fs/qnx6/inode.c 		   struct list_head *pages, unsigned nr_pages)
nr_pages          105 fs/qnx6/inode.c 	return mpage_readpages(mapping, pages, nr_pages, qnx6_get_block);
nr_pages         1165 fs/reiserfs/inode.c 		   struct list_head *pages, unsigned nr_pages)
nr_pages         1167 fs/reiserfs/inode.c 	return mpage_readpages(mapping, pages, nr_pages, reiserfs_get_block);
nr_pages          187 fs/splice.c    	unsigned int spd_pages = spd->nr_pages;
nr_pages          214 fs/splice.c    		if (!--spd->nr_pages)
nr_pages          372 fs/splice.c    	unsigned int nr_pages;
nr_pages          392 fs/splice.c    	nr_pages = DIV_ROUND_UP(res + base, PAGE_SIZE);
nr_pages          395 fs/splice.c    	if (nr_pages > PIPE_DEF_BUFFERS) {
nr_pages          396 fs/splice.c    		vec = kmalloc_array(nr_pages, sizeof(struct kvec), GFP_KERNEL);
nr_pages          406 fs/splice.c    	for (i = 0; i < nr_pages; i++) {
nr_pages          414 fs/splice.c    	res = kernel_readv(in, vec, nr_pages, *ppos);
nr_pages          423 fs/splice.c    	for (i = 0; i < nr_pages; i++)
nr_pages          199 fs/udf/inode.c 			struct list_head *pages, unsigned nr_pages)
nr_pages          201 fs/udf/inode.c 	return mpage_readpages(mapping, pages, nr_pages, udf_get_block);
nr_pages         1172 fs/xfs/xfs_aops.c 	unsigned		nr_pages)
nr_pages         1174 fs/xfs/xfs_aops.c 	trace_xfs_vm_readpages(mapping->host, nr_pages);
nr_pages         1175 fs/xfs/xfs_aops.c 	return iomap_readpages(mapping, pages, nr_pages, &xfs_iomap_ops);
nr_pages         1269 fs/xfs/xfs_buf.c 	int		nr_pages;
nr_pages         1293 fs/xfs/xfs_buf.c 	nr_pages = min(total_nr_pages, BIO_MAX_PAGES);
nr_pages         1295 fs/xfs/xfs_buf.c 	bio = bio_alloc(GFP_NOIO, nr_pages);
nr_pages         1302 fs/xfs/xfs_buf.c 	for (; size && nr_pages; nr_pages--, page_index++) {
nr_pages         1201 fs/xfs/xfs_trace.h 	TP_PROTO(struct inode *inode, int nr_pages),
nr_pages         1202 fs/xfs/xfs_trace.h 	TP_ARGS(inode, nr_pages),
nr_pages         1206 fs/xfs/xfs_trace.h 		__field(int, nr_pages)
nr_pages         1211 fs/xfs/xfs_trace.h 		__entry->nr_pages = nr_pages;
nr_pages         1216 fs/xfs/xfs_trace.h 		  __entry->nr_pages)
nr_pages         1221 fs/xfs/xfs_trace.h 	TP_PROTO(struct inode *inode, int nr_pages), \
nr_pages         1222 fs/xfs/xfs_trace.h 	TP_ARGS(inode, nr_pages))
nr_pages           93 include/drm/drm_prime.h struct sg_table *drm_prime_pages_to_sg(struct page **pages, unsigned int nr_pages);
nr_pages          517 include/linux/ceph/osd_client.h 			       struct page **pages, int nr_pages,
nr_pages          527 include/linux/ceph/osd_client.h 				struct page **pages, int nr_pages);
nr_pages          804 include/linux/compat.h asmlinkage long compat_sys_move_pages(pid_t pid, compat_ulong_t nr_pages,
nr_pages          214 include/linux/coresight.h 			      int nr_pages, bool overwrite);
nr_pages          204 include/linux/dax.h long dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, long nr_pages,
nr_pages          140 include/linux/device-mapper.h 		long nr_pages, void **kaddr, pfn_t *pfn);
nr_pages          452 include/linux/edac.h 	u32 nr_pages;			/* number of pages on this dimm */
nr_pages          377 include/linux/fs.h 			struct list_head *pages, unsigned nr_pages);
nr_pages          160 include/linux/fscache-cache.h 					      unsigned *nr_pages,
nr_pages          592 include/linux/fscache.h 				unsigned *nr_pages,
nr_pages          599 include/linux/fscache.h 						     nr_pages, end_io_func,
nr_pages          616 include/linux/gfp.h void free_contig_range(unsigned long pfn, unsigned int nr_pages);
nr_pages           54 include/linux/hugetlb_cgroup.h extern int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages,
nr_pages           56 include/linux/hugetlb_cgroup.h extern void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages,
nr_pages           59 include/linux/hugetlb_cgroup.h extern void hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages,
nr_pages           61 include/linux/hugetlb_cgroup.h extern void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages,
nr_pages           85 include/linux/hugetlb_cgroup.h hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages,
nr_pages           92 include/linux/hugetlb_cgroup.h hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages,
nr_pages           99 include/linux/hugetlb_cgroup.h hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, struct page *page)
nr_pages          104 include/linux/hugetlb_cgroup.h hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages,
nr_pages          156 include/linux/iomap.h 		unsigned nr_pages, const struct iomap_ops *ops);
nr_pages          280 include/linux/ioport.h walk_system_ram_range(unsigned long start_pfn, unsigned long nr_pages,
nr_pages          703 include/linux/kvm_host.h 			    struct page **pages, int nr_pages);
nr_pages          537 include/linux/memcontrol.h 		int zid, int nr_pages);
nr_pages         1336 include/linux/memcontrol.h bool mem_cgroup_charge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages);
nr_pages         1337 include/linux/memcontrol.h void mem_cgroup_uncharge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages);
nr_pages         1382 include/linux/memcontrol.h 				 unsigned int nr_pages);
nr_pages           52 include/linux/memory.h 	unsigned long nr_pages;
nr_pages           67 include/linux/memory.h 	unsigned int nr_pages;		/* # pages in range to check */
nr_pages           94 include/linux/memory_hotplug.h extern int zone_grow_waitqueues(struct zone *zone, unsigned long nr_pages);
nr_pages          128 include/linux/memory_hotplug.h extern void __remove_pages(unsigned long start_pfn, unsigned long nr_pages,
nr_pages          132 include/linux/memory_hotplug.h extern int __add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages,
nr_pages          137 include/linux/memory_hotplug.h 		unsigned long nr_pages, struct mhp_restrictions *restrictions)
nr_pages          139 include/linux/memory_hotplug.h 	return __add_pages(nid, start_pfn, nr_pages, restrictions);
nr_pages          142 include/linux/memory_hotplug.h int add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages,
nr_pages          314 include/linux/memory_hotplug.h extern bool is_mem_section_removable(unsigned long pfn, unsigned long nr_pages);
nr_pages          316 include/linux/memory_hotplug.h extern int offline_pages(unsigned long start_pfn, unsigned long nr_pages);
nr_pages          322 include/linux/memory_hotplug.h 					unsigned long nr_pages)
nr_pages          329 include/linux/memory_hotplug.h static inline int offline_pages(unsigned long start_pfn, unsigned long nr_pages)
nr_pages          347 include/linux/memory_hotplug.h 		unsigned long nr_pages, struct vmem_altmap *altmap);
nr_pages          350 include/linux/memory_hotplug.h 				       unsigned long nr_pages);
nr_pages          353 include/linux/memory_hotplug.h 		unsigned long nr_pages, struct vmem_altmap *altmap);
nr_pages          355 include/linux/memory_hotplug.h 		unsigned long pfn, unsigned long nr_pages,
nr_pages          359 include/linux/memory_hotplug.h extern bool allow_online_pfn_range(int nid, unsigned long pfn, unsigned long nr_pages,
nr_pages          362 include/linux/memory_hotplug.h 		unsigned long nr_pages);
nr_pages         1533 include/linux/mm.h 			    unsigned long start, unsigned long nr_pages,
nr_pages         1536 include/linux/mm.h long get_user_pages(unsigned long start, unsigned long nr_pages,
nr_pages         1539 include/linux/mm.h long get_user_pages_locked(unsigned long start, unsigned long nr_pages,
nr_pages         1541 include/linux/mm.h long get_user_pages_unlocked(unsigned long start, unsigned long nr_pages,
nr_pages         1544 include/linux/mm.h int get_user_pages_fast(unsigned long start, int nr_pages,
nr_pages         1594 include/linux/mm.h int get_kernel_pages(const struct kvec *iov, int nr_pages, int write,
nr_pages         1638 include/linux/mm.h int __get_user_pages_fast(unsigned long start, int nr_pages, int write,
nr_pages         2774 include/linux/mm.h 		unsigned long nr_pages, int nid, struct vmem_altmap *altmap);
nr_pages         2795 include/linux/mm.h 				  unsigned long nr_pages);
nr_pages           28 include/linux/mm_inline.h 				int nr_pages)
nr_pages           32 include/linux/mm_inline.h 	__mod_lruvec_state(lruvec, NR_LRU_BASE + lru, nr_pages);
nr_pages           34 include/linux/mm_inline.h 				NR_ZONE_LRU_BASE + lru, nr_pages);
nr_pages           39 include/linux/mm_inline.h 				int nr_pages)
nr_pages           41 include/linux/mm_inline.h 	__update_lru_size(lruvec, lru, zid, nr_pages);
nr_pages           43 include/linux/mm_inline.h 	mem_cgroup_update_lru_size(lruvec, lru, zid, nr_pages);
nr_pages          616 include/linux/mmzone.h 		unsigned long start_pfn, unsigned long nr_pages)
nr_pages          621 include/linux/mmzone.h 	    start_pfn + nr_pages <= zone->zone_start_pfn)
nr_pages         1180 include/linux/mmzone.h void subsection_map_init(unsigned long pfn, unsigned long nr_pages);
nr_pages           18 include/linux/mpage.h 				unsigned nr_pages, get_block_t get_block);
nr_pages           50 include/linux/page_counter.h void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages);
nr_pages           51 include/linux/page_counter.h void page_counter_charge(struct page_counter *counter, unsigned long nr_pages);
nr_pages           53 include/linux/page_counter.h 			     unsigned long nr_pages,
nr_pages           55 include/linux/page_counter.h void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages);
nr_pages           56 include/linux/page_counter.h void page_counter_set_min(struct page_counter *counter, unsigned long nr_pages);
nr_pages           57 include/linux/page_counter.h void page_counter_set_low(struct page_counter *counter, unsigned long nr_pages);
nr_pages           58 include/linux/page_counter.h int page_counter_set_max(struct page_counter *counter, unsigned long nr_pages);
nr_pages           60 include/linux/page_counter.h 			  unsigned long *nr_pages);
nr_pages          352 include/linux/pagemap.h 			pgoff_t end, unsigned int nr_pages,
nr_pages          355 include/linux/pagemap.h 			pgoff_t *start, unsigned int nr_pages,
nr_pages          358 include/linux/pagemap.h 	return find_get_pages_range(mapping, start, (pgoff_t)-1, nr_pages,
nr_pages          362 include/linux/pagemap.h 			       unsigned int nr_pages, struct page **pages);
nr_pages          364 include/linux/pagemap.h 			pgoff_t end, xa_mark_t tag, unsigned int nr_pages,
nr_pages          367 include/linux/pagemap.h 			pgoff_t *index, xa_mark_t tag, unsigned int nr_pages,
nr_pages          371 include/linux/pagemap.h 					nr_pages, pages);
nr_pages          417 include/linux/perf_event.h 					 int nr_pages, bool overwrite);
nr_pages          112 include/linux/scif.h 	int nr_pages;
nr_pages           59 include/linux/splice.h 	int nr_pages;			/* number of populated pages in map */
nr_pages          153 include/linux/swap.h 	pgoff_t nr_pages;
nr_pages          356 include/linux/swap.h 						  unsigned long nr_pages,
nr_pages          363 include/linux/swap.h extern unsigned long shrink_all_memory(unsigned long nr_pages);
nr_pages          396 include/linux/swap.h 		unsigned long nr_pages, sector_t start_block);
nr_pages          473 include/linux/swap.h extern int init_swap_address_space(unsigned int type, unsigned long nr_pages);
nr_pages          661 include/linux/swap.h extern void mem_cgroup_uncharge_swap(swp_entry_t entry, unsigned int nr_pages);
nr_pages          676 include/linux/swap.h 					    unsigned int nr_pages)
nr_pages          894 include/linux/syscalls.h asmlinkage long sys_move_pages(pid_t pid, unsigned long nr_pages,
nr_pages           47 include/linux/vmalloc.h 	unsigned int		nr_pages;
nr_pages          374 include/linux/vmstat.h static inline void __mod_zone_freepage_state(struct zone *zone, int nr_pages,
nr_pages          377 include/linux/vmstat.h 	__mod_zone_page_state(zone, NR_FREE_PAGES, nr_pages);
nr_pages          379 include/linux/vmstat.h 		__mod_zone_page_state(zone, NR_FREE_CMA_PAGES, nr_pages);
nr_pages          220 include/linux/writeback.h int cgroup_writeback_by_id(u64 bdi_id, int memcg_id, unsigned long nr_pages,
nr_pages           63 include/media/videobuf-dma-sg.h 	int                 nr_pages;
nr_pages          353 include/trace/events/writeback.h 		__field(long, nr_pages)
nr_pages          364 include/trace/events/writeback.h 		__entry->nr_pages = work->nr_pages;
nr_pages          377 include/trace/events/writeback.h 		  __entry->nr_pages,
nr_pages           98 include/xen/arm/page.h 		unsigned long nr_pages);
nr_pages           27 include/xen/balloon.h int alloc_xenballooned_pages(int nr_pages, struct page **pages);
nr_pages           28 include/xen/balloon.h void free_xenballooned_pages(int nr_pages, struct page **pages);
nr_pages          198 include/xen/grant_table.h int gnttab_alloc_pages(int nr_pages, struct page **pages);
nr_pages          199 include/xen/grant_table.h void gnttab_free_pages(int nr_pages, struct page **pages);
nr_pages          208 include/xen/grant_table.h 	int nr_pages;
nr_pages          219 include/xen/grant_table.h int gnttab_pages_set_private(int nr_pages, struct page **pages);
nr_pages          220 include/xen/grant_table.h void gnttab_pages_clear_private(int nr_pages, struct page **pages);
nr_pages          638 include/xen/interface/xen.h 	unsigned long nr_pages;     /* Total pages allocated to this domain.  */
nr_pages          205 include/xen/xenbus.h 		      unsigned int nr_pages, grant_ref_t *grefs);
nr_pages          934 kernel/dma/debug.c 	int i, nr_pages;
nr_pages          949 kernel/dma/debug.c 	nr_pages = DIV_ROUND_UP(nr_prealloc_entries, DMA_DEBUG_DYNAMIC_ENTRIES);
nr_pages          950 kernel/dma/debug.c 	for (i = 0; i < nr_pages; ++i)
nr_pages         1110 kernel/dma/debug.c 		for (i = 0; i < stack_vm_area->nr_pages; i++) {
nr_pages          126 kernel/dma/remap.c 	unsigned long nr_pages = atomic_pool_size >> PAGE_SHIFT;
nr_pages          132 kernel/dma/remap.c 		page = dma_alloc_from_contiguous(NULL, nr_pages,
nr_pages          167 kernel/dma/remap.c 	if (!dma_release_from_contiguous(NULL, page, nr_pages))
nr_pages         5224 kernel/events/core.c 		if (!rb || !rb->nr_pages) {
nr_pages         5712 kernel/events/core.c 	unsigned long nr_pages;
nr_pages         5730 kernel/events/core.c 		nr_pages = (vma_size / PAGE_SIZE) - 1;
nr_pages         5742 kernel/events/core.c 		nr_pages = vma_size / PAGE_SIZE;
nr_pages         5764 kernel/events/core.c 		if (aux_size != vma_size || aux_size != nr_pages * PAGE_SIZE)
nr_pages         5768 kernel/events/core.c 		if (rb_has_aux(rb) && rb->aux_nr_pages != nr_pages)
nr_pages         5771 kernel/events/core.c 		if (!is_power_of_2(nr_pages))
nr_pages         5784 kernel/events/core.c 		user_extra = nr_pages;
nr_pages         5793 kernel/events/core.c 	if (nr_pages != 0 && !is_power_of_2(nr_pages))
nr_pages         5796 kernel/events/core.c 	if (vma_size != PAGE_SIZE * (1 + nr_pages))
nr_pages         5803 kernel/events/core.c 		if (event->rb->nr_pages != nr_pages) {
nr_pages         5821 kernel/events/core.c 	user_extra = nr_pages + 1;
nr_pages         5872 kernel/events/core.c 		rb = rb_alloc(nr_pages,
nr_pages         5890 kernel/events/core.c 		ret = rb_alloc_aux(rb, event, vma->vm_pgoff, nr_pages,
nr_pages           20 kernel/events/internal.h 	int				nr_pages;	/* nr of data pages  */
nr_pages           72 kernel/events/internal.h 	if (!pause && rb->nr_pages)
nr_pages           79 kernel/events/internal.h rb_alloc(int nr_pages, long watermark, int cpu, int flags);
nr_pages           82 kernel/events/internal.h 			pgoff_t pgoff, int nr_pages, long watermark, int flags);
nr_pages          120 kernel/events/internal.h 	return rb->nr_pages << (PAGE_SHIFT + page_order(rb));
nr_pages          146 kernel/events/internal.h 			handle->page &= rb->nr_pages - 1;		\
nr_pages          174 kernel/events/ring_buffer.c 		if (rb->nr_pages)
nr_pages          234 kernel/events/ring_buffer.c 	handle->page = (offset >> page_shift) & (rb->nr_pages - 1);
nr_pages          328 kernel/events/ring_buffer.c 	if (!rb->nr_pages)
nr_pages          630 kernel/events/ring_buffer.c 		 pgoff_t pgoff, int nr_pages, long watermark, int flags)
nr_pages          643 kernel/events/ring_buffer.c 	max_order = ilog2(nr_pages);
nr_pages          656 kernel/events/ring_buffer.c 	rb->aux_pages = kcalloc_node(nr_pages, sizeof(void *), GFP_KERNEL,
nr_pages          662 kernel/events/ring_buffer.c 	for (rb->aux_nr_pages = 0; rb->aux_nr_pages < nr_pages;) {
nr_pages          666 kernel/events/ring_buffer.c 		order = min(max_order, ilog2(nr_pages - rb->aux_nr_pages));
nr_pages          690 kernel/events/ring_buffer.c 	rb->aux_priv = event->pmu->setup_aux(event, rb->aux_pages, nr_pages,
nr_pages          709 kernel/events/ring_buffer.c 		rb->aux_watermark = nr_pages << (PAGE_SHIFT - 1);
nr_pages          735 kernel/events/ring_buffer.c 	if (pgoff > rb->nr_pages)
nr_pages          757 kernel/events/ring_buffer.c struct ring_buffer *rb_alloc(int nr_pages, long watermark, int cpu, int flags)
nr_pages          764 kernel/events/ring_buffer.c 	size += nr_pages * sizeof(void *);
nr_pages          777 kernel/events/ring_buffer.c 	for (i = 0; i < nr_pages; i++) {
nr_pages          783 kernel/events/ring_buffer.c 	rb->nr_pages = nr_pages;
nr_pages          815 kernel/events/ring_buffer.c 	for (i = 0; i < rb->nr_pages; i++)
nr_pages          823 kernel/events/ring_buffer.c 	return rb->nr_pages << page_order(rb);
nr_pages          866 kernel/events/ring_buffer.c struct ring_buffer *rb_alloc(int nr_pages, long watermark, int cpu, int flags)
nr_pages          881 kernel/events/ring_buffer.c 	all_buf = vmalloc_user((nr_pages + 1) * PAGE_SIZE);
nr_pages          887 kernel/events/ring_buffer.c 	if (nr_pages) {
nr_pages          888 kernel/events/ring_buffer.c 		rb->nr_pages = 1;
nr_pages          889 kernel/events/ring_buffer.c 		rb->page_order = ilog2(nr_pages);
nr_pages          380 kernel/fork.c  		BUG_ON(vm->nr_pages != THREAD_SIZE / PAGE_SIZE);
nr_pages          760 kernel/fork.c  	unsigned long nr_pages = totalram_pages();
nr_pages          766 kernel/fork.c  	if (fls64(nr_pages) + fls64(PAGE_SIZE) > 64)
nr_pages          769 kernel/fork.c  		threads = div64_u64((u64) nr_pages * (u64) PAGE_SIZE,
nr_pages          153 kernel/kexec_core.c 	unsigned long nr_pages = totalram_pages();
nr_pages          219 kernel/kexec_core.c 		if (PAGE_COUNT(image->segment[i].memsz) > nr_pages / 2)
nr_pages          225 kernel/kexec_core.c 	if (total_pages > nr_pages / 2)
nr_pages          240 kernel/power/hibernate.c 		      unsigned nr_pages, char *msg)
nr_pages          253 kernel/power/hibernate.c 	k = nr_pages * (PAGE_SIZE / 1024);
nr_pages         1521 kernel/power/snapshot.c static unsigned long preallocate_image_pages(unsigned long nr_pages, gfp_t mask)
nr_pages         1525 kernel/power/snapshot.c 	while (nr_pages > 0) {
nr_pages         1536 kernel/power/snapshot.c 		nr_pages--;
nr_pages         1543 kernel/power/snapshot.c static unsigned long preallocate_image_memory(unsigned long nr_pages,
nr_pages         1552 kernel/power/snapshot.c 	if (nr_pages < alloc)
nr_pages         1553 kernel/power/snapshot.c 		alloc = nr_pages;
nr_pages         1559 kernel/power/snapshot.c static unsigned long preallocate_image_highmem(unsigned long nr_pages)
nr_pages         1561 kernel/power/snapshot.c 	return preallocate_image_pages(nr_pages, GFP_IMAGE | __GFP_HIGHMEM);
nr_pages         1574 kernel/power/snapshot.c static unsigned long preallocate_highmem_fraction(unsigned long nr_pages,
nr_pages         1578 kernel/power/snapshot.c 	unsigned long alloc = __fraction(nr_pages, highmem, total);
nr_pages         1583 kernel/power/snapshot.c static inline unsigned long preallocate_image_highmem(unsigned long nr_pages)
nr_pages         1588 kernel/power/snapshot.c static inline unsigned long preallocate_highmem_fraction(unsigned long nr_pages,
nr_pages         1875 kernel/power/snapshot.c static int enough_free_mem(unsigned int nr_pages, unsigned int nr_highmem)
nr_pages         1884 kernel/power/snapshot.c 	nr_pages += count_pages_for_highmem(nr_highmem);
nr_pages         1886 kernel/power/snapshot.c 		 nr_pages, PAGES_FOR_IO, free);
nr_pages         1888 kernel/power/snapshot.c 	return free > nr_pages + PAGES_FOR_IO;
nr_pages         1946 kernel/power/snapshot.c 			unsigned int nr_pages, unsigned int nr_highmem)
nr_pages         1953 kernel/power/snapshot.c 			nr_pages += alloc_highmem_pages(copy_bm, nr_highmem);
nr_pages         1956 kernel/power/snapshot.c 	if (nr_pages > alloc_normal) {
nr_pages         1957 kernel/power/snapshot.c 		nr_pages -= alloc_normal;
nr_pages         1958 kernel/power/snapshot.c 		while (nr_pages-- > 0) {
nr_pages         1977 kernel/power/snapshot.c 	unsigned int nr_pages, nr_highmem;
nr_pages         1982 kernel/power/snapshot.c 	nr_pages = count_data_pages();
nr_pages         1984 kernel/power/snapshot.c 	pr_info("Need to copy %u pages\n", nr_pages + nr_highmem);
nr_pages         1986 kernel/power/snapshot.c 	if (!enough_free_mem(nr_pages, nr_highmem)) {
nr_pages         1991 kernel/power/snapshot.c 	if (swsusp_alloc(&copy_bm, nr_pages, nr_highmem)) {
nr_pages         2009 kernel/power/snapshot.c 	nr_pages += nr_highmem;
nr_pages         2010 kernel/power/snapshot.c 	nr_copy_pages = nr_pages;
nr_pages         2011 kernel/power/snapshot.c 	nr_meta_pages = DIV_ROUND_UP(nr_pages * sizeof(long), PAGE_SIZE);
nr_pages         2013 kernel/power/snapshot.c 	pr_info("Hibernation image created (%d pages copied)\n", nr_pages);
nr_pages         2464 kernel/power/snapshot.c 	unsigned int nr_pages, nr_highmem;
nr_pages         2495 kernel/power/snapshot.c 	nr_pages = nr_copy_pages - nr_highmem - allocated_unsafe_pages;
nr_pages         2496 kernel/power/snapshot.c 	nr_pages = DIV_ROUND_UP(nr_pages, PBES_PER_LINKED_PAGE);
nr_pages         2497 kernel/power/snapshot.c 	while (nr_pages > 0) {
nr_pages         2505 kernel/power/snapshot.c 		nr_pages--;
nr_pages         2508 kernel/power/snapshot.c 	nr_pages = nr_copy_pages - nr_highmem - allocated_unsafe_pages;
nr_pages         2509 kernel/power/snapshot.c 	while (nr_pages > 0) {
nr_pages         2523 kernel/power/snapshot.c 		nr_pages--;
nr_pages          536 kernel/power/swap.c 	int nr_pages;
nr_pages          549 kernel/power/swap.c 	nr_pages = 0;
nr_pages          558 kernel/power/swap.c 		if (!(nr_pages % m))
nr_pages          560 kernel/power/swap.c 				nr_pages / m * 10);
nr_pages          561 kernel/power/swap.c 		nr_pages++;
nr_pages          672 kernel/power/swap.c 	int nr_pages;
nr_pages          767 kernel/power/swap.c 	nr_pages = 0;
nr_pages          782 kernel/power/swap.c 				if (!(nr_pages % m))
nr_pages          784 kernel/power/swap.c 						nr_pages / m * 10);
nr_pages          785 kernel/power/swap.c 				nr_pages++;
nr_pages          880 kernel/power/swap.c static int enough_swap(unsigned int nr_pages)
nr_pages          887 kernel/power/swap.c 	required = PAGES_FOR_IO + nr_pages;
nr_pages         1059 kernel/power/swap.c 	unsigned nr_pages;
nr_pages         1068 kernel/power/swap.c 	nr_pages = 0;
nr_pages         1081 kernel/power/swap.c 		if (!(nr_pages % m))
nr_pages         1083 kernel/power/swap.c 				nr_pages / m * 10);
nr_pages         1084 kernel/power/swap.c 		nr_pages++;
nr_pages         1164 kernel/power/swap.c 	unsigned nr_pages;
nr_pages         1284 kernel/power/swap.c 	nr_pages = 0;
nr_pages         1408 kernel/power/swap.c 				if (!(nr_pages % m))
nr_pages         1410 kernel/power/swap.c 						nr_pages / m * 10);
nr_pages         1411 kernel/power/swap.c 				nr_pages++;
nr_pages         1205 kernel/relay.c 	unsigned int pidx, poff, total_len, subbuf_pages, nr_pages;
nr_pages         1218 kernel/relay.c 		.nr_pages = 0,
nr_pages         1240 kernel/relay.c 	nr_pages = min_t(unsigned int, subbuf_pages, spd.nr_pages_max);
nr_pages         1242 kernel/relay.c 	for (total_len = 0; spd.nr_pages < nr_pages; spd.nr_pages++) {
nr_pages         1252 kernel/relay.c 		spd.pages[spd.nr_pages] = rbuf->page_array[pidx];
nr_pages         1253 kernel/relay.c 		spd.partial[spd.nr_pages].offset = poff;
nr_pages         1260 kernel/relay.c 		spd.partial[spd.nr_pages].len = this_len;
nr_pages         1261 kernel/relay.c 		spd.partial[spd.nr_pages].private = private;
nr_pages         1269 kernel/relay.c 			spd.nr_pages++;
nr_pages         1275 kernel/relay.c 	if (!spd.nr_pages)
nr_pages          475 kernel/resource.c int walk_system_ram_range(unsigned long start_pfn, unsigned long nr_pages,
nr_pages          485 kernel/resource.c 	end = ((u64)(start_pfn + nr_pages) << PAGE_SHIFT) - 1;
nr_pages          501 kernel/resource.c static int __is_ram(unsigned long pfn, unsigned long nr_pages, void *arg)
nr_pages          451 kernel/trace/ring_buffer.c 	unsigned long			nr_pages;
nr_pages          523 kernel/trace/ring_buffer.c 	return buffer->buffers[cpu]->nr_pages;
nr_pages          643 kernel/trace/ring_buffer.c 			size_t nr_pages;
nr_pages          651 kernel/trace/ring_buffer.c 			nr_pages = cpu_buffer->nr_pages;
nr_pages          658 kernel/trace/ring_buffer.c 			    (!nr_pages || (dirty * 100) > full * nr_pages))
nr_pages         1189 kernel/trace/ring_buffer.c static int __rb_allocate_pages(long nr_pages, struct list_head *pages, int cpu)
nr_pages         1204 kernel/trace/ring_buffer.c 	if (i < nr_pages)
nr_pages         1225 kernel/trace/ring_buffer.c 	for (i = 0; i < nr_pages; i++) {
nr_pages         1261 kernel/trace/ring_buffer.c 			     unsigned long nr_pages)
nr_pages         1265 kernel/trace/ring_buffer.c 	WARN_ON(!nr_pages);
nr_pages         1267 kernel/trace/ring_buffer.c 	if (__rb_allocate_pages(nr_pages, &pages, cpu_buffer->cpu))
nr_pages         1278 kernel/trace/ring_buffer.c 	cpu_buffer->nr_pages = nr_pages;
nr_pages         1286 kernel/trace/ring_buffer.c rb_allocate_cpu_buffer(struct ring_buffer *buffer, long nr_pages, int cpu)
nr_pages         1326 kernel/trace/ring_buffer.c 	ret = rb_allocate_pages(cpu_buffer, nr_pages);
nr_pages         1381 kernel/trace/ring_buffer.c 	long nr_pages;
nr_pages         1395 kernel/trace/ring_buffer.c 	nr_pages = DIV_ROUND_UP(size, BUF_PAGE_SIZE);
nr_pages         1404 kernel/trace/ring_buffer.c 	if (nr_pages < 2)
nr_pages         1405 kernel/trace/ring_buffer.c 		nr_pages = 2;
nr_pages         1417 kernel/trace/ring_buffer.c 	buffer->buffers[cpu] = rb_allocate_cpu_buffer(buffer, nr_pages, cpu);
nr_pages         1495 kernel/trace/ring_buffer.c rb_remove_pages(struct ring_buffer_per_cpu *cpu_buffer, unsigned long nr_pages)
nr_pages         1531 kernel/trace/ring_buffer.c 	for (nr_removed = 0; nr_removed < nr_pages; nr_removed++) {
nr_pages         1694 kernel/trace/ring_buffer.c 		cpu_buffer->nr_pages += cpu_buffer->nr_pages_to_update;
nr_pages         1719 kernel/trace/ring_buffer.c 	unsigned long nr_pages;
nr_pages         1733 kernel/trace/ring_buffer.c 	nr_pages = DIV_ROUND_UP(size, BUF_PAGE_SIZE);
nr_pages         1736 kernel/trace/ring_buffer.c 	if (nr_pages < 2)
nr_pages         1737 kernel/trace/ring_buffer.c 		nr_pages = 2;
nr_pages         1739 kernel/trace/ring_buffer.c 	size = nr_pages * BUF_PAGE_SIZE;
nr_pages         1757 kernel/trace/ring_buffer.c 			cpu_buffer->nr_pages_to_update = nr_pages -
nr_pages         1758 kernel/trace/ring_buffer.c 							cpu_buffer->nr_pages;
nr_pages         1817 kernel/trace/ring_buffer.c 		if (nr_pages == cpu_buffer->nr_pages)
nr_pages         1820 kernel/trace/ring_buffer.c 		cpu_buffer->nr_pages_to_update = nr_pages -
nr_pages         1821 kernel/trace/ring_buffer.c 						cpu_buffer->nr_pages;
nr_pages         2474 kernel/trace/ring_buffer.c 	max_count = cpu_buffer->nr_pages * 100;
nr_pages         2614 kernel/trace/ring_buffer.c 	size_t nr_pages;
nr_pages         2642 kernel/trace/ring_buffer.c 	nr_pages = cpu_buffer->nr_pages;
nr_pages         2644 kernel/trace/ring_buffer.c 	if (full && nr_pages && (dirty * 100) <= full * nr_pages)
nr_pages         4346 kernel/trace/ring_buffer.c 	return BUF_PAGE_SIZE * buffer->buffers[cpu]->nr_pages;
nr_pages         4528 kernel/trace/ring_buffer.c 	if (cpu_buffer_a->nr_pages != cpu_buffer_b->nr_pages)
nr_pages         4874 kernel/trace/ring_buffer.c 	unsigned long nr_pages;
nr_pages         4880 kernel/trace/ring_buffer.c 	nr_pages = 0;
nr_pages         4885 kernel/trace/ring_buffer.c 		if (nr_pages == 0)
nr_pages         4886 kernel/trace/ring_buffer.c 			nr_pages = buffer->buffers[cpu_i]->nr_pages;
nr_pages         4887 kernel/trace/ring_buffer.c 		if (nr_pages != buffer->buffers[cpu_i]->nr_pages) {
nr_pages         4894 kernel/trace/ring_buffer.c 		nr_pages = 2;
nr_pages         4896 kernel/trace/ring_buffer.c 		rb_allocate_cpu_buffer(buffer, nr_pages, cpu);
nr_pages         6175 kernel/trace/trace.c 		.nr_pages	= 0, /* This gets updated below. */
nr_pages         6234 kernel/trace/trace.c 	spd.nr_pages = i;
nr_pages         7482 kernel/trace/trace.c 		spd.nr_pages++;
nr_pages         7489 kernel/trace/trace.c 	spd.nr_pages = i;
nr_pages         7492 kernel/trace/trace.c 	if (!spd.nr_pages) {
nr_pages           71 mm/compaction.c 	unsigned int i, order, nr_pages;
nr_pages           79 mm/compaction.c 		nr_pages = 1 << order;
nr_pages           85 mm/compaction.c 		for (i = 0; i < nr_pages; i++) {
nr_pages          509 mm/filemap.c   	int nr_pages;
nr_pages          518 mm/filemap.c   		nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index,
nr_pages          520 mm/filemap.c   		if (!nr_pages)
nr_pages          523 mm/filemap.c   		for (i = 0; i < nr_pages; i++) {
nr_pages         1792 mm/filemap.c   			      pgoff_t end, unsigned int nr_pages,
nr_pages         1799 mm/filemap.c   	if (unlikely(!nr_pages))
nr_pages         1818 mm/filemap.c   		if (++ret == nr_pages) {
nr_pages         1858 mm/filemap.c   			       unsigned int nr_pages, struct page **pages)
nr_pages         1864 mm/filemap.c   	if (unlikely(!nr_pages))
nr_pages         1886 mm/filemap.c   		if (++ret == nr_pages)
nr_pages         1914 mm/filemap.c   			pgoff_t end, xa_mark_t tag, unsigned int nr_pages,
nr_pages         1921 mm/filemap.c   	if (unlikely(!nr_pages))
nr_pages         1944 mm/filemap.c   		if (++ret == nr_pages) {
nr_pages          789 mm/gup.c       		unsigned long start, unsigned long nr_pages,
nr_pages          797 mm/gup.c       	if (!nr_pages)
nr_pages          838 mm/gup.c       						&start, &nr_pages, i,
nr_pages          898 mm/gup.c       		if (page_increm > nr_pages)
nr_pages          899 mm/gup.c       			page_increm = nr_pages;
nr_pages          902 mm/gup.c       		nr_pages -= page_increm;
nr_pages          903 mm/gup.c       	} while (nr_pages);
nr_pages         1016 mm/gup.c       						unsigned long nr_pages,
nr_pages         1038 mm/gup.c       		ret = __get_user_pages(tsk, mm, start, nr_pages, flags, pages,
nr_pages         1047 mm/gup.c       			BUG_ON(ret >= nr_pages);
nr_pages         1051 mm/gup.c       			nr_pages -= ret;
nr_pages         1053 mm/gup.c       			if (!nr_pages)
nr_pages         1089 mm/gup.c       		nr_pages--;
nr_pages         1091 mm/gup.c       		if (!nr_pages)
nr_pages         1165 mm/gup.c       		unsigned long start, unsigned long nr_pages,
nr_pages         1178 mm/gup.c       	return __get_user_pages_locked(tsk, mm, start, nr_pages, pages, vmas,
nr_pages         1207 mm/gup.c       	unsigned long nr_pages = (end - start) / PAGE_SIZE;
nr_pages         1238 mm/gup.c       	return __get_user_pages(current, mm, start, nr_pages, gup_flags,
nr_pages         1333 mm/gup.c       		unsigned long nr_pages, struct page **pages,
nr_pages         1349 mm/gup.c       	for (i = 0; i < nr_pages; i++) {
nr_pages         1377 mm/gup.c       static bool check_dax_vmas(struct vm_area_struct **vmas, long nr_pages)
nr_pages         1382 mm/gup.c       	for (i = 0; i < nr_pages; i++) {
nr_pages         1451 mm/gup.c       					unsigned long nr_pages,
nr_pages         1463 mm/gup.c       	for (i = 0; i < nr_pages;) {
nr_pages         1503 mm/gup.c       		for (i = 0; i < nr_pages; i++)
nr_pages         1522 mm/gup.c       		nr_pages = __get_user_pages_locked(tsk, mm, start, nr_pages,
nr_pages         1526 mm/gup.c       		if ((nr_pages > 0) && migrate_allow) {
nr_pages         1532 mm/gup.c       	return nr_pages;
nr_pages         1538 mm/gup.c       					unsigned long nr_pages,
nr_pages         1543 mm/gup.c       	return nr_pages;
nr_pages         1554 mm/gup.c       				  unsigned long nr_pages,
nr_pages         1568 mm/gup.c       			vmas_tmp = kcalloc(nr_pages,
nr_pages         1577 mm/gup.c       	rc = __get_user_pages_locked(tsk, mm, start, nr_pages, pages,
nr_pages         1605 mm/gup.c       						  unsigned long nr_pages,
nr_pages         1610 mm/gup.c       	return __get_user_pages_locked(tsk, mm, start, nr_pages, pages, vmas,
nr_pages         1622 mm/gup.c       long get_user_pages(unsigned long start, unsigned long nr_pages,
nr_pages         1626 mm/gup.c       	return __gup_longterm_locked(current, current->mm, start, nr_pages,
nr_pages         1652 mm/gup.c       long get_user_pages_locked(unsigned long start, unsigned long nr_pages,
nr_pages         1665 mm/gup.c       	return __get_user_pages_locked(current, current->mm, start, nr_pages,
nr_pages         1686 mm/gup.c       long get_user_pages_unlocked(unsigned long start, unsigned long nr_pages,
nr_pages         1703 mm/gup.c       	ret = __get_user_pages_locked(current, mm, start, nr_pages, pages, NULL,
nr_pages         2339 mm/gup.c       int __get_user_pages_fast(unsigned long start, int nr_pages, int write,
nr_pages         2347 mm/gup.c       	len = (unsigned long) nr_pages << PAGE_SHIFT;
nr_pages         2384 mm/gup.c       static int __gup_longterm_unlocked(unsigned long start, int nr_pages,
nr_pages         2396 mm/gup.c       					    start, nr_pages,
nr_pages         2400 mm/gup.c       		ret = get_user_pages_unlocked(start, nr_pages,
nr_pages         2423 mm/gup.c       int get_user_pages_fast(unsigned long start, int nr_pages,
nr_pages         2435 mm/gup.c       	len = (unsigned long) nr_pages << PAGE_SHIFT;
nr_pages         2458 mm/gup.c       	if (nr < nr_pages) {
nr_pages         2463 mm/gup.c       		ret = __gup_longterm_unlocked(start, nr_pages - nr,
nr_pages           26 mm/gup_benchmark.c 	unsigned long i, nr_pages, addr, next;
nr_pages           34 mm/gup_benchmark.c 	nr_pages = gup->size / PAGE_SIZE;
nr_pages           35 mm/gup_benchmark.c 	pages = kvcalloc(nr_pages, sizeof(void *), GFP_KERNEL);
nr_pages           82 mm/gup_benchmark.c 	for (i = 0; i < nr_pages; i++) {
nr_pages         1054 mm/hugetlb.c   	int nr_pages = 1 << order;
nr_pages         1058 mm/hugetlb.c   	for (i = 1; i < nr_pages; i++, p = mem_map_next(p, page, i)) {
nr_pages         1074 mm/hugetlb.c   				unsigned long nr_pages, gfp_t gfp_mask)
nr_pages         1076 mm/hugetlb.c   	unsigned long end_pfn = start_pfn + nr_pages;
nr_pages         1082 mm/hugetlb.c   			unsigned long start_pfn, unsigned long nr_pages)
nr_pages         1084 mm/hugetlb.c   	unsigned long i, end_pfn = start_pfn + nr_pages;
nr_pages         1109 mm/hugetlb.c   			unsigned long start_pfn, unsigned long nr_pages)
nr_pages         1111 mm/hugetlb.c   	unsigned long last_pfn = start_pfn + nr_pages - 1;
nr_pages         1119 mm/hugetlb.c   	unsigned long nr_pages = 1 << order;
nr_pages         1129 mm/hugetlb.c   		pfn = ALIGN(zone->zone_start_pfn, nr_pages);
nr_pages         1130 mm/hugetlb.c   		while (zone_spans_last_pfn(zone, pfn, nr_pages)) {
nr_pages         1131 mm/hugetlb.c   			if (pfn_range_valid_gigantic(zone, pfn, nr_pages)) {
nr_pages         1140 mm/hugetlb.c   				ret = __alloc_gigantic_page(pfn, nr_pages, gfp_mask);
nr_pages         1145 mm/hugetlb.c   			pfn += nr_pages;
nr_pages         1384 mm/hugetlb.c   	int nr_pages = 1 << order;
nr_pages         1391 mm/hugetlb.c   	for (i = 1; i < nr_pages; i++, p = mem_map_next(p, page, i)) {
nr_pages         1917 mm/hugetlb.c   	unsigned long nr_pages;
nr_pages         1927 mm/hugetlb.c   	nr_pages = min(unused_resv_pages, h->surplus_huge_pages);
nr_pages         1941 mm/hugetlb.c   	while (nr_pages--) {
nr_pages         4389 mm/hugetlb.c   			 unsigned long *position, unsigned long *nr_pages,
nr_pages         4394 mm/hugetlb.c   	unsigned long remainder = *nr_pages;
nr_pages         4481 mm/hugetlb.c   				*nr_pages = 0;
nr_pages         4534 mm/hugetlb.c   	*nr_pages = remainder;
nr_pages          128 mm/hugetlb_cgroup.c 	unsigned int nr_pages;
nr_pages          142 mm/hugetlb_cgroup.c 	nr_pages = compound_nr(page);
nr_pages          146 mm/hugetlb_cgroup.c 		page_counter_charge(&parent->hugepage[idx], nr_pages);
nr_pages          150 mm/hugetlb_cgroup.c 	page_counter_cancel(counter, nr_pages);
nr_pages          181 mm/hugetlb_cgroup.c int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages,
nr_pages          205 mm/hugetlb_cgroup.c 	if (!page_counter_try_charge(&h_cg->hugepage[idx], nr_pages, &counter))
nr_pages          214 mm/hugetlb_cgroup.c void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages,
nr_pages          228 mm/hugetlb_cgroup.c void hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages,
nr_pages          240 mm/hugetlb_cgroup.c 	page_counter_uncharge(&h_cg->hugepage[idx], nr_pages);
nr_pages          244 mm/hugetlb_cgroup.c void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages,
nr_pages          253 mm/hugetlb_cgroup.c 	page_counter_uncharge(&h_cg->hugepage[idx], nr_pages);
nr_pages          292 mm/hugetlb_cgroup.c 	unsigned long nr_pages;
nr_pages          299 mm/hugetlb_cgroup.c 	ret = page_counter_memparse(buf, "-1", &nr_pages);
nr_pages          304 mm/hugetlb_cgroup.c 	nr_pages = round_down(nr_pages, 1 << huge_page_order(&hstates[idx]));
nr_pages          309 mm/hugetlb_cgroup.c 		ret = page_counter_set_max(&h_cg->hugepage[idx], nr_pages);
nr_pages          330 mm/internal.h  		int nr_pages = hpage_nr_pages(page);
nr_pages          333 mm/internal.h  		__mod_zone_page_state(page_zone(page), NR_MLOCK, -nr_pages);
nr_pages          335 mm/internal.h  		__mod_zone_page_state(page_zone(newpage), NR_MLOCK, nr_pages);
nr_pages          680 mm/kasan/common.c 	nr_shadow_pages = mem_data->nr_pages >> KASAN_SHADOW_SCALE_SHIFT;
nr_pages          686 mm/kasan/common.c 	if (WARN_ON(mem_data->nr_pages % KASAN_SHADOW_SCALE_SIZE) ||
nr_pages         2822 mm/ksm.c       				      mn->start_pfn + mn->nr_pages);
nr_pages         2888 mm/ksm.c       	unsigned long nr_pages;
nr_pages         2890 mm/ksm.c       	err = kstrtoul(buf, 10, &nr_pages);
nr_pages         2891 mm/ksm.c       	if (err || nr_pages > UINT_MAX)
nr_pages         2894 mm/ksm.c       	ksm_thread_pages_to_scan = nr_pages;
nr_pages          588 mm/memcontrol.c 	unsigned long nr_pages = page_counter_read(&memcg->memory);
nr_pages          592 mm/memcontrol.c 	if (nr_pages > soft_limit)
nr_pages          593 mm/memcontrol.c 		excess = nr_pages - soft_limit;
nr_pages          847 mm/memcontrol.c 					 bool compound, int nr_pages)
nr_pages          854 mm/memcontrol.c 		__mod_memcg_state(memcg, MEMCG_RSS, nr_pages);
nr_pages          856 mm/memcontrol.c 		__mod_memcg_state(memcg, MEMCG_CACHE, nr_pages);
nr_pages          858 mm/memcontrol.c 			__mod_memcg_state(memcg, NR_SHMEM, nr_pages);
nr_pages          863 mm/memcontrol.c 		__mod_memcg_state(memcg, MEMCG_RSS_HUGE, nr_pages);
nr_pages          867 mm/memcontrol.c 	if (nr_pages > 0)
nr_pages          871 mm/memcontrol.c 		nr_pages = -nr_pages; /* for event */
nr_pages          874 mm/memcontrol.c 	__this_cpu_add(memcg->vmstats_percpu->nr_page_events, nr_pages);
nr_pages         1291 mm/memcontrol.c 				int zid, int nr_pages)
nr_pages         1303 mm/memcontrol.c 	if (nr_pages < 0)
nr_pages         1304 mm/memcontrol.c 		*lru_size += nr_pages;
nr_pages         1309 mm/memcontrol.c 		__func__, lruvec, lru, nr_pages, size)) {
nr_pages         1314 mm/memcontrol.c 	if (nr_pages > 0)
nr_pages         1315 mm/memcontrol.c 		*lru_size += nr_pages;
nr_pages         2164 mm/memcontrol.c 	unsigned int nr_pages;
nr_pages         2183 mm/memcontrol.c static bool consume_stock(struct mem_cgroup *memcg, unsigned int nr_pages)
nr_pages         2189 mm/memcontrol.c 	if (nr_pages > MEMCG_CHARGE_BATCH)
nr_pages         2195 mm/memcontrol.c 	if (memcg == stock->cached && stock->nr_pages >= nr_pages) {
nr_pages         2196 mm/memcontrol.c 		stock->nr_pages -= nr_pages;
nr_pages         2212 mm/memcontrol.c 	if (stock->nr_pages) {
nr_pages         2213 mm/memcontrol.c 		page_counter_uncharge(&old->memory, stock->nr_pages);
nr_pages         2215 mm/memcontrol.c 			page_counter_uncharge(&old->memsw, stock->nr_pages);
nr_pages         2216 mm/memcontrol.c 		css_put_many(&old->css, stock->nr_pages);
nr_pages         2217 mm/memcontrol.c 		stock->nr_pages = 0;
nr_pages         2244 mm/memcontrol.c static void refill_stock(struct mem_cgroup *memcg, unsigned int nr_pages)
nr_pages         2256 mm/memcontrol.c 	stock->nr_pages += nr_pages;
nr_pages         2258 mm/memcontrol.c 	if (stock->nr_pages > MEMCG_CHARGE_BATCH)
nr_pages         2289 mm/memcontrol.c 		if (memcg && stock->nr_pages &&
nr_pages         2355 mm/memcontrol.c 			 unsigned int nr_pages,
nr_pages         2362 mm/memcontrol.c 		try_to_free_mem_cgroup_pages(memcg, nr_pages, gfp_mask, true);
nr_pages         2432 mm/memcontrol.c 					  unsigned int nr_pages)
nr_pages         2485 mm/memcontrol.c 	penalty_jiffies = penalty_jiffies * nr_pages / MEMCG_CHARGE_BATCH;
nr_pages         2503 mm/memcontrol.c 	unsigned int nr_pages = current->memcg_nr_pages_over_high;
nr_pages         2506 mm/memcontrol.c 	if (likely(!nr_pages))
nr_pages         2510 mm/memcontrol.c 	reclaim_high(memcg, nr_pages, GFP_KERNEL);
nr_pages         2517 mm/memcontrol.c 	penalty_jiffies = calculate_high_delay(memcg, nr_pages);
nr_pages         2542 mm/memcontrol.c 		      unsigned int nr_pages)
nr_pages         2544 mm/memcontrol.c 	unsigned int batch = max(MEMCG_CHARGE_BATCH, nr_pages);
nr_pages         2556 mm/memcontrol.c 	if (consume_stock(memcg, nr_pages))
nr_pages         2571 mm/memcontrol.c 	if (batch > nr_pages) {
nr_pages         2572 mm/memcontrol.c 		batch = nr_pages;
nr_pages         2611 mm/memcontrol.c 	nr_reclaimed = try_to_free_mem_cgroup_pages(mem_over_limit, nr_pages,
nr_pages         2614 mm/memcontrol.c 	if (mem_cgroup_margin(mem_over_limit) >= nr_pages)
nr_pages         2634 mm/memcontrol.c 	if (nr_reclaimed && nr_pages <= (1 << PAGE_ALLOC_COSTLY_ORDER))
nr_pages         2661 mm/memcontrol.c 		       get_order(nr_pages * PAGE_SIZE));
nr_pages         2680 mm/memcontrol.c 	page_counter_charge(&memcg->memory, nr_pages);
nr_pages         2682 mm/memcontrol.c 		page_counter_charge(&memcg->memsw, nr_pages);
nr_pages         2683 mm/memcontrol.c 	css_get_many(&memcg->css, nr_pages);
nr_pages         2689 mm/memcontrol.c 	if (batch > nr_pages)
nr_pages         2690 mm/memcontrol.c 		refill_stock(memcg, batch - nr_pages);
nr_pages         2717 mm/memcontrol.c static void cancel_charge(struct mem_cgroup *memcg, unsigned int nr_pages)
nr_pages         2722 mm/memcontrol.c 	page_counter_uncharge(&memcg->memory, nr_pages);
nr_pages         2724 mm/memcontrol.c 		page_counter_uncharge(&memcg->memsw, nr_pages);
nr_pages         2726 mm/memcontrol.c 	css_put_many(&memcg->css, nr_pages);
nr_pages         3016 mm/memcontrol.c 	unsigned int nr_pages = 1 << order;
nr_pages         3020 mm/memcontrol.c 	ret = try_charge(memcg, gfp, nr_pages);
nr_pages         3025 mm/memcontrol.c 	    !page_counter_try_charge(&memcg->kmem, nr_pages, &counter)) {
nr_pages         3033 mm/memcontrol.c 			page_counter_charge(&memcg->kmem, nr_pages);
nr_pages         3036 mm/memcontrol.c 		cancel_charge(memcg, nr_pages);
nr_pages         3076 mm/memcontrol.c 				 unsigned int nr_pages)
nr_pages         3079 mm/memcontrol.c 		page_counter_uncharge(&memcg->kmem, nr_pages);
nr_pages         3081 mm/memcontrol.c 	page_counter_uncharge(&memcg->memory, nr_pages);
nr_pages         3083 mm/memcontrol.c 		page_counter_uncharge(&memcg->memsw, nr_pages);
nr_pages         3093 mm/memcontrol.c 	unsigned int nr_pages = 1 << order;
nr_pages         3099 mm/memcontrol.c 	__memcg_kmem_uncharge_memcg(memcg, nr_pages);
nr_pages         3106 mm/memcontrol.c 	css_put_many(&memcg->css, nr_pages);
nr_pages         3682 mm/memcontrol.c 	unsigned long nr_pages;
nr_pages         3686 mm/memcontrol.c 	ret = page_counter_memparse(buf, "-1", &nr_pages);
nr_pages         3698 mm/memcontrol.c 			ret = mem_cgroup_resize_max(memcg, nr_pages, false);
nr_pages         3701 mm/memcontrol.c 			ret = mem_cgroup_resize_max(memcg, nr_pages, true);
nr_pages         3707 mm/memcontrol.c 			ret = memcg_update_kmem_max(memcg, nr_pages);
nr_pages         3710 mm/memcontrol.c 			ret = memcg_update_tcp_max(memcg, nr_pages);
nr_pages         3715 mm/memcontrol.c 		memcg->soft_limit = nr_pages;
nr_pages         5491 mm/memcontrol.c 	unsigned int nr_pages = compound ? hpage_nr_pages(page) : 1;
nr_pages         5520 mm/memcontrol.c 		__mod_lruvec_state(from_vec, NR_FILE_MAPPED, -nr_pages);
nr_pages         5521 mm/memcontrol.c 		__mod_lruvec_state(to_vec, NR_FILE_MAPPED, nr_pages);
nr_pages         5533 mm/memcontrol.c 			__mod_lruvec_state(from_vec, NR_FILE_DIRTY, -nr_pages);
nr_pages         5534 mm/memcontrol.c 			__mod_lruvec_state(to_vec, NR_FILE_DIRTY, nr_pages);
nr_pages         5539 mm/memcontrol.c 		__mod_lruvec_state(from_vec, NR_WRITEBACK, -nr_pages);
nr_pages         5540 mm/memcontrol.c 		__mod_lruvec_state(to_vec, NR_WRITEBACK, nr_pages);
nr_pages         5557 mm/memcontrol.c 	mem_cgroup_charge_statistics(to, page, compound, nr_pages);
nr_pages         5559 mm/memcontrol.c 	mem_cgroup_charge_statistics(from, page, compound, -nr_pages);
nr_pages         6137 mm/memcontrol.c 	unsigned long nr_pages;
nr_pages         6148 mm/memcontrol.c 	nr_pages = page_counter_read(&memcg->memory);
nr_pages         6149 mm/memcontrol.c 	if (nr_pages > high)
nr_pages         6150 mm/memcontrol.c 		try_to_free_mem_cgroup_pages(memcg, nr_pages - high,
nr_pages         6180 mm/memcontrol.c 		unsigned long nr_pages = page_counter_read(&memcg->memory);
nr_pages         6182 mm/memcontrol.c 		if (nr_pages <= max)
nr_pages         6197 mm/memcontrol.c 			if (!try_to_free_mem_cgroup_pages(memcg, nr_pages - max,
nr_pages         6518 mm/memcontrol.c 	unsigned int nr_pages = compound ? hpage_nr_pages(page) : 1;
nr_pages         6551 mm/memcontrol.c 	ret = try_charge(memcg, gfp_mask, nr_pages);
nr_pages         6592 mm/memcontrol.c 	unsigned int nr_pages = compound ? hpage_nr_pages(page) : 1;
nr_pages         6610 mm/memcontrol.c 	mem_cgroup_charge_statistics(memcg, page, compound, nr_pages);
nr_pages         6621 mm/memcontrol.c 		mem_cgroup_uncharge_swap(entry, nr_pages);
nr_pages         6636 mm/memcontrol.c 	unsigned int nr_pages = compound ? hpage_nr_pages(page) : 1;
nr_pages         6648 mm/memcontrol.c 	cancel_charge(memcg, nr_pages);
nr_pages         6669 mm/memcontrol.c 	unsigned long nr_pages = ug->nr_anon + ug->nr_file + ug->nr_kmem;
nr_pages         6673 mm/memcontrol.c 		page_counter_uncharge(&ug->memcg->memory, nr_pages);
nr_pages         6675 mm/memcontrol.c 			page_counter_uncharge(&ug->memcg->memsw, nr_pages);
nr_pages         6687 mm/memcontrol.c 	__this_cpu_add(ug->memcg->vmstats_percpu->nr_page_events, nr_pages);
nr_pages         6692 mm/memcontrol.c 		css_put_many(&ug->memcg->css, nr_pages);
nr_pages         6719 mm/memcontrol.c 		unsigned int nr_pages = 1;
nr_pages         6722 mm/memcontrol.c 			nr_pages = compound_nr(page);
nr_pages         6723 mm/memcontrol.c 			ug->nr_huge += nr_pages;
nr_pages         6726 mm/memcontrol.c 			ug->nr_anon += nr_pages;
nr_pages         6728 mm/memcontrol.c 			ug->nr_file += nr_pages;
nr_pages         6730 mm/memcontrol.c 				ug->nr_shmem += nr_pages;
nr_pages         6819 mm/memcontrol.c 	unsigned int nr_pages;
nr_pages         6843 mm/memcontrol.c 	nr_pages = compound ? hpage_nr_pages(newpage) : 1;
nr_pages         6845 mm/memcontrol.c 	page_counter_charge(&memcg->memory, nr_pages);
nr_pages         6847 mm/memcontrol.c 		page_counter_charge(&memcg->memsw, nr_pages);
nr_pages         6848 mm/memcontrol.c 	css_get_many(&memcg->css, nr_pages);
nr_pages         6853 mm/memcontrol.c 	mem_cgroup_charge_statistics(memcg, newpage, compound, nr_pages);
nr_pages         6898 mm/memcontrol.c bool mem_cgroup_charge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages)
nr_pages         6905 mm/memcontrol.c 		if (page_counter_try_charge(&memcg->tcpmem, nr_pages, &fail)) {
nr_pages         6909 mm/memcontrol.c 		page_counter_charge(&memcg->tcpmem, nr_pages);
nr_pages         6918 mm/memcontrol.c 	mod_memcg_state(memcg, MEMCG_SOCK, nr_pages);
nr_pages         6920 mm/memcontrol.c 	if (try_charge(memcg, gfp_mask, nr_pages) == 0)
nr_pages         6923 mm/memcontrol.c 	try_charge(memcg, gfp_mask|__GFP_NOFAIL, nr_pages);
nr_pages         6932 mm/memcontrol.c void mem_cgroup_uncharge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages)
nr_pages         6935 mm/memcontrol.c 		page_counter_uncharge(&memcg->tcpmem, nr_pages);
nr_pages         6939 mm/memcontrol.c 	mod_memcg_state(memcg, MEMCG_SOCK, -nr_pages);
nr_pages         6941 mm/memcontrol.c 	refill_stock(memcg, nr_pages);
nr_pages         7102 mm/memcontrol.c 	unsigned int nr_pages = hpage_nr_pages(page);
nr_pages         7124 mm/memcontrol.c 	    !page_counter_try_charge(&memcg->swap, nr_pages, &counter)) {
nr_pages         7132 mm/memcontrol.c 	if (nr_pages > 1)
nr_pages         7133 mm/memcontrol.c 		mem_cgroup_id_get_many(memcg, nr_pages - 1);
nr_pages         7134 mm/memcontrol.c 	oldid = swap_cgroup_record(entry, mem_cgroup_id(memcg), nr_pages);
nr_pages         7136 mm/memcontrol.c 	mod_memcg_state(memcg, MEMCG_SWAP, nr_pages);
nr_pages         7146 mm/memcontrol.c void mem_cgroup_uncharge_swap(swp_entry_t entry, unsigned int nr_pages)
nr_pages         7154 mm/memcontrol.c 	id = swap_cgroup_record(entry, 0, nr_pages);
nr_pages         7160 mm/memcontrol.c 				page_counter_uncharge(&memcg->swap, nr_pages);
nr_pages         7162 mm/memcontrol.c 				page_counter_uncharge(&memcg->memsw, nr_pages);
nr_pages         7164 mm/memcontrol.c 		mod_memcg_state(memcg, MEMCG_SWAP, -nr_pages);
nr_pages         7165 mm/memcontrol.c 		mem_cgroup_id_put_many(memcg, nr_pages);
nr_pages         3436 mm/memory.c    	unsigned long address = vmf->address, nr_pages, mask;
nr_pages         3442 mm/memory.c    	nr_pages = READ_ONCE(fault_around_bytes) >> PAGE_SHIFT;
nr_pages         3443 mm/memory.c    	mask = ~(nr_pages * PAGE_SIZE - 1) & PAGE_MASK;
nr_pages         3457 mm/memory.c    			start_pgoff + nr_pages - 1);
nr_pages          227 mm/memory_hotplug.c 	unsigned long i, pfn, end_pfn, nr_pages;
nr_pages          231 mm/memory_hotplug.c 	nr_pages = PAGE_ALIGN(sizeof(struct pglist_data)) >> PAGE_SHIFT;
nr_pages          234 mm/memory_hotplug.c 	for (i = 0; i < nr_pages; i++, page++)
nr_pages          254 mm/memory_hotplug.c static int check_pfn_span(unsigned long pfn, unsigned long nr_pages,
nr_pages          273 mm/memory_hotplug.c 			|| !IS_ALIGNED(nr_pages, min_align)) {
nr_pages          275 mm/memory_hotplug.c 				reason, pfn, pfn + nr_pages - 1);
nr_pages          287 mm/memory_hotplug.c int __ref __add_pages(int nid, unsigned long pfn, unsigned long nr_pages,
nr_pages          299 mm/memory_hotplug.c 				|| vmem_altmap_offset(altmap) > nr_pages) {
nr_pages          306 mm/memory_hotplug.c 	err = check_pfn_span(pfn, nr_pages, "add");
nr_pages          311 mm/memory_hotplug.c 	end_sec = pfn_to_section_nr(pfn + nr_pages - 1);
nr_pages          315 mm/memory_hotplug.c 		pfns = min(nr_pages, PAGES_PER_SECTION
nr_pages          321 mm/memory_hotplug.c 		nr_pages -= pfns;
nr_pages          470 mm/memory_hotplug.c 				      unsigned long nr_pages)
nr_pages          488 mm/memory_hotplug.c 	shrink_zone_span(zone, start_pfn, start_pfn + nr_pages);
nr_pages          495 mm/memory_hotplug.c static void __remove_section(unsigned long pfn, unsigned long nr_pages,
nr_pages          504 mm/memory_hotplug.c 	sparse_remove_section(ms, pfn, nr_pages, map_offset, altmap);
nr_pages          518 mm/memory_hotplug.c void __remove_pages(unsigned long pfn, unsigned long nr_pages,
nr_pages          526 mm/memory_hotplug.c 	if (check_pfn_span(pfn, nr_pages, "remove"))
nr_pages          530 mm/memory_hotplug.c 	end_sec = pfn_to_section_nr(pfn + nr_pages - 1);
nr_pages          535 mm/memory_hotplug.c 		pfns = min(nr_pages, PAGES_PER_SECTION
nr_pages          539 mm/memory_hotplug.c 		nr_pages -= pfns;
nr_pages          616 mm/memory_hotplug.c static int online_pages_range(unsigned long start_pfn, unsigned long nr_pages,
nr_pages          619 mm/memory_hotplug.c 	const unsigned long end_pfn = start_pfn + nr_pages;
nr_pages          639 mm/memory_hotplug.c 	*(unsigned long *)arg += nr_pages;
nr_pages          644 mm/memory_hotplug.c static void node_states_check_changes_online(unsigned long nr_pages,
nr_pages          676 mm/memory_hotplug.c 		unsigned long nr_pages)
nr_pages          683 mm/memory_hotplug.c 	zone->spanned_pages = max(start_pfn + nr_pages, old_end_pfn) - zone->zone_start_pfn;
nr_pages          687 mm/memory_hotplug.c                                      unsigned long nr_pages)
nr_pages          694 mm/memory_hotplug.c 	pgdat->node_spanned_pages = max(start_pfn + nr_pages, old_end_pfn) - pgdat->node_start_pfn;
nr_pages          703 mm/memory_hotplug.c 		unsigned long nr_pages, struct vmem_altmap *altmap)
nr_pages          715 mm/memory_hotplug.c 		init_currently_empty_zone(zone, start_pfn, nr_pages);
nr_pages          716 mm/memory_hotplug.c 	resize_zone_range(zone, start_pfn, nr_pages);
nr_pages          718 mm/memory_hotplug.c 	resize_pgdat_range(pgdat, start_pfn, nr_pages);
nr_pages          727 mm/memory_hotplug.c 	memmap_init_zone(nr_pages, nid, zone_idx(zone), start_pfn,
nr_pages          739 mm/memory_hotplug.c 		unsigned long nr_pages)
nr_pages          747 mm/memory_hotplug.c 		if (zone_intersects(zone, start_pfn, nr_pages))
nr_pages          755 mm/memory_hotplug.c 		unsigned long nr_pages)
nr_pages          758 mm/memory_hotplug.c 			nr_pages);
nr_pages          760 mm/memory_hotplug.c 	bool in_kernel = zone_intersects(kernel_zone, start_pfn, nr_pages);
nr_pages          761 mm/memory_hotplug.c 	bool in_movable = zone_intersects(movable_zone, start_pfn, nr_pages);
nr_pages          779 mm/memory_hotplug.c 		unsigned long nr_pages)
nr_pages          782 mm/memory_hotplug.c 		return default_kernel_zone_for_pfn(nid, start_pfn, nr_pages);
nr_pages          787 mm/memory_hotplug.c 	return default_zone_for_pfn(nid, start_pfn, nr_pages);
nr_pages          790 mm/memory_hotplug.c int __ref online_pages(unsigned long pfn, unsigned long nr_pages, int online_type)
nr_pages          812 mm/memory_hotplug.c 	zone = zone_for_pfn_range(online_type, nid, pfn, nr_pages);
nr_pages          813 mm/memory_hotplug.c 	move_pfn_range_to_zone(zone, pfn, nr_pages, NULL);
nr_pages          816 mm/memory_hotplug.c 	arg.nr_pages = nr_pages;
nr_pages          817 mm/memory_hotplug.c 	node_states_check_changes_online(nr_pages, zone, &arg);
nr_pages          834 mm/memory_hotplug.c 	ret = walk_system_ram_range(pfn, nr_pages, &onlined_pages,
nr_pages          873 mm/memory_hotplug.c 		 (((unsigned long long) pfn + nr_pages) << PAGE_SHIFT) - 1);
nr_pages          875 mm/memory_hotplug.c 	remove_pfn_range_from_zone(zone, pfn, nr_pages);
nr_pages         1193 mm/memory_hotplug.c bool is_mem_section_removable(unsigned long start_pfn, unsigned long nr_pages)
nr_pages         1197 mm/memory_hotplug.c 	end_pfn = min(start_pfn + nr_pages,
nr_pages         1390 mm/memory_hotplug.c offline_isolated_pages_cb(unsigned long start, unsigned long nr_pages,
nr_pages         1395 mm/memory_hotplug.c 	*offlined_pages += __offline_isolated_pages(start, start + nr_pages);
nr_pages         1403 mm/memory_hotplug.c check_pages_isolated_cb(unsigned long start_pfn, unsigned long nr_pages,
nr_pages         1406 mm/memory_hotplug.c 	return test_pages_isolated(start_pfn, start_pfn + nr_pages, true);
nr_pages         1421 mm/memory_hotplug.c static void node_states_check_changes_offline(unsigned long nr_pages,
nr_pages         1442 mm/memory_hotplug.c 	if (zone_idx(zone) <= ZONE_NORMAL && nr_pages >= present_pages)
nr_pages         1455 mm/memory_hotplug.c 	if (zone_idx(zone) <= ZONE_HIGHMEM && nr_pages >= present_pages)
nr_pages         1471 mm/memory_hotplug.c 	if (nr_pages >= present_pages)
nr_pages         1490 mm/memory_hotplug.c 	unsigned long pfn, nr_pages;
nr_pages         1512 mm/memory_hotplug.c 	nr_pages = end_pfn - start_pfn;
nr_pages         1525 mm/memory_hotplug.c 	arg.nr_pages = nr_pages;
nr_pages         1526 mm/memory_hotplug.c 	node_states_check_changes_offline(nr_pages, zone, &arg);
nr_pages         1611 mm/memory_hotplug.c 	remove_pfn_range_from_zone(zone, start_pfn, nr_pages);
nr_pages         1628 mm/memory_hotplug.c int offline_pages(unsigned long start_pfn, unsigned long nr_pages)
nr_pages         1630 mm/memory_hotplug.c 	return __offline_pages(start_pfn, start_pfn + nr_pages);
nr_pages          549 mm/migrate.c   				int nr_pages)
nr_pages          555 mm/migrate.c   	for (i = 0; i < nr_pages; ) {
nr_pages          568 mm/migrate.c   	int nr_pages;
nr_pages          573 mm/migrate.c   		nr_pages = pages_per_huge_page(h);
nr_pages          575 mm/migrate.c   		if (unlikely(nr_pages > MAX_ORDER_NR_PAGES)) {
nr_pages          576 mm/migrate.c   			__copy_gigantic_page(dst, src, nr_pages);
nr_pages          582 mm/migrate.c   		nr_pages = hpage_nr_pages(src);
nr_pages          585 mm/migrate.c   	for (i = 0; i < nr_pages; i++) {
nr_pages         1595 mm/migrate.c   			 unsigned long nr_pages,
nr_pages         1607 mm/migrate.c   	for (i = start = 0; i < nr_pages; i++) {
nr_pages         1644 mm/migrate.c   					err += nr_pages - i - 1;
nr_pages         1679 mm/migrate.c   				err += nr_pages - i - 1;
nr_pages         1713 mm/migrate.c   static void do_pages_stat_array(struct mm_struct *mm, unsigned long nr_pages,
nr_pages         1720 mm/migrate.c   	for (i = 0; i < nr_pages; i++) {
nr_pages         1752 mm/migrate.c   static int do_pages_stat(struct mm_struct *mm, unsigned long nr_pages,
nr_pages         1760 mm/migrate.c   	while (nr_pages) {
nr_pages         1763 mm/migrate.c   		chunk_nr = nr_pages;
nr_pages         1777 mm/migrate.c   		nr_pages -= chunk_nr;
nr_pages         1779 mm/migrate.c   	return nr_pages ? -EFAULT : 0;
nr_pages         1786 mm/migrate.c   static int kernel_move_pages(pid_t pid, unsigned long nr_pages,
nr_pages         1835 mm/migrate.c   		err = do_pages_move(mm, task_nodes, nr_pages, pages,
nr_pages         1838 mm/migrate.c   		err = do_pages_stat(mm, nr_pages, pages, status);
nr_pages         1848 mm/migrate.c   SYSCALL_DEFINE6(move_pages, pid_t, pid, unsigned long, nr_pages,
nr_pages         1853 mm/migrate.c   	return kernel_move_pages(pid, nr_pages, pages, nodes, status, flags);
nr_pages         1857 mm/migrate.c   COMPAT_SYSCALL_DEFINE6(move_pages, pid_t, pid, compat_ulong_t, nr_pages,
nr_pages         1866 mm/migrate.c   	pages = compat_alloc_user_space(nr_pages * sizeof(void *));
nr_pages         1867 mm/migrate.c   	for (i = 0; i < nr_pages; i++) {
nr_pages         1874 mm/migrate.c   	return kernel_move_pages(pid, nr_pages, pages, nodes, status, flags);
nr_pages         2665 mm/migrate.c   	long nr_pages = (args->end - args->start) >> PAGE_SHIFT;
nr_pages         2672 mm/migrate.c   	if (nr_pages <= 0)
nr_pages         2682 mm/migrate.c   	memset(args->src, 0, sizeof(*args->src) * nr_pages);
nr_pages          184 mm/mlock.c     	int nr_pages;
nr_pages          201 mm/mlock.c     		nr_pages = 1;
nr_pages          205 mm/mlock.c     	nr_pages = hpage_nr_pages(page);
nr_pages          206 mm/mlock.c     	__mod_zone_page_state(page_zone(page), NR_MLOCK, -nr_pages);
nr_pages          219 mm/mlock.c     	return nr_pages - 1;
nr_pages          524 mm/mlock.c     	int nr_pages;
nr_pages          560 mm/mlock.c     	nr_pages = (end - start) >> PAGE_SHIFT;
nr_pages          562 mm/mlock.c     		nr_pages = -nr_pages;
nr_pages          564 mm/mlock.c     		nr_pages = 0;
nr_pages          565 mm/mlock.c     	mm->locked_vm += nr_pages;
nr_pages          563 mm/mmap.c      	unsigned long nr_pages = 0;
nr_pages          571 mm/mmap.c      	nr_pages = (min(end, vma->vm_end) -
nr_pages          582 mm/mmap.c      		nr_pages += overlap_len >> PAGE_SHIFT;
nr_pages          585 mm/mmap.c      	return nr_pages;
nr_pages         1723 mm/mmap.c      		unsigned long nr_pages;
nr_pages         1729 mm/mmap.c      		nr_pages = count_vma_pages_range(mm, addr, addr + len);
nr_pages         1732 mm/mmap.c      					(len >> PAGE_SHIFT) - nr_pages))
nr_pages          279 mm/page-writeback.c 	unsigned long nr_pages = 0;
nr_pages          288 mm/page-writeback.c 		nr_pages += zone_page_state(zone, NR_FREE_PAGES);
nr_pages          296 mm/page-writeback.c 	nr_pages -= min(nr_pages, pgdat->totalreserve_pages);
nr_pages          298 mm/page-writeback.c 	nr_pages += node_page_state(pgdat, NR_INACTIVE_FILE);
nr_pages          299 mm/page-writeback.c 	nr_pages += node_page_state(pgdat, NR_ACTIVE_FILE);
nr_pages          301 mm/page-writeback.c 	return nr_pages;
nr_pages          314 mm/page-writeback.c 			unsigned long nr_pages;
nr_pages          323 mm/page-writeback.c 			nr_pages = zone_page_state(z, NR_FREE_PAGES);
nr_pages          325 mm/page-writeback.c 			nr_pages -= min(nr_pages, high_wmark_pages(z));
nr_pages          326 mm/page-writeback.c 			nr_pages += zone_page_state(z, NR_ZONE_INACTIVE_FILE);
nr_pages          327 mm/page-writeback.c 			nr_pages += zone_page_state(z, NR_ZONE_ACTIVE_FILE);
nr_pages          328 mm/page-writeback.c 			x += nr_pages;
nr_pages          505 mm/page-writeback.c 	unsigned long nr_pages = 0;
nr_pages          507 mm/page-writeback.c 	nr_pages += node_page_state(pgdat, NR_FILE_DIRTY);
nr_pages          508 mm/page-writeback.c 	nr_pages += node_page_state(pgdat, NR_UNSTABLE_NFS);
nr_pages          509 mm/page-writeback.c 	nr_pages += node_page_state(pgdat, NR_WRITEBACK);
nr_pages          511 mm/page-writeback.c 	return nr_pages <= limit;
nr_pages         2166 mm/page-writeback.c 	int nr_pages;
nr_pages         2195 mm/page-writeback.c 		nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end,
nr_pages         2197 mm/page-writeback.c 		if (nr_pages == 0)
nr_pages         2200 mm/page-writeback.c 		for (i = 0; i < nr_pages; i++) {
nr_pages          680 mm/page_alloc.c 	int nr_pages = 1 << order;
nr_pages          685 mm/page_alloc.c 	for (i = 1; i < nr_pages; i++) {
nr_pages         1432 mm/page_alloc.c 	unsigned int nr_pages = 1 << order;
nr_pages         1437 mm/page_alloc.c 	for (loop = 0; loop < (nr_pages - 1); loop++, p++) {
nr_pages         1445 mm/page_alloc.c 	atomic_long_add(nr_pages, &page_zone(page)->managed_pages);
nr_pages         1572 mm/page_alloc.c 				       unsigned long nr_pages)
nr_pages         1577 mm/page_alloc.c 	if (!nr_pages)
nr_pages         1583 mm/page_alloc.c 	if (nr_pages == pageblock_nr_pages &&
nr_pages         1590 mm/page_alloc.c 	for (i = 0; i < nr_pages; i++, page++, pfn++) {
nr_pages         1663 mm/page_alloc.c 	unsigned long nr_pages = 0;
nr_pages         1678 mm/page_alloc.c 		nr_pages++;
nr_pages         1680 mm/page_alloc.c 	return (nr_pages);
nr_pages         1729 mm/page_alloc.c 	unsigned long nr_pages = 0;
nr_pages         1740 mm/page_alloc.c 		nr_pages += deferred_init_pages(zone, *start_pfn, t);
nr_pages         1764 mm/page_alloc.c 	return nr_pages;
nr_pages         1772 mm/page_alloc.c 	unsigned long spfn = 0, epfn = 0, nr_pages = 0;
nr_pages         1814 mm/page_alloc.c 		nr_pages += deferred_init_maxorder(&i, zone, &spfn, &epfn);
nr_pages         1822 mm/page_alloc.c 		pgdat->node_id,	nr_pages, jiffies_to_msecs(jiffies - start));
nr_pages         1850 mm/page_alloc.c 	unsigned long nr_pages = 0;
nr_pages         1897 mm/page_alloc.c 		nr_pages += deferred_init_maxorder(&i, zone, &spfn, &epfn);
nr_pages         1904 mm/page_alloc.c 		if (nr_pages >= nr_pages_needed)
nr_pages         1911 mm/page_alloc.c 	return nr_pages > 0;
nr_pages         8515 mm/page_alloc.c void free_contig_range(unsigned long pfn, unsigned int nr_pages)
nr_pages         8519 mm/page_alloc.c 	for (; nr_pages--; pfn++) {
nr_pages           55 mm/page_counter.c void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages)
nr_pages           59 mm/page_counter.c 	new = atomic_long_sub_return(nr_pages, &counter->usage);
nr_pages           72 mm/page_counter.c void page_counter_charge(struct page_counter *counter, unsigned long nr_pages)
nr_pages           79 mm/page_counter.c 		new = atomic_long_add_return(nr_pages, &c->usage);
nr_pages          100 mm/page_counter.c 			     unsigned long nr_pages,
nr_pages          121 mm/page_counter.c 		new = atomic_long_add_return(nr_pages, &c->usage);
nr_pages          123 mm/page_counter.c 			atomic_long_sub(nr_pages, &c->usage);
nr_pages          145 mm/page_counter.c 		page_counter_cancel(c, nr_pages);
nr_pages          155 mm/page_counter.c void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages)
nr_pages          160 mm/page_counter.c 		page_counter_cancel(c, nr_pages);
nr_pages          173 mm/page_counter.c int page_counter_set_max(struct page_counter *counter, unsigned long nr_pages)
nr_pages          192 mm/page_counter.c 		if (usage > nr_pages)
nr_pages          195 mm/page_counter.c 		old = xchg(&counter->max, nr_pages);
nr_pages          212 mm/page_counter.c void page_counter_set_min(struct page_counter *counter, unsigned long nr_pages)
nr_pages          216 mm/page_counter.c 	counter->min = nr_pages;
nr_pages          229 mm/page_counter.c void page_counter_set_low(struct page_counter *counter, unsigned long nr_pages)
nr_pages          233 mm/page_counter.c 	counter->low = nr_pages;
nr_pages          249 mm/page_counter.c 			  unsigned long *nr_pages)
nr_pages          255 mm/page_counter.c 		*nr_pages = PAGE_COUNTER_MAX;
nr_pages          263 mm/page_counter.c 	*nr_pages = min(bytes / PAGE_SIZE, (u64)PAGE_COUNTER_MAX);
nr_pages          139 mm/page_ext.c  	unsigned long nr_pages;
nr_pages          141 mm/page_ext.c  	nr_pages = NODE_DATA(nid)->node_spanned_pages;
nr_pages          142 mm/page_ext.c  	if (!nr_pages)
nr_pages          152 mm/page_ext.c  		nr_pages += MAX_ORDER_NR_PAGES;
nr_pages          154 mm/page_ext.c  	table_size = page_ext_size * nr_pages;
nr_pages          287 mm/page_ext.c  				unsigned long nr_pages,
nr_pages          294 mm/page_ext.c  	end = SECTION_ALIGN_UP(start_pfn + nr_pages);
nr_pages          322 mm/page_ext.c  				unsigned long nr_pages, int nid)
nr_pages          327 mm/page_ext.c  	end = SECTION_ALIGN_UP(start_pfn + nr_pages);
nr_pages          344 mm/page_ext.c  				   mn->nr_pages, mn->status_change_nid);
nr_pages          348 mm/page_ext.c  				mn->nr_pages, mn->status_change_nid);
nr_pages          352 mm/page_ext.c  				mn->nr_pages, mn->status_change_nid);
nr_pages           40 mm/page_isolation.c 	arg.nr_pages = pageblock_nr_pages;
nr_pages           73 mm/page_isolation.c 		unsigned long nr_pages;
nr_pages           78 mm/page_isolation.c 		nr_pages = move_freepages_block(zone, page, MIGRATE_ISOLATE,
nr_pages           81 mm/page_isolation.c 		__mod_zone_freepage_state(zone, -nr_pages, mt);
nr_pages           93 mm/page_isolation.c 	unsigned long flags, nr_pages;
nr_pages          133 mm/page_isolation.c 		nr_pages = move_freepages_block(zone, page, migratetype, NULL);
nr_pages          134 mm/page_isolation.c 		__mod_zone_freepage_state(zone, nr_pages, migratetype);
nr_pages          147 mm/page_isolation.c __first_valid_page(unsigned long pfn, unsigned long nr_pages)
nr_pages          151 mm/page_isolation.c 	for (i = 0; i < nr_pages; i++) {
nr_pages           58 mm/percpu-internal.h 	int			nr_pages;	/* # of pages served by this chunk */
nr_pages           82 mm/percpu-internal.h 	return chunk->nr_pages * PAGE_SIZE / PCPU_BITMAP_BLOCK_SIZE;
nr_pages          106 mm/percpu-internal.h 	return pcpu_nr_pages_to_map_bits(chunk->nr_pages);
nr_pages           49 mm/percpu-km.c 	const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT;
nr_pages           59 mm/percpu-km.c 	pages = alloc_pages(gfp, order_base_2(nr_pages));
nr_pages           65 mm/percpu-km.c 	for (i = 0; i < nr_pages; i++)
nr_pages           72 mm/percpu-km.c 	pcpu_chunk_populated(chunk, 0, nr_pages);
nr_pages           83 mm/percpu-km.c 	const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT;
nr_pages           92 mm/percpu-km.c 		__free_pages(chunk->data, order_base_2(nr_pages));
nr_pages          103 mm/percpu-km.c 	size_t nr_pages, alloc_pages;
nr_pages          111 mm/percpu-km.c 	nr_pages = (ai->groups[0].nr_units * ai->unit_size) >> PAGE_SHIFT;
nr_pages          112 mm/percpu-km.c 	alloc_pages = roundup_pow_of_two(nr_pages);
nr_pages          114 mm/percpu-km.c 	if (alloc_pages > nr_pages)
nr_pages          116 mm/percpu-km.c 			alloc_pages - nr_pages);
nr_pages          134 mm/percpu-vm.c static void __pcpu_unmap_pages(unsigned long addr, int nr_pages)
nr_pages          136 mm/percpu-vm.c 	unmap_kernel_range_noflush(addr, nr_pages << PAGE_SHIFT);
nr_pages          193 mm/percpu-vm.c 			    int nr_pages)
nr_pages          195 mm/percpu-vm.c 	return map_kernel_range_noflush(addr, nr_pages << PAGE_SHIFT,
nr_pages          214 mm/percpu.c    	end_addr = chunk->base_addr + chunk->nr_pages * PAGE_SIZE -
nr_pages         1343 mm/percpu.c    	chunk->nr_pages = region_size >> PAGE_SHIFT;
nr_pages         1369 mm/percpu.c    	bitmap_fill(chunk->populated, chunk->nr_pages);
nr_pages         1370 mm/percpu.c    	chunk->nr_populated = chunk->nr_pages;
nr_pages         1371 mm/percpu.c    	chunk->nr_empty_pop_pages = chunk->nr_pages;
nr_pages         1414 mm/percpu.c    	chunk->nr_pages = pcpu_unit_pages;
nr_pages         1435 mm/percpu.c    	chunk->free_bytes = chunk->nr_pages * PAGE_SIZE;
nr_pages         1864 mm/percpu.c    					 chunk->nr_pages) {
nr_pages         1903 mm/percpu.c    			nr_unpop = chunk->nr_pages - chunk->nr_populated;
nr_pages         1914 mm/percpu.c    					   chunk->nr_pages) {
nr_pages           84 mm/process_vm_access.c 	unsigned long nr_pages;
nr_pages           93 mm/process_vm_access.c 	nr_pages = (addr + len - 1) / PAGE_SIZE - addr / PAGE_SIZE + 1;
nr_pages           98 mm/process_vm_access.c 	while (!rc && nr_pages && iov_iter_count(iter)) {
nr_pages           99 mm/process_vm_access.c 		int pages = min(nr_pages, max_pages_per_loop);
nr_pages          125 mm/process_vm_access.c 		nr_pages -= pages;
nr_pages          162 mm/process_vm_access.c 	unsigned long nr_pages = 0;
nr_pages          178 mm/process_vm_access.c 			nr_pages = max(nr_pages, nr_pages_iov);
nr_pages          182 mm/process_vm_access.c 	if (nr_pages == 0)
nr_pages          185 mm/process_vm_access.c 	if (nr_pages > PVM_MAX_PP_ARRAY_COUNT) {
nr_pages          189 mm/process_vm_access.c 					      sizeof(struct pages *)*nr_pages),
nr_pages          117 mm/readahead.c 		struct list_head *pages, unsigned int nr_pages, gfp_t gfp)
nr_pages          126 mm/readahead.c 		ret = mapping->a_ops->readpages(filp, mapping, pages, nr_pages);
nr_pages          132 mm/readahead.c 	for (page_idx = 0; page_idx < nr_pages; page_idx++) {
nr_pages          164 mm/readahead.c 	unsigned int nr_pages = 0;
nr_pages          189 mm/readahead.c 			if (nr_pages)
nr_pages          190 mm/readahead.c 				read_pages(mapping, filp, &page_pool, nr_pages,
nr_pages          192 mm/readahead.c 			nr_pages = 0;
nr_pages          203 mm/readahead.c 		nr_pages++;
nr_pages          211 mm/readahead.c 	if (nr_pages)
nr_pages          212 mm/readahead.c 		read_pages(mapping, filp, &page_pool, nr_pages, gfp_mask);
nr_pages          215 mm/readahead.c 	return nr_pages;
nr_pages          133 mm/shmem.c     	unsigned long nr_pages = totalram_pages();
nr_pages          135 mm/shmem.c     	return min(nr_pages - totalhigh_pages(), nr_pages / 2);
nr_pages          249 mm/sparse-vmemmap.c 		unsigned long nr_pages, int nid, struct vmem_altmap *altmap)
nr_pages          259 mm/sparse-vmemmap.c 	end = ALIGN(pfn + nr_pages, PAGES_PER_SUBSECTION);
nr_pages          261 mm/sparse-vmemmap.c 	nr_pages = end - pfn;
nr_pages          264 mm/sparse-vmemmap.c 	end = start + nr_pages * sizeof(struct page);
nr_pages          223 mm/sparse.c    		unsigned long nr_pages)
nr_pages          226 mm/sparse.c    	int end = subsection_map_index(pfn + nr_pages - 1);
nr_pages          231 mm/sparse.c    void __init subsection_map_init(unsigned long pfn, unsigned long nr_pages)
nr_pages          233 mm/sparse.c    	int end_sec = pfn_to_section_nr(pfn + nr_pages - 1);
nr_pages          236 mm/sparse.c    	if (!nr_pages)
nr_pages          243 mm/sparse.c    		pfns = min(nr_pages, PAGES_PER_SECTION
nr_pages          253 mm/sparse.c    		nr_pages -= pfns;
nr_pages          452 mm/sparse.c    		unsigned long nr_pages, int nid, struct vmem_altmap *altmap)
nr_pages          651 mm/sparse.c    		unsigned long nr_pages, int nid, struct vmem_altmap *altmap)
nr_pages          653 mm/sparse.c    	return __populate_section_memmap(pfn, nr_pages, nid, altmap);
nr_pages          656 mm/sparse.c    static void depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages,
nr_pages          660 mm/sparse.c    	unsigned long end = start + nr_pages * sizeof(struct page);
nr_pages          673 mm/sparse.c    		unsigned long nr_pages, int nid, struct vmem_altmap *altmap)
nr_pages          694 mm/sparse.c    static void depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages,
nr_pages          709 mm/sparse.c    	unsigned long magic, nr_pages;
nr_pages          712 mm/sparse.c    	nr_pages = PAGE_ALIGN(PAGES_PER_SECTION * sizeof(struct page))
nr_pages          715 mm/sparse.c    	for (i = 0; i < nr_pages; i++, page++) {
nr_pages          737 mm/sparse.c    static void section_deactivate(unsigned long pfn, unsigned long nr_pages,
nr_pages          749 mm/sparse.c    	subsection_mask_set(map, pfn, nr_pages);
nr_pages          755 mm/sparse.c    				pfn, nr_pages))
nr_pages          803 mm/sparse.c    		depopulate_section_memmap(pfn, nr_pages, altmap);
nr_pages          810 mm/sparse.c    		unsigned long nr_pages, struct vmem_altmap *altmap)
nr_pages          819 mm/sparse.c    	subsection_mask_set(map, pfn, nr_pages);
nr_pages          851 mm/sparse.c    	if (nr_pages < PAGES_PER_SECTION && early_section(ms))
nr_pages          854 mm/sparse.c    	memmap = populate_section_memmap(pfn, nr_pages, nid, altmap);
nr_pages          856 mm/sparse.c    		section_deactivate(pfn, nr_pages, altmap);
nr_pages          878 mm/sparse.c    		unsigned long nr_pages, struct vmem_altmap *altmap)
nr_pages          889 mm/sparse.c    	memmap = section_activate(nid, start_pfn, nr_pages, altmap);
nr_pages          897 mm/sparse.c    	page_init_poison(memmap, sizeof(struct page) * nr_pages);
nr_pages          912 mm/sparse.c    static void clear_hwpoisoned_pages(struct page *memmap, int nr_pages)
nr_pages          925 mm/sparse.c    	for (i = 0; i < nr_pages; i++) {
nr_pages          933 mm/sparse.c    static inline void clear_hwpoisoned_pages(struct page *memmap, int nr_pages)
nr_pages          939 mm/sparse.c    		unsigned long nr_pages, unsigned long map_offset,
nr_pages          943 mm/sparse.c    			nr_pages - map_offset);
nr_pages          944 mm/sparse.c    	section_deactivate(pfn, nr_pages, altmap);
nr_pages          597 mm/swap_state.c int init_swap_address_space(unsigned int type, unsigned long nr_pages)
nr_pages          602 mm/swap_state.c 	nr = DIV_ROUND_UP(nr_pages, SWAP_ADDRESS_SPACE_PAGES);
nr_pages          181 mm/swapfile.c  	nr_blocks = ((sector_t)se->nr_pages - 1) << (PAGE_SHIFT - 9);
nr_pages          192 mm/swapfile.c  		nr_blocks = (sector_t)se->nr_pages << (PAGE_SHIFT - 9);
nr_pages          215 mm/swapfile.c  		else if (offset >= se->start_page + se->nr_pages)
nr_pages          229 mm/swapfile.c  				 pgoff_t start_page, pgoff_t nr_pages)
nr_pages          233 mm/swapfile.c  	while (nr_pages) {
nr_pages          236 mm/swapfile.c  		sector_t nr_blocks = se->nr_pages - offset;
nr_pages          238 mm/swapfile.c  		if (nr_blocks > nr_pages)
nr_pages          239 mm/swapfile.c  			nr_blocks = nr_pages;
nr_pages          241 mm/swapfile.c  		nr_pages -= nr_blocks;
nr_pages         2312 mm/swapfile.c  		unsigned long nr_pages, sector_t start_block)
nr_pages         2329 mm/swapfile.c  		BUG_ON(se->start_page + se->nr_pages != start_page);
nr_pages         2330 mm/swapfile.c  		if (se->start_block + se->nr_pages == start_block) {
nr_pages         2332 mm/swapfile.c  			se->nr_pages += nr_pages;
nr_pages         2342 mm/swapfile.c  	new_se->nr_pages = nr_pages;
nr_pages         2176 mm/vmalloc.c   	for (i = 0; i < area->nr_pages; i++)
nr_pages         2209 mm/vmalloc.c   	for (i = 0; i < area->nr_pages; i++) {
nr_pages         2254 mm/vmalloc.c   		for (i = 0; i < area->nr_pages; i++) {
nr_pages         2260 mm/vmalloc.c   		atomic_long_sub(area->nr_pages, &nr_vmalloc_pages);
nr_pages         2402 mm/vmalloc.c   	unsigned int nr_pages, array_size, i;
nr_pages         2409 mm/vmalloc.c   	nr_pages = get_vm_area_size(area) >> PAGE_SHIFT;
nr_pages         2410 mm/vmalloc.c   	array_size = (nr_pages * sizeof(struct page *));
nr_pages         2427 mm/vmalloc.c   	area->nr_pages = nr_pages;
nr_pages         2429 mm/vmalloc.c   	for (i = 0; i < area->nr_pages; i++) {
nr_pages         2439 mm/vmalloc.c   			area->nr_pages = i;
nr_pages         2440 mm/vmalloc.c   			atomic_long_add(area->nr_pages, &nr_vmalloc_pages);
nr_pages         2447 mm/vmalloc.c   	atomic_long_add(area->nr_pages, &nr_vmalloc_pages);
nr_pages         2456 mm/vmalloc.c   			  (area->nr_pages*PAGE_SIZE), area->size);
nr_pages         3465 mm/vmalloc.c   		for (nr = 0; nr < v->nr_pages; nr++)
nr_pages         3517 mm/vmalloc.c   	if (v->nr_pages)
nr_pages         3518 mm/vmalloc.c   		seq_printf(m, " pages=%d", v->nr_pages);
nr_pages         1140 mm/vmscan.c    		unsigned int nr_pages;
nr_pages         1152 mm/vmscan.c    		nr_pages = compound_nr(page);
nr_pages         1155 mm/vmscan.c    		sc->nr_scanned += nr_pages;
nr_pages         1276 mm/vmscan.c    			stat->nr_ref_keep += nr_pages;
nr_pages         1338 mm/vmscan.c    		if ((nr_pages > 1) && !PageTransHuge(page)) {
nr_pages         1339 mm/vmscan.c    			sc->nr_scanned -= (nr_pages - 1);
nr_pages         1340 mm/vmscan.c    			nr_pages = 1;
nr_pages         1353 mm/vmscan.c    				stat->nr_unmap_fail += nr_pages;
nr_pages         1484 mm/vmscan.c    		nr_reclaimed += nr_pages;
nr_pages         1501 mm/vmscan.c    		if (nr_pages > 1) {
nr_pages         1502 mm/vmscan.c    			sc->nr_scanned -= (nr_pages - 1);
nr_pages         1503 mm/vmscan.c    			nr_pages = 1;
nr_pages         1514 mm/vmscan.c    			stat->nr_activate[type] += nr_pages;
nr_pages         1693 mm/vmscan.c    	unsigned long scan, total_scan, nr_pages;
nr_pages         1707 mm/vmscan.c    		nr_pages = compound_nr(page);
nr_pages         1708 mm/vmscan.c    		total_scan += nr_pages;
nr_pages         1712 mm/vmscan.c    			nr_skipped[page_zonenum(page)] += nr_pages;
nr_pages         1726 mm/vmscan.c    		scan += nr_pages;
nr_pages         1729 mm/vmscan.c    			nr_taken += nr_pages;
nr_pages         1730 mm/vmscan.c    			nr_zone_taken[page_zonenum(page)] += nr_pages;
nr_pages         1882 mm/vmscan.c    	int nr_pages, nr_moved = 0;
nr_pages         1902 mm/vmscan.c    		nr_pages = hpage_nr_pages(page);
nr_pages         1903 mm/vmscan.c    		update_lru_size(lruvec, lru, page_zonenum(page), nr_pages);
nr_pages         1918 mm/vmscan.c    			nr_moved += nr_pages;
nr_pages         3350 mm/vmscan.c    					   unsigned long nr_pages,
nr_pages         3360 mm/vmscan.c    		.nr_to_reclaim = max(nr_pages, SWAP_CLUSTER_MAX),
nr_pages         4212 mm/vmscan.c    	const unsigned long nr_pages = 1 << order;
nr_pages         4216 mm/vmscan.c    		.nr_to_reclaim = max(nr_pages, SWAP_CLUSTER_MAX),
nr_pages         4247 mm/vmscan.c    		} while (sc.nr_reclaimed < nr_pages && --sc.priority >= 0);
nr_pages         4257 mm/vmscan.c    	return sc.nr_reclaimed >= nr_pages;
nr_pages         1033 mm/zsmalloc.c  	int nr_pages = class->pages_per_zspage;
nr_pages         1043 mm/zsmalloc.c  	for (i = 0; i < nr_pages; i++) {
nr_pages           22 net/9p/trans_common.c void p9_release_pages(struct page **pages, int nr_pages)
nr_pages           26 net/9p/trans_common.c 	for (i = 0; i < nr_pages; i++)
nr_pages          214 net/9p/trans_virtio.c 	       struct page **pdata, int nr_pages, size_t offs, int count)
nr_pages          220 net/9p/trans_virtio.c 	BUG_ON(nr_pages > (limit - start));
nr_pages          225 net/9p/trans_virtio.c 	while (nr_pages) {
nr_pages          235 net/9p/trans_virtio.c 		nr_pages--;
nr_pages          311 net/9p/trans_virtio.c 	int nr_pages;
nr_pages          333 net/9p/trans_virtio.c 		nr_pages = DIV_ROUND_UP(n + *offs, PAGE_SIZE);
nr_pages          334 net/9p/trans_virtio.c 		atomic_add(nr_pages, &vp_pinned);
nr_pages          354 net/9p/trans_virtio.c 		nr_pages = DIV_ROUND_UP((unsigned long)p + len, PAGE_SIZE) -
nr_pages          357 net/9p/trans_virtio.c 		*pages = kmalloc_array(nr_pages, sizeof(struct page *),
nr_pages          364 net/9p/trans_virtio.c 		for (index = 0; index < nr_pages; index++) {
nr_pages         2298 net/core/skbuff.c 	return	spd->nr_pages &&
nr_pages         2299 net/core/skbuff.c 		spd->pages[spd->nr_pages - 1] == page &&
nr_pages         2300 net/core/skbuff.c 		(spd->partial[spd->nr_pages - 1].offset +
nr_pages         2301 net/core/skbuff.c 		 spd->partial[spd->nr_pages - 1].len == offset);
nr_pages         2313 net/core/skbuff.c 	if (unlikely(spd->nr_pages == MAX_SKB_FRAGS))
nr_pages         2322 net/core/skbuff.c 		spd->partial[spd->nr_pages - 1].len += *len;
nr_pages         2326 net/core/skbuff.c 	spd->pages[spd->nr_pages] = page;
nr_pages         2327 net/core/skbuff.c 	spd->partial[spd->nr_pages].len = *len;
nr_pages         2328 net/core/skbuff.c 	spd->partial[spd->nr_pages].offset = offset;
nr_pages         2329 net/core/skbuff.c 	spd->nr_pages++;
nr_pages         2442 net/core/skbuff.c 	if (spd.nr_pages)
nr_pages         1130 net/dccp/proto.c 	unsigned long nr_pages = totalram_pages();
nr_pages         1157 net/dccp/proto.c 	if (nr_pages >= (128 * 1024))
nr_pages         1158 net/dccp/proto.c 		goal = nr_pages >> (21 - PAGE_SHIFT);
nr_pages         1160 net/dccp/proto.c 		goal = nr_pages >> (23 - PAGE_SHIFT);
nr_pages         2438 net/netfilter/nf_conntrack_core.c 	unsigned long nr_pages = totalram_pages();
nr_pages         2458 net/netfilter/nf_conntrack_core.c 			= (((nr_pages << PAGE_SHIFT) / 16384)
nr_pages         2460 net/netfilter/nf_conntrack_core.c 		if (nr_pages > (4 * (1024 * 1024 * 1024 / PAGE_SIZE)))
nr_pages         2462 net/netfilter/nf_conntrack_core.c 		else if (nr_pages > (1024 * 1024 * 1024 / PAGE_SIZE))
nr_pages          284 net/netfilter/xt_hashlimit.c 	unsigned long nr_pages = totalram_pages();
nr_pages          290 net/netfilter/xt_hashlimit.c 		size = (nr_pages << PAGE_SHIFT) / 16384 /
nr_pages          292 net/netfilter/xt_hashlimit.c 		if (nr_pages > 1024 * 1024 * 1024 / PAGE_SIZE)
nr_pages          163 net/rds/info.c 	unsigned long nr_pages = 0;
nr_pages          188 net/rds/info.c 	nr_pages = (PAGE_ALIGN(start + len) - (start & PAGE_MASK))
nr_pages          191 net/rds/info.c 	pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_KERNEL);
nr_pages          196 net/rds/info.c 	ret = get_user_pages_fast(start, nr_pages, FOLL_WRITE, pages);
nr_pages          197 net/rds/info.c 	if (ret != nr_pages) {
nr_pages          199 net/rds/info.c 			nr_pages = ret;
nr_pages          201 net/rds/info.c 			nr_pages = 0;
nr_pages          206 net/rds/info.c 	rdsdebug("len %d nr_pages %lu\n", len, nr_pages);
nr_pages          238 net/rds/info.c 	for (i = 0; pages && i < nr_pages; i++)
nr_pages          156 net/rds/rdma.c static int rds_pin_pages(unsigned long user_addr, unsigned int nr_pages,
nr_pages          161 net/rds/rdma.c 	ret = get_user_pages_fast(user_addr, nr_pages, write ? FOLL_WRITE : 0,
nr_pages          164 net/rds/rdma.c 	if (ret >= 0 && ret < nr_pages) {
nr_pages          178 net/rds/rdma.c 	unsigned int nr_pages;
nr_pages          198 net/rds/rdma.c 	nr_pages = rds_pages_in_vec(&args->vec);
nr_pages          199 net/rds/rdma.c 	if (nr_pages == 0) {
nr_pages          207 net/rds/rdma.c 	if ((nr_pages - 1) > (RDS_MAX_MSG_SIZE >> PAGE_SHIFT)) {
nr_pages          213 net/rds/rdma.c 		args->vec.addr, args->vec.bytes, nr_pages);
nr_pages          216 net/rds/rdma.c 	pages = kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL);
nr_pages          250 net/rds/rdma.c 	ret = rds_pin_pages(args->vec.addr, nr_pages, pages, 1);
nr_pages          499 net/rds/rdma.c 	unsigned int nr_pages;
nr_pages          504 net/rds/rdma.c 		nr_pages = rds_pages_in_vec(&iov[i]);
nr_pages          505 net/rds/rdma.c 		if (nr_pages == 0)
nr_pages          508 net/rds/rdma.c 		tot_pages += nr_pages;
nr_pages          527 net/rds/rdma.c 	unsigned int nr_pages;
nr_pages          551 net/rds/rdma.c 		nr_pages = rds_pages_in_vec(vec);
nr_pages          552 net/rds/rdma.c 		if (nr_pages == 0)
nr_pages          555 net/rds/rdma.c 		tot_pages += nr_pages;
nr_pages          578 net/rds/rdma.c 	int nr_pages;
nr_pages          608 net/rds/rdma.c 	nr_pages = rds_rdma_pages(iovs, args->nr_local);
nr_pages          609 net/rds/rdma.c 	if (nr_pages < 0) {
nr_pages          614 net/rds/rdma.c 	pages = kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL);
nr_pages          626 net/rds/rdma.c 	WARN_ON(!nr_pages);
nr_pages          627 net/rds/rdma.c 	op->op_sg = rds_message_alloc_sgs(rm, nr_pages);
nr_pages         1375 net/sctp/protocol.c 	unsigned long nr_pages = totalram_pages();
nr_pages         1434 net/sctp/protocol.c 	if (nr_pages >= (128 * 1024))
nr_pages         1435 net/sctp/protocol.c 		goal = nr_pages >> (22 - PAGE_SHIFT);
nr_pages         1437 net/sctp/protocol.c 		goal = nr_pages >> (24 - PAGE_SHIFT);
nr_pages          169 net/smc/smc_rx.c 	spd.nr_pages = 1;
nr_pages           24 security/keys/big_key.c 	unsigned int		nr_pages;
nr_pages          144 security/keys/big_key.c 		memset(buf->virt, 0, buf->nr_pages * PAGE_SIZE);
nr_pages          148 security/keys/big_key.c 	for (i = 0; i < buf->nr_pages; i++)
nr_pages          172 security/keys/big_key.c 	buf->nr_pages = npg;
nr_pages          176 security/keys/big_key.c 	for (i = 0; i < buf->nr_pages; i++) {
nr_pages          186 security/keys/big_key.c 	buf->virt = vmap(buf->pages, buf->nr_pages, VM_MAP, PAGE_KERNEL);
nr_pages           11 tools/testing/nvdimm/pmem-dax.c 		long nr_pages, void **kaddr, pfn_t *pfn)
nr_pages           16 tools/testing/nvdimm/pmem-dax.c 					PFN_PHYS(nr_pages))))
nr_pages           47 tools/testing/nvdimm/pmem-dax.c 		return nr_pages;
nr_pages           80 tools/testing/radix-tree/regression1.c 			    unsigned int nr_pages, struct page **pages)
nr_pages           35 tools/testing/selftests/vm/gup_benchmark.c 	int i, fd, filed, opt, nr_pages = 1, thp = -1, repeats = 1, write = 0;
nr_pages           49 tools/testing/selftests/vm/gup_benchmark.c 			nr_pages = atoi(optarg);
nr_pages           87 tools/testing/selftests/vm/gup_benchmark.c 	gup.nr_pages_per_call = nr_pages;
nr_pages          104 tools/testing/selftests/vm/transhuge-stress.c 		int nr_succeed = 0, nr_failed = 0, nr_pages = 0;
nr_pages          128 tools/testing/selftests/vm/transhuge-stress.c 					nr_pages++;
nr_pages          142 tools/testing/selftests/vm/transhuge-stress.c 		      nr_succeed, nr_failed, nr_pages);
nr_pages           62 tools/testing/selftests/vm/userfaultfd.c static unsigned long nr_cpus, nr_pages, nr_pages_per_cpu, page_size;
nr_pages          132 tools/testing/selftests/vm/userfaultfd.c 	if (madvise(rel_area, nr_pages * page_size, MADV_DONTNEED)) {
nr_pages          142 tools/testing/selftests/vm/userfaultfd.c 	if (posix_memalign(alloc_area, page_size, nr_pages * page_size)) {
nr_pages          159 tools/testing/selftests/vm/userfaultfd.c 				nr_pages * page_size,
nr_pages          160 tools/testing/selftests/vm/userfaultfd.c 				nr_pages * page_size)) {
nr_pages          173 tools/testing/selftests/vm/userfaultfd.c 	*alloc_area = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE,
nr_pages          177 tools/testing/selftests/vm/userfaultfd.c 			   nr_pages * page_size);
nr_pages          184 tools/testing/selftests/vm/userfaultfd.c 		area_alias = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE,
nr_pages          187 tools/testing/selftests/vm/userfaultfd.c 				  nr_pages * page_size);
nr_pages          189 tools/testing/selftests/vm/userfaultfd.c 			if (munmap(*alloc_area, nr_pages * page_size) < 0)
nr_pages          223 tools/testing/selftests/vm/userfaultfd.c 	if (madvise(rel_area, nr_pages * page_size, MADV_REMOVE)) {
nr_pages          233 tools/testing/selftests/vm/userfaultfd.c 	*alloc_area = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE,
nr_pages          322 tools/testing/selftests/vm/userfaultfd.c 		page_nr %= nr_pages;
nr_pages          416 tools/testing/selftests/vm/userfaultfd.c 	if (offset >= nr_pages * page_size)
nr_pages          721 tools/testing/selftests/vm/userfaultfd.c 		split_nr_pages = (nr_pages + 1) / 2;
nr_pages          723 tools/testing/selftests/vm/userfaultfd.c 		split_nr_pages = nr_pages;
nr_pages          771 tools/testing/selftests/vm/userfaultfd.c 	area_dst = mremap(area_dst, nr_pages * page_size,  nr_pages * page_size,
nr_pages          776 tools/testing/selftests/vm/userfaultfd.c 	for (; nr < nr_pages; nr++) {
nr_pages          789 tools/testing/selftests/vm/userfaultfd.c 	for (nr = 0; nr < nr_pages; nr++) {
nr_pages          822 tools/testing/selftests/vm/userfaultfd.c 	if (offset >= nr_pages * page_size)
nr_pages          885 tools/testing/selftests/vm/userfaultfd.c 	uffdio_register.range.len = nr_pages * page_size;
nr_pages          930 tools/testing/selftests/vm/userfaultfd.c 	uffdio_register.range.len = nr_pages * page_size;
nr_pages          964 tools/testing/selftests/vm/userfaultfd.c 	return userfaults != nr_pages;
nr_pages          989 tools/testing/selftests/vm/userfaultfd.c 	uffdio_register.range.len = nr_pages * page_size;
nr_pages         1053 tools/testing/selftests/vm/userfaultfd.c 	count_verify = malloc(nr_pages * sizeof(unsigned long long));
nr_pages         1059 tools/testing/selftests/vm/userfaultfd.c 	for (nr = 0; nr < nr_pages; nr++) {
nr_pages         1120 tools/testing/selftests/vm/userfaultfd.c 		uffdio_register.range.len = nr_pages * page_size;
nr_pages         1190 tools/testing/selftests/vm/userfaultfd.c 			for (nr = 0; nr < nr_pages; nr++) {
nr_pages         1316 tools/testing/selftests/vm/userfaultfd.c 	nr_pages = nr_pages_per_cpu * nr_cpus;
nr_pages         1334 tools/testing/selftests/vm/userfaultfd.c 	       nr_pages, nr_pages_per_cpu);
nr_pages          201 tools/vm/page-types.c static unsigned long	nr_pages[HASH_SIZE];
nr_pages          435 tools/vm/page-types.c 	for (i = 0; i < ARRAY_SIZE(nr_pages); i++) {
nr_pages          436 tools/vm/page-types.c 		if (nr_pages[i])
nr_pages          439 tools/vm/page-types.c 				nr_pages[i],
nr_pages          440 tools/vm/page-types.c 				pages2mb(nr_pages[i]),
nr_pages          656 tools/vm/page-types.c 	nr_pages[hash_slot(flags)]++;
nr_pages          720 tools/vm/page-types.c 	nr_pages[hash_slot(flags)]++;
nr_pages          974 tools/vm/page-types.c 	unsigned long nr_pages, pfn, i;
nr_pages          984 tools/vm/page-types.c 		nr_pages = (end - off + page_size - 1) / page_size;
nr_pages          985 tools/vm/page-types.c 		if (nr_pages > PAGEMAP_BATCH)
nr_pages          986 tools/vm/page-types.c 			nr_pages = PAGEMAP_BATCH;
nr_pages          987 tools/vm/page-types.c 		len = nr_pages * page_size;
nr_pages         1009 tools/vm/page-types.c 		for (i = 0; i < nr_pages ; i++) {
nr_pages         1020 tools/vm/page-types.c 					nr_pages) != nr_pages)
nr_pages         1025 tools/vm/page-types.c 		for (i = 0; i < nr_pages; i++) {
nr_pages         1433 virt/kvm/kvm_main.c 				       gfn_t *nr_pages, bool write)
nr_pages         1441 virt/kvm/kvm_main.c 	if (nr_pages)
nr_pages         1442 virt/kvm/kvm_main.c 		*nr_pages = slot->npages - (gfn - slot->base_gfn);
nr_pages         1448 virt/kvm/kvm_main.c 				     gfn_t *nr_pages)
nr_pages         1450 virt/kvm/kvm_main.c 	return __gfn_to_hva_many(slot, gfn, nr_pages, true);
nr_pages         1779 virt/kvm/kvm_main.c 			    struct page **pages, int nr_pages)
nr_pages         1788 virt/kvm/kvm_main.c 	if (entry < nr_pages)
nr_pages         1791 virt/kvm/kvm_main.c 	return __get_user_pages_fast(addr, nr_pages, 1, pages);