npage              90 drivers/gpu/drm/gma500/framebuffer.c 	if (gtt->npage) {
npage             393 drivers/gpu/drm/gma500/gma_display.c 		if (gt->npage > 4)
npage             396 drivers/gpu/drm/gma500/gma_display.c 			cursor_pages = gt->npage;
npage              95 drivers/gpu/drm/gma500/gtt.c 		set_pages_array_wc(pages, r->npage);
npage              99 drivers/gpu/drm/gma500/gtt.c 	for (i = r->roll; i < r->npage; i++) {
npage             137 drivers/gpu/drm/gma500/gtt.c 	for (i = 0; i < r->npage; i++)
npage             140 drivers/gpu/drm/gma500/gtt.c 	set_pages_array_wb(r->pages, r->npage);
npage             159 drivers/gpu/drm/gma500/gtt.c 	if (roll >= r->npage) {
npage             173 drivers/gpu/drm/gma500/gtt.c 	for (i = r->roll; i < r->npage; i++) {
npage             204 drivers/gpu/drm/gma500/gtt.c 	gt->npage = gt->gem.size / PAGE_SIZE;
npage             255 drivers/gpu/drm/gma500/gtt.c 				     gt->npage, 0, 0, PSB_MMU_CACHED_MEMORY);
npage             296 drivers/gpu/drm/gma500/gtt.c 				     (gpu_base + gt->offset), gt->npage, 0, 0);
npage              39 drivers/gpu/drm/gma500/gtt.h 	int npage;			/* Number of backing pages */
npage             124 drivers/gpu/drm/i915/gvt/kvmgt.c 	int npage;
npage             129 drivers/gpu/drm/i915/gvt/kvmgt.c 	for (npage = 0; npage < total_pages; npage++) {
npage             130 drivers/gpu/drm/i915/gvt/kvmgt.c 		unsigned long cur_gfn = gfn + npage;
npage             143 drivers/gpu/drm/i915/gvt/kvmgt.c 	int npage;
npage             151 drivers/gpu/drm/i915/gvt/kvmgt.c 	for (npage = 0; npage < total_pages; npage++) {
npage             152 drivers/gpu/drm/i915/gvt/kvmgt.c 		unsigned long cur_gfn = gfn + npage;
npage             165 drivers/gpu/drm/i915/gvt/kvmgt.c 			npage++;
npage             170 drivers/gpu/drm/i915/gvt/kvmgt.c 		if (npage == 0)
npage             172 drivers/gpu/drm/i915/gvt/kvmgt.c 		else if (base_pfn + npage != pfn) {
npage             175 drivers/gpu/drm/i915/gvt/kvmgt.c 			npage++;
npage             183 drivers/gpu/drm/i915/gvt/kvmgt.c 	gvt_unpin_guest_page(vgpu, gfn, npage * PAGE_SIZE);
npage            1025 drivers/infiniband/hw/hns/hns_roce_mr.c 	int npage = 0;
npage            1062 drivers/infiniband/hw/hns/hns_roce_mr.c 		if (!(npage % (1 << (mtt->page_shift - PAGE_SHIFT)))) {
npage            1072 drivers/infiniband/hw/hns/hns_roce_mr.c 		npage++;
npage            1583 drivers/infiniband/hw/hns/hns_roce_mr.c 	int npage;
npage            1590 drivers/infiniband/hw/hns/hns_roce_mr.c 	npage = 0;
npage            1600 drivers/infiniband/hw/hns/hns_roce_mr.c 				mtts[i] = bufs[npage] >> PAGE_ADDR_SHIFT;
npage            1602 drivers/infiniband/hw/hns/hns_roce_mr.c 				mtts[i] = bufs[npage];
npage            1604 drivers/infiniband/hw/hns/hns_roce_mr.c 			npage++;
npage            1651 drivers/infiniband/hw/hns/hns_roce_mr.c 	int npage;
npage            1666 drivers/infiniband/hw/hns/hns_roce_mr.c 		npage = min(mtt_count, left);
npage            1667 drivers/infiniband/hw/hns/hns_roce_mr.c 		memcpy(&mtts[total], addr, BA_BYTE_LEN * npage);
npage            1668 drivers/infiniband/hw/hns/hns_roce_mr.c 		left -= npage;
npage            1669 drivers/infiniband/hw/hns/hns_roce_mr.c 		total += npage;
npage             162 drivers/infiniband/hw/mthca/mthca_allocator.c 	int npage = (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE;
npage             165 drivers/infiniband/hw/mthca/mthca_allocator.c 	array->page_list = kmalloc_array(npage, sizeof(*array->page_list),
npage             170 drivers/infiniband/hw/mthca/mthca_allocator.c 	for (i = 0; i < npage; ++i) {
npage            1913 drivers/vfio/vfio.c int vfio_pin_pages(struct device *dev, unsigned long *user_pfn, int npage,
npage            1921 drivers/vfio/vfio.c 	if (!dev || !user_pfn || !phys_pfn || !npage)
npage            1924 drivers/vfio/vfio.c 	if (npage > VFIO_PIN_PAGES_MAX_ENTRIES)
npage            1939 drivers/vfio/vfio.c 					     npage, prot, phys_pfn);
npage            1960 drivers/vfio/vfio.c int vfio_unpin_pages(struct device *dev, unsigned long *user_pfn, int npage)
npage            1967 drivers/vfio/vfio.c 	if (!dev || !user_pfn || !npage)
npage            1970 drivers/vfio/vfio.c 	if (npage > VFIO_PIN_PAGES_MAX_ENTRIES)
npage            1985 drivers/vfio/vfio.c 					       npage);
npage             268 drivers/vfio/vfio_iommu_type1.c static int vfio_lock_acct(struct vfio_dma *dma, long npage, bool async)
npage             273 drivers/vfio/vfio_iommu_type1.c 	if (!npage)
npage             282 drivers/vfio/vfio_iommu_type1.c 		ret = __account_locked_vm(mm, abs(npage), npage > 0, dma->task,
npage             398 drivers/vfio/vfio_iommu_type1.c 				  long npage, unsigned long *pfn_base,
npage             435 drivers/vfio/vfio_iommu_type1.c 	for (vaddr += PAGE_SIZE, iova += PAGE_SIZE; pinned < npage;
npage             477 drivers/vfio/vfio_iommu_type1.c 				    unsigned long pfn, long npage,
npage             483 drivers/vfio/vfio_iommu_type1.c 	for (i = 0; i < npage; i++, iova += PAGE_SIZE) {
npage             543 drivers/vfio/vfio_iommu_type1.c 				      int npage, int prot,
npage             574 drivers/vfio/vfio_iommu_type1.c 	for (i = 0; i < npage; i++) {
npage             629 drivers/vfio/vfio_iommu_type1.c 					int npage)
npage             645 drivers/vfio/vfio_iommu_type1.c 	for (i = 0; i < npage; i++) {
npage             658 drivers/vfio/vfio_iommu_type1.c 	return i > npage ? npage : (i > 0 ? i : -EINVAL);
npage             986 drivers/vfio/vfio_iommu_type1.c 			  unsigned long pfn, long npage, int prot)
npage             993 drivers/vfio/vfio_iommu_type1.c 				npage << PAGE_SHIFT, prot | d->prot);
npage            1004 drivers/vfio/vfio_iommu_type1.c 		iommu_unmap(d->domain, iova, npage << PAGE_SHIFT);
npage            1015 drivers/vfio/vfio_iommu_type1.c 	long npage;
npage            1021 drivers/vfio/vfio_iommu_type1.c 		npage = vfio_pin_pages_remote(dma, vaddr + dma->size,
npage            1023 drivers/vfio/vfio_iommu_type1.c 		if (npage <= 0) {
npage            1024 drivers/vfio/vfio_iommu_type1.c 			WARN_ON(!npage);
npage            1025 drivers/vfio/vfio_iommu_type1.c 			ret = (int)npage;
npage            1030 drivers/vfio/vfio_iommu_type1.c 		ret = vfio_iommu_map(iommu, iova + dma->size, pfn, npage,
npage            1034 drivers/vfio/vfio_iommu_type1.c 						npage, true);
npage            1038 drivers/vfio/vfio_iommu_type1.c 		size -= npage << PAGE_SHIFT;
npage            1039 drivers/vfio/vfio_iommu_type1.c 		dma->size += npage << PAGE_SHIFT;
npage            1235 drivers/vfio/vfio_iommu_type1.c 				long npage;
npage            1237 drivers/vfio/vfio_iommu_type1.c 				npage = vfio_pin_pages_remote(dma, vaddr,
npage            1240 drivers/vfio/vfio_iommu_type1.c 				if (npage <= 0) {
npage            1241 drivers/vfio/vfio_iommu_type1.c 					WARN_ON(!npage);
npage            1242 drivers/vfio/vfio_iommu_type1.c 					ret = (int)npage;
npage            1247 drivers/vfio/vfio_iommu_type1.c 				size = npage << PAGE_SHIFT;
npage             860 fs/f2fs/f2fs.h 		struct page *ipage, struct page *npage, nid_t nid)
npage             865 fs/f2fs/f2fs.h 	dn->node_page = npage;
npage            3500 fs/f2fs/f2fs.h bool f2fs_recover_inline_data(struct inode *inode, struct page *npage);
npage             255 fs/f2fs/inline.c bool f2fs_recover_inline_data(struct inode *inode, struct page *npage)
npage             270 fs/f2fs/inline.c 	if (IS_INODE(npage))
npage             271 fs/f2fs/inline.c 		ri = F2FS_INODE(npage);
npage             281 fs/f2fs/inline.c 		src_addr = inline_data_addr(inode, npage);
npage             725 fs/f2fs/node.c 	struct page *npage[4];
npage             738 fs/f2fs/node.c 	npage[0] = dn->inode_page;
npage             740 fs/f2fs/node.c 	if (!npage[0]) {
npage             741 fs/f2fs/node.c 		npage[0] = f2fs_get_node_page(sbi, nids[0]);
npage             742 fs/f2fs/node.c 		if (IS_ERR(npage[0]))
npage             743 fs/f2fs/node.c 			return PTR_ERR(npage[0]);
npage             749 fs/f2fs/node.c 		f2fs_put_page(npage[0], 1);
npage             753 fs/f2fs/node.c 	parent = npage[0];
npage             756 fs/f2fs/node.c 	dn->inode_page = npage[0];
npage             771 fs/f2fs/node.c 			npage[i] = f2fs_new_node_page(dn, noffset[i]);
npage             772 fs/f2fs/node.c 			if (IS_ERR(npage[i])) {
npage             774 fs/f2fs/node.c 				err = PTR_ERR(npage[i]);
npage             782 fs/f2fs/node.c 			npage[i] = f2fs_get_node_page_ra(parent, offset[i - 1]);
npage             783 fs/f2fs/node.c 			if (IS_ERR(npage[i])) {
npage             784 fs/f2fs/node.c 				err = PTR_ERR(npage[i]);
npage             797 fs/f2fs/node.c 			npage[i] = f2fs_get_node_page(sbi, nids[i]);
npage             798 fs/f2fs/node.c 			if (IS_ERR(npage[i])) {
npage             799 fs/f2fs/node.c 				err = PTR_ERR(npage[i]);
npage             800 fs/f2fs/node.c 				f2fs_put_page(npage[0], 0);
npage             805 fs/f2fs/node.c 			parent = npage[i];
npage             811 fs/f2fs/node.c 	dn->node_page = npage[level];
npage             819 fs/f2fs/node.c 		f2fs_put_page(npage[0], 0);
npage            1137 fs/f2fs/node.c 	struct page *npage;
npage            1143 fs/f2fs/node.c 	npage = f2fs_get_node_page(sbi, nid);
npage            1144 fs/f2fs/node.c 	if (IS_ERR(npage))
npage            1145 fs/f2fs/node.c 		return PTR_ERR(npage);
npage            1147 fs/f2fs/node.c 	set_new_dnode(&dn, inode, NULL, npage, nid);
npage            1150 fs/f2fs/node.c 		f2fs_put_page(npage, 1);
npage              76 include/linux/vfio.h 				     int npage, int prot,
npage              79 include/linux/vfio.h 				       unsigned long *user_pfn, int npage);
npage             106 include/linux/vfio.h 			  int npage, int prot, unsigned long *phys_pfn);
npage             108 include/linux/vfio.h 			    int npage);
npage             315 lib/kfifo.c    		struct page *npage;
npage             319 lib/kfifo.c    		npage = virt_to_page(buf);
npage             320 lib/kfifo.c    		if (page_to_phys(page) != page_to_phys(npage) - l) {
npage             325 lib/kfifo.c    			page = npage;