npage 90 drivers/gpu/drm/gma500/framebuffer.c if (gtt->npage) { npage 393 drivers/gpu/drm/gma500/gma_display.c if (gt->npage > 4) npage 396 drivers/gpu/drm/gma500/gma_display.c cursor_pages = gt->npage; npage 95 drivers/gpu/drm/gma500/gtt.c set_pages_array_wc(pages, r->npage); npage 99 drivers/gpu/drm/gma500/gtt.c for (i = r->roll; i < r->npage; i++) { npage 137 drivers/gpu/drm/gma500/gtt.c for (i = 0; i < r->npage; i++) npage 140 drivers/gpu/drm/gma500/gtt.c set_pages_array_wb(r->pages, r->npage); npage 159 drivers/gpu/drm/gma500/gtt.c if (roll >= r->npage) { npage 173 drivers/gpu/drm/gma500/gtt.c for (i = r->roll; i < r->npage; i++) { npage 204 drivers/gpu/drm/gma500/gtt.c gt->npage = gt->gem.size / PAGE_SIZE; npage 255 drivers/gpu/drm/gma500/gtt.c gt->npage, 0, 0, PSB_MMU_CACHED_MEMORY); npage 296 drivers/gpu/drm/gma500/gtt.c (gpu_base + gt->offset), gt->npage, 0, 0); npage 39 drivers/gpu/drm/gma500/gtt.h int npage; /* Number of backing pages */ npage 124 drivers/gpu/drm/i915/gvt/kvmgt.c int npage; npage 129 drivers/gpu/drm/i915/gvt/kvmgt.c for (npage = 0; npage < total_pages; npage++) { npage 130 drivers/gpu/drm/i915/gvt/kvmgt.c unsigned long cur_gfn = gfn + npage; npage 143 drivers/gpu/drm/i915/gvt/kvmgt.c int npage; npage 151 drivers/gpu/drm/i915/gvt/kvmgt.c for (npage = 0; npage < total_pages; npage++) { npage 152 drivers/gpu/drm/i915/gvt/kvmgt.c unsigned long cur_gfn = gfn + npage; npage 165 drivers/gpu/drm/i915/gvt/kvmgt.c npage++; npage 170 drivers/gpu/drm/i915/gvt/kvmgt.c if (npage == 0) npage 172 drivers/gpu/drm/i915/gvt/kvmgt.c else if (base_pfn + npage != pfn) { npage 175 drivers/gpu/drm/i915/gvt/kvmgt.c npage++; npage 183 drivers/gpu/drm/i915/gvt/kvmgt.c gvt_unpin_guest_page(vgpu, gfn, npage * PAGE_SIZE); npage 1025 drivers/infiniband/hw/hns/hns_roce_mr.c int npage = 0; npage 1062 drivers/infiniband/hw/hns/hns_roce_mr.c if (!(npage % (1 << (mtt->page_shift - PAGE_SHIFT)))) { npage 1072 drivers/infiniband/hw/hns/hns_roce_mr.c npage++; npage 1583 drivers/infiniband/hw/hns/hns_roce_mr.c int npage; npage 1590 drivers/infiniband/hw/hns/hns_roce_mr.c npage = 0; npage 1600 drivers/infiniband/hw/hns/hns_roce_mr.c mtts[i] = bufs[npage] >> PAGE_ADDR_SHIFT; npage 1602 drivers/infiniband/hw/hns/hns_roce_mr.c mtts[i] = bufs[npage]; npage 1604 drivers/infiniband/hw/hns/hns_roce_mr.c npage++; npage 1651 drivers/infiniband/hw/hns/hns_roce_mr.c int npage; npage 1666 drivers/infiniband/hw/hns/hns_roce_mr.c npage = min(mtt_count, left); npage 1667 drivers/infiniband/hw/hns/hns_roce_mr.c memcpy(&mtts[total], addr, BA_BYTE_LEN * npage); npage 1668 drivers/infiniband/hw/hns/hns_roce_mr.c left -= npage; npage 1669 drivers/infiniband/hw/hns/hns_roce_mr.c total += npage; npage 162 drivers/infiniband/hw/mthca/mthca_allocator.c int npage = (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; npage 165 drivers/infiniband/hw/mthca/mthca_allocator.c array->page_list = kmalloc_array(npage, sizeof(*array->page_list), npage 170 drivers/infiniband/hw/mthca/mthca_allocator.c for (i = 0; i < npage; ++i) { npage 1913 drivers/vfio/vfio.c int vfio_pin_pages(struct device *dev, unsigned long *user_pfn, int npage, npage 1921 drivers/vfio/vfio.c if (!dev || !user_pfn || !phys_pfn || !npage) npage 1924 drivers/vfio/vfio.c if (npage > VFIO_PIN_PAGES_MAX_ENTRIES) npage 1939 drivers/vfio/vfio.c npage, prot, phys_pfn); npage 1960 drivers/vfio/vfio.c int vfio_unpin_pages(struct device *dev, unsigned long *user_pfn, int npage) npage 1967 drivers/vfio/vfio.c if (!dev || !user_pfn || !npage) npage 1970 drivers/vfio/vfio.c if (npage > VFIO_PIN_PAGES_MAX_ENTRIES) npage 1985 drivers/vfio/vfio.c npage); npage 268 drivers/vfio/vfio_iommu_type1.c static int vfio_lock_acct(struct vfio_dma *dma, long npage, bool async) npage 273 drivers/vfio/vfio_iommu_type1.c if (!npage) npage 282 drivers/vfio/vfio_iommu_type1.c ret = __account_locked_vm(mm, abs(npage), npage > 0, dma->task, npage 398 drivers/vfio/vfio_iommu_type1.c long npage, unsigned long *pfn_base, npage 435 drivers/vfio/vfio_iommu_type1.c for (vaddr += PAGE_SIZE, iova += PAGE_SIZE; pinned < npage; npage 477 drivers/vfio/vfio_iommu_type1.c unsigned long pfn, long npage, npage 483 drivers/vfio/vfio_iommu_type1.c for (i = 0; i < npage; i++, iova += PAGE_SIZE) { npage 543 drivers/vfio/vfio_iommu_type1.c int npage, int prot, npage 574 drivers/vfio/vfio_iommu_type1.c for (i = 0; i < npage; i++) { npage 629 drivers/vfio/vfio_iommu_type1.c int npage) npage 645 drivers/vfio/vfio_iommu_type1.c for (i = 0; i < npage; i++) { npage 658 drivers/vfio/vfio_iommu_type1.c return i > npage ? npage : (i > 0 ? i : -EINVAL); npage 986 drivers/vfio/vfio_iommu_type1.c unsigned long pfn, long npage, int prot) npage 993 drivers/vfio/vfio_iommu_type1.c npage << PAGE_SHIFT, prot | d->prot); npage 1004 drivers/vfio/vfio_iommu_type1.c iommu_unmap(d->domain, iova, npage << PAGE_SHIFT); npage 1015 drivers/vfio/vfio_iommu_type1.c long npage; npage 1021 drivers/vfio/vfio_iommu_type1.c npage = vfio_pin_pages_remote(dma, vaddr + dma->size, npage 1023 drivers/vfio/vfio_iommu_type1.c if (npage <= 0) { npage 1024 drivers/vfio/vfio_iommu_type1.c WARN_ON(!npage); npage 1025 drivers/vfio/vfio_iommu_type1.c ret = (int)npage; npage 1030 drivers/vfio/vfio_iommu_type1.c ret = vfio_iommu_map(iommu, iova + dma->size, pfn, npage, npage 1034 drivers/vfio/vfio_iommu_type1.c npage, true); npage 1038 drivers/vfio/vfio_iommu_type1.c size -= npage << PAGE_SHIFT; npage 1039 drivers/vfio/vfio_iommu_type1.c dma->size += npage << PAGE_SHIFT; npage 1235 drivers/vfio/vfio_iommu_type1.c long npage; npage 1237 drivers/vfio/vfio_iommu_type1.c npage = vfio_pin_pages_remote(dma, vaddr, npage 1240 drivers/vfio/vfio_iommu_type1.c if (npage <= 0) { npage 1241 drivers/vfio/vfio_iommu_type1.c WARN_ON(!npage); npage 1242 drivers/vfio/vfio_iommu_type1.c ret = (int)npage; npage 1247 drivers/vfio/vfio_iommu_type1.c size = npage << PAGE_SHIFT; npage 860 fs/f2fs/f2fs.h struct page *ipage, struct page *npage, nid_t nid) npage 865 fs/f2fs/f2fs.h dn->node_page = npage; npage 3500 fs/f2fs/f2fs.h bool f2fs_recover_inline_data(struct inode *inode, struct page *npage); npage 255 fs/f2fs/inline.c bool f2fs_recover_inline_data(struct inode *inode, struct page *npage) npage 270 fs/f2fs/inline.c if (IS_INODE(npage)) npage 271 fs/f2fs/inline.c ri = F2FS_INODE(npage); npage 281 fs/f2fs/inline.c src_addr = inline_data_addr(inode, npage); npage 725 fs/f2fs/node.c struct page *npage[4]; npage 738 fs/f2fs/node.c npage[0] = dn->inode_page; npage 740 fs/f2fs/node.c if (!npage[0]) { npage 741 fs/f2fs/node.c npage[0] = f2fs_get_node_page(sbi, nids[0]); npage 742 fs/f2fs/node.c if (IS_ERR(npage[0])) npage 743 fs/f2fs/node.c return PTR_ERR(npage[0]); npage 749 fs/f2fs/node.c f2fs_put_page(npage[0], 1); npage 753 fs/f2fs/node.c parent = npage[0]; npage 756 fs/f2fs/node.c dn->inode_page = npage[0]; npage 771 fs/f2fs/node.c npage[i] = f2fs_new_node_page(dn, noffset[i]); npage 772 fs/f2fs/node.c if (IS_ERR(npage[i])) { npage 774 fs/f2fs/node.c err = PTR_ERR(npage[i]); npage 782 fs/f2fs/node.c npage[i] = f2fs_get_node_page_ra(parent, offset[i - 1]); npage 783 fs/f2fs/node.c if (IS_ERR(npage[i])) { npage 784 fs/f2fs/node.c err = PTR_ERR(npage[i]); npage 797 fs/f2fs/node.c npage[i] = f2fs_get_node_page(sbi, nids[i]); npage 798 fs/f2fs/node.c if (IS_ERR(npage[i])) { npage 799 fs/f2fs/node.c err = PTR_ERR(npage[i]); npage 800 fs/f2fs/node.c f2fs_put_page(npage[0], 0); npage 805 fs/f2fs/node.c parent = npage[i]; npage 811 fs/f2fs/node.c dn->node_page = npage[level]; npage 819 fs/f2fs/node.c f2fs_put_page(npage[0], 0); npage 1137 fs/f2fs/node.c struct page *npage; npage 1143 fs/f2fs/node.c npage = f2fs_get_node_page(sbi, nid); npage 1144 fs/f2fs/node.c if (IS_ERR(npage)) npage 1145 fs/f2fs/node.c return PTR_ERR(npage); npage 1147 fs/f2fs/node.c set_new_dnode(&dn, inode, NULL, npage, nid); npage 1150 fs/f2fs/node.c f2fs_put_page(npage, 1); npage 76 include/linux/vfio.h int npage, int prot, npage 79 include/linux/vfio.h unsigned long *user_pfn, int npage); npage 106 include/linux/vfio.h int npage, int prot, unsigned long *phys_pfn); npage 108 include/linux/vfio.h int npage); npage 315 lib/kfifo.c struct page *npage; npage 319 lib/kfifo.c npage = virt_to_page(buf); npage 320 lib/kfifo.c if (page_to_phys(page) != page_to_phys(npage) - l) { npage 325 lib/kfifo.c page = npage;