dpage            1323 drivers/crypto/hifn_795x.c 	struct page *spage, *dpage;
dpage            1344 drivers/crypto/hifn_795x.c 			dpage = sg_page(t);
dpage            1349 drivers/crypto/hifn_795x.c 			dpage = sg_page(dst);
dpage            1355 drivers/crypto/hifn_795x.c 		hifn_setup_dst_desc(dev, dpage, doff, len, n - len == 0);
dpage             686 drivers/dma/nbpfaxi.c 	struct nbpf_desc_page *dpage = (void *)get_zeroed_page(GFP_KERNEL | GFP_DMA);
dpage             695 drivers/dma/nbpfaxi.c 	if (!dpage)
dpage             699 drivers/dma/nbpfaxi.c 		__func__, NBPF_DESCS_PER_PAGE, NBPF_SEGMENTS_PER_PAGE, sizeof(*dpage));
dpage             701 drivers/dma/nbpfaxi.c 	for (i = 0, ldesc = dpage->ldesc, hwdesc = dpage->hwdesc;
dpage             702 drivers/dma/nbpfaxi.c 	     i < ARRAY_SIZE(dpage->ldesc);
dpage             713 drivers/dma/nbpfaxi.c 	for (i = 0, desc = dpage->desc;
dpage             714 drivers/dma/nbpfaxi.c 	     i < ARRAY_SIZE(dpage->desc);
dpage             730 drivers/dma/nbpfaxi.c 	list_add(&dpage->node, &chan->desc_page);
dpage             733 drivers/dma/nbpfaxi.c 	return ARRAY_SIZE(dpage->desc);
dpage            1059 drivers/dma/nbpfaxi.c 	struct nbpf_desc_page *dpage, *tmp;
dpage            1068 drivers/dma/nbpfaxi.c 	list_for_each_entry_safe(dpage, tmp, &chan->desc_page, node) {
dpage            1071 drivers/dma/nbpfaxi.c 		list_del(&dpage->node);
dpage            1072 drivers/dma/nbpfaxi.c 		for (i = 0, ldesc = dpage->ldesc;
dpage            1073 drivers/dma/nbpfaxi.c 		     i < ARRAY_SIZE(dpage->ldesc);
dpage            1077 drivers/dma/nbpfaxi.c 		free_page((unsigned long)dpage);
dpage             136 drivers/gpu/drm/nouveau/nouveau_dmem.c 	struct page *dpage, *spage;
dpage             142 drivers/gpu/drm/nouveau/nouveau_dmem.c 	dpage = alloc_page_vma(GFP_HIGHUSER, vmf->vma, vmf->address);
dpage             143 drivers/gpu/drm/nouveau/nouveau_dmem.c 	if (!dpage)
dpage             145 drivers/gpu/drm/nouveau/nouveau_dmem.c 	lock_page(dpage);
dpage             147 drivers/gpu/drm/nouveau/nouveau_dmem.c 	*dma_addr = dma_map_page(dev, dpage, 0, PAGE_SIZE, DMA_BIDIRECTIONAL);
dpage             155 drivers/gpu/drm/nouveau/nouveau_dmem.c 	args->dst[0] = migrate_pfn(page_to_pfn(dpage)) | MIGRATE_PFN_LOCKED;
dpage             161 drivers/gpu/drm/nouveau/nouveau_dmem.c 	__free_page(dpage);
dpage             564 drivers/gpu/drm/nouveau/nouveau_dmem.c 	struct page *dpage, *spage;
dpage             570 drivers/gpu/drm/nouveau/nouveau_dmem.c 	dpage = nouveau_dmem_page_alloc_locked(drm);
dpage             571 drivers/gpu/drm/nouveau/nouveau_dmem.c 	if (!dpage)
dpage             579 drivers/gpu/drm/nouveau/nouveau_dmem.c 			nouveau_dmem_page_addr(dpage), NOUVEAU_APER_HOST,
dpage             583 drivers/gpu/drm/nouveau/nouveau_dmem.c 	return migrate_pfn(page_to_pfn(dpage)) | MIGRATE_PFN_LOCKED;
dpage             588 drivers/gpu/drm/nouveau/nouveau_dmem.c 	nouveau_dmem_page_free_locked(drm, dpage);
dpage            1337 drivers/net/ethernet/3com/typhoon.c 	void *dpage;
dpage            1358 drivers/net/ethernet/3com/typhoon.c 	dpage = pci_alloc_consistent(pdev, PAGE_SIZE, &dpage_dma);
dpage            1359 drivers/net/ethernet/3com/typhoon.c 	if(!dpage) {
dpage            1423 drivers/net/ethernet/3com/typhoon.c 								   dpage, len,
dpage            1464 drivers/net/ethernet/3com/typhoon.c 	pci_free_consistent(pdev, PAGE_SIZE, dpage, dpage_dma);
dpage            4080 drivers/scsi/st.c 		struct page *dpage = st_bp->reserved_pages[dst_seg];
dpage            4084 drivers/scsi/st.c 		memmove(page_address(dpage) + dst_offset,
dpage             168 fs/f2fs/acl.c  						struct page *dpage)
dpage             178 fs/f2fs/acl.c  	retval = f2fs_getxattr(inode, name_index, "", NULL, 0, dpage);
dpage             184 fs/f2fs/acl.c  							retval, dpage);
dpage             334 fs/f2fs/acl.c  		struct page *dpage)
dpage             346 fs/f2fs/acl.c  	p = __f2fs_get_acl(dir, ACL_TYPE_DEFAULT, dpage);
dpage             384 fs/f2fs/acl.c  							struct page *dpage)
dpage             389 fs/f2fs/acl.c  	error = f2fs_acl_create(dir, &inode->i_mode, &default_acl, &acl, dpage);
dpage              45 fs/f2fs/acl.h  				struct page *ipage, struct page *dpage)
dpage             471 fs/f2fs/dir.c  			struct page *dpage)
dpage             493 fs/f2fs/dir.c  		err = f2fs_init_acl(inode, dir, page, dpage);
dpage            2980 fs/f2fs/f2fs.h 			const struct qstr *orig_name, struct page *dpage);
dpage             144 fs/f2fs/xattr.h 			size_t buffer_size, struct page *dpage)
dpage             104 fs/nilfs2/page.c 	struct page *spage = sbh->b_page, *dpage = dbh->b_page;
dpage             108 fs/nilfs2/page.c 	kaddr1 = kmap_atomic(dpage);
dpage             125 fs/nilfs2/page.c 		SetPageUptodate(dpage);
dpage             127 fs/nilfs2/page.c 		ClearPageUptodate(dpage);
dpage             129 fs/nilfs2/page.c 		SetPageMappedToDisk(dpage);
dpage             131 fs/nilfs2/page.c 		ClearPageMappedToDisk(dpage);
dpage             255 fs/nilfs2/page.c 		struct page *page = pvec.pages[i], *dpage;
dpage             261 fs/nilfs2/page.c 		dpage = grab_cache_page(dmap, page->index);
dpage             262 fs/nilfs2/page.c 		if (unlikely(!dpage)) {
dpage             272 fs/nilfs2/page.c 		nilfs_copy_page(dpage, page, 1);
dpage             273 fs/nilfs2/page.c 		__set_page_dirty_nobuffers(dpage);
dpage             275 fs/nilfs2/page.c 		unlock_page(dpage);
dpage             276 fs/nilfs2/page.c 		put_page(dpage);
dpage             309 fs/nilfs2/page.c 		struct page *page = pvec.pages[i], *dpage;
dpage             313 fs/nilfs2/page.c 		dpage = find_lock_page(dmap, offset);
dpage             314 fs/nilfs2/page.c 		if (dpage) {
dpage             316 fs/nilfs2/page.c 			WARN_ON(PageDirty(dpage));
dpage             317 fs/nilfs2/page.c 			nilfs_copy_page(dpage, page, 0);
dpage             318 fs/nilfs2/page.c 			unlock_page(dpage);
dpage             319 fs/nilfs2/page.c 			put_page(dpage);