page_block 1662 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (unlikely((getref || vmm->func->page_block) && !shift)) { page_block 1664 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c getref, vmm->func->page_block); page_block 1711 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (vmm->func->page_block && prev && prev->page != p) page_block 1712 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c addr = ALIGN(addr, vmm->func->page_block); page_block 1716 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (vmm->func->page_block && next && next->page != p) page_block 1717 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c tail = ALIGN_DOWN(tail, vmm->func->page_block); page_block 153 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h u64 page_block; page_block 30 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmmcp77.c .page_block = 1 << 29, page_block 370 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c .page_block = 1 << 29, page_block 240 fs/ext4/readpage.c unsigned page_block; page_block 274 fs/ext4/readpage.c page_block = 0; page_block 291 fs/ext4/readpage.c if (page_block == blocks_per_page) page_block 293 fs/ext4/readpage.c blocks[page_block] = map.m_pblk + map_offset + page_block 295 fs/ext4/readpage.c page_block++; page_block 304 fs/ext4/readpage.c while (page_block < blocks_per_page) { page_block 321 fs/ext4/readpage.c first_hole = page_block; page_block 322 fs/ext4/readpage.c page_block++; page_block 330 fs/ext4/readpage.c if (page_block && blocks[page_block-1] != map.m_pblk-1) page_block 337 fs/ext4/readpage.c } else if (page_block == blocks_per_page) page_block 339 fs/ext4/readpage.c blocks[page_block] = map.m_pblk+relative_block; page_block 340 fs/ext4/readpage.c page_block++; page_block 104 fs/mpage.c map_buffer_to_page(struct page *page, struct buffer_head *bh, int page_block) page_block 125 fs/mpage.c if (block == page_block) { page_block 168 fs/mpage.c unsigned page_block; page_block 194 fs/mpage.c page_block = 0; page_block 211 fs/mpage.c if (page_block == blocks_per_page) page_block 213 fs/mpage.c blocks[page_block] = map_bh->b_blocknr + map_offset + page_block 215 fs/mpage.c page_block++; page_block 225 fs/mpage.c while (page_block < blocks_per_page) { page_block 239 fs/mpage.c first_hole = page_block; page_block 240 fs/mpage.c page_block++; page_block 252 fs/mpage.c map_buffer_to_page(page, map_bh, page_block); page_block 260 fs/mpage.c if (page_block && blocks[page_block-1] != map_bh->b_blocknr-1) page_block 267 fs/mpage.c } else if (page_block == blocks_per_page) page_block 269 fs/mpage.c blocks[page_block] = map_bh->b_blocknr+relative_block; page_block 270 fs/mpage.c page_block++; page_block 509 fs/mpage.c unsigned page_block; page_block 526 fs/mpage.c page_block = 0; page_block 537 fs/mpage.c first_unmapped = page_block; page_block 546 fs/mpage.c if (page_block) { page_block 547 fs/mpage.c if (bh->b_blocknr != blocks[page_block-1] + 1) page_block 550 fs/mpage.c blocks[page_block++] = bh->b_blocknr; page_block 578 fs/mpage.c for (page_block = 0; page_block < blocks_per_page; ) { page_block 590 fs/mpage.c if (page_block) { page_block 591 fs/mpage.c if (map_bh.b_blocknr != blocks[page_block-1] + 1) page_block 594 fs/mpage.c blocks[page_block++] = map_bh.b_blocknr; page_block 601 fs/mpage.c BUG_ON(page_block == 0); page_block 603 fs/mpage.c first_unmapped = page_block; page_block 612 fs/ubifs/file.c unsigned int page_block; page_block 628 fs/ubifs/file.c page_block = page->index << UBIFS_BLOCKS_PER_PAGE_SHIFT; page_block 635 fs/ubifs/file.c } else if (key_block(c, &bu->zbranch[nn].key) == page_block) { page_block 651 fs/ubifs/file.c err = ubifs_decrypt(inode, dn, &dlen, page_block); page_block 666 fs/ubifs/file.c } else if (key_block(c, &bu->zbranch[nn].key) < page_block) { page_block 676 fs/ubifs/file.c page_block += 1; page_block 705 fs/ubifs/file.c page_block, inode->i_ino);