cur_end           336 arch/mips/kvm/mmu.c 	unsigned long cur_end = ~0ul;					\
cur_end           347 arch/mips/kvm/mmu.c 			cur_end = end;					\
cur_end           349 arch/mips/kvm/mmu.c 		ret |= kvm_mips_##name##_pte(pte, start, cur_end);	\
cur_end           359 arch/mips/kvm/mmu.c 	unsigned long cur_end = ~0ul;					\
cur_end           370 arch/mips/kvm/mmu.c 			cur_end = end;					\
cur_end           372 arch/mips/kvm/mmu.c 		ret |= kvm_mips_##name##_pmd(pmd, start, cur_end);	\
cur_end           382 arch/mips/kvm/mmu.c 	unsigned long cur_end = ~0ul;					\
cur_end           393 arch/mips/kvm/mmu.c 			cur_end = end;					\
cur_end           395 arch/mips/kvm/mmu.c 		ret |= kvm_mips_##name##_pud(pud, start, cur_end);	\
cur_end          3070 fs/btrfs/extent_io.c 	u64 cur_end;
cur_end          3140 fs/btrfs/extent_io.c 		cur_end = min(extent_map_end(em) - 1, end);
cur_end          3218 fs/btrfs/extent_io.c 		if (test_range_bit(tree, cur, cur_end,
cur_end          1218 fs/btrfs/inode.c 	u64 cur_end;
cur_end          1256 fs/btrfs/inode.c 			cur_end = min(end, start + SZ_512K - 1);
cur_end          1258 fs/btrfs/inode.c 			cur_end = end;
cur_end          1268 fs/btrfs/inode.c 		async_chunk[i].end = cur_end;
cur_end          1291 fs/btrfs/inode.c 		nr_pages = DIV_ROUND_UP(cur_end - start, PAGE_SIZE);
cur_end          1297 fs/btrfs/inode.c 		start = cur_end + 1;