cur_end 336 arch/mips/kvm/mmu.c unsigned long cur_end = ~0ul; \ cur_end 347 arch/mips/kvm/mmu.c cur_end = end; \ cur_end 349 arch/mips/kvm/mmu.c ret |= kvm_mips_##name##_pte(pte, start, cur_end); \ cur_end 359 arch/mips/kvm/mmu.c unsigned long cur_end = ~0ul; \ cur_end 370 arch/mips/kvm/mmu.c cur_end = end; \ cur_end 372 arch/mips/kvm/mmu.c ret |= kvm_mips_##name##_pmd(pmd, start, cur_end); \ cur_end 382 arch/mips/kvm/mmu.c unsigned long cur_end = ~0ul; \ cur_end 393 arch/mips/kvm/mmu.c cur_end = end; \ cur_end 395 arch/mips/kvm/mmu.c ret |= kvm_mips_##name##_pud(pud, start, cur_end); \ cur_end 3070 fs/btrfs/extent_io.c u64 cur_end; cur_end 3140 fs/btrfs/extent_io.c cur_end = min(extent_map_end(em) - 1, end); cur_end 3218 fs/btrfs/extent_io.c if (test_range_bit(tree, cur, cur_end, cur_end 1218 fs/btrfs/inode.c u64 cur_end; cur_end 1256 fs/btrfs/inode.c cur_end = min(end, start + SZ_512K - 1); cur_end 1258 fs/btrfs/inode.c cur_end = end; cur_end 1268 fs/btrfs/inode.c async_chunk[i].end = cur_end; cur_end 1291 fs/btrfs/inode.c nr_pages = DIV_ROUND_UP(cur_end - start, PAGE_SIZE); cur_end 1297 fs/btrfs/inode.c start = cur_end + 1;