hole_end 163 arch/sparc/kernel/adi_64.c unsigned long end_addr, hole_start, hole_end; hole_end 168 arch/sparc/kernel/adi_64.c hole_end = ULONG_MAX; hole_end 195 arch/sparc/kernel/adi_64.c (tag_desc->start < hole_end)) hole_end 196 arch/sparc/kernel/adi_64.c hole_end = tag_desc->start; hole_end 251 arch/sparc/kernel/adi_64.c if (hole_end < end_addr) { hole_end 258 arch/sparc/kernel/adi_64.c end_addr = hole_end - 1; hole_end 404 drivers/gpu/drm/drm_mm.c u64 hole_start, hole_end; hole_end 417 drivers/gpu/drm/drm_mm.c adj_end = hole_end = hole_start + hole->hole_size; hole_end 435 drivers/gpu/drm/drm_mm.c if (end < hole_end) hole_end 494 drivers/gpu/drm/drm_mm.c u64 hole_end = hole_start + hole->hole_size; hole_end 501 drivers/gpu/drm/drm_mm.c if (mode == DRM_MM_INSERT_HIGH && hole_end <= range_start) hole_end 505 drivers/gpu/drm/drm_mm.c col_end = hole_end; hole_end 553 drivers/gpu/drm/drm_mm.c if (adj_start + size < hole_end) hole_end 729 drivers/gpu/drm/drm_mm.c u64 hole_start, hole_end; hole_end 749 drivers/gpu/drm/drm_mm.c hole_end = __drm_mm_hole_node_end(hole); hole_end 752 drivers/gpu/drm/drm_mm.c col_end = hole_end; hole_end 790 drivers/gpu/drm/drm_mm.c DRM_MM_BUG_ON(scan->hit_end > hole_end); hole_end 860 drivers/gpu/drm/drm_mm.c u64 hole_start, hole_end; hole_end 874 drivers/gpu/drm/drm_mm.c hole_end = hole_start + hole->hole_size; hole_end 877 drivers/gpu/drm/drm_mm.c hole_end >= scan->hit_end) hole_end 887 drivers/gpu/drm/drm_mm.c DRM_MM_BUG_ON(hole_end < scan->hit_end); hole_end 889 drivers/gpu/drm/drm_mm.c mm->color_adjust(hole, scan->color, &hole_start, &hole_end); hole_end 892 drivers/gpu/drm/drm_mm.c if (hole_end < scan->hit_end) hole_end 425 drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c u64 hole_start, hole_end; hole_end 434 drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c drm_mm_for_each_hole(hole, mm, hole_start, hole_end) { hole_end 436 drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c resv.size = hole_end - hole_start - 1; /* PAGE_SIZE units */ hole_end 2680 drivers/gpu/drm/i915/i915_gem_gtt.c unsigned long hole_start, hole_end; hole_end 2715 drivers/gpu/drm/i915/i915_gem_gtt.c drm_mm_for_each_hole(entry, &ggtt->vm.mm, hole_start, hole_end) { hole_end 2717 drivers/gpu/drm/i915/i915_gem_gtt.c hole_start, hole_end); hole_end 2719 drivers/gpu/drm/i915/i915_gem_gtt.c hole_end - hole_start); hole_end 217 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c u64 hole_start, u64 hole_end, hole_end 227 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c for (size = 12; (hole_end - hole_start) >> size; size++) { hole_end 233 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c hole_size = (hole_end - hole_start) >> size; hole_end 239 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c __func__, hole_start, hole_end, size, hole_size); hole_end 253 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c GEM_BUG_ON(hole_start + count * BIT_ULL(size) > hole_end); hole_end 284 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c hole_end = hole_start; /* quit */ hole_end 344 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c u64 hole_start, u64 hole_end, hole_end 347 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c const u64 hole_size = hole_end - hole_start; hole_end 371 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c { "top-down", hole_end, -1, }, hole_end 420 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (offset + obj->base.size > hole_end) hole_end 456 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (offset + obj->base.size > hole_end) hole_end 493 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (offset + obj->base.size > hole_end) hole_end 529 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (offset + obj->base.size > hole_end) hole_end 556 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c u64 hole_start, u64 hole_end, hole_end 559 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c const u64 hole_size = hole_end - hole_start; hole_end 588 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c addr + obj->base.size < hole_end; hole_end 594 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c hole_start, hole_end, err); hole_end 640 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c u64 hole_start, u64 hole_end, hole_end 664 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c for (pot = fls64(hole_end - 1) - 1; hole_end 671 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c addr <= round_down(hole_end - 2*I915_GTT_PAGE_SIZE, step) - I915_GTT_PAGE_SIZE; hole_end 678 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c hole_start, hole_end, hole_end 700 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c __func__, pot, fls64(hole_end - 1) - 1)) { hole_end 716 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c u64 hole_start, u64 hole_end, hole_end 728 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c for (size = 12; (hole_end - hole_start) >> size; size++) { hole_end 735 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c hole_size = (hole_end - hole_start) >> size; hole_end 741 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c __func__, hole_start, hole_end, size, hole_size); hole_end 782 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c hole_start, hole_end, hole_end 826 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c u64 hole_start, u64 hole_end, hole_end 837 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c for (addr = hole_start; addr < hole_end; ) { hole_end 841 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c size = min(size, hole_end - addr); hole_end 861 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c __func__, addr, size, hole_start, hole_end, err); hole_end 880 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c __func__, addr, hole_start, hole_end)) { hole_end 893 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c u64 hole_start, u64 hole_end, hole_end 904 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = __shrink_hole(i915, vm, hole_start, hole_end, end_time); hole_end 916 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c u64 hole_start, u64 hole_end, hole_end 996 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c u64 hole_start, u64 hole_end, hole_end 1079 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c u64 hole_start, u64 hole_end, hole_end 1083 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c u64 hole_start, hole_end, last = 0; hole_end 1091 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c drm_mm_for_each_hole(node, &ggtt->vm.mm, hole_start, hole_end) { hole_end 1097 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c &hole_start, &hole_end); hole_end 1098 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (hole_start >= hole_end) hole_end 1101 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = func(i915, &ggtt->vm, hole_start, hole_end, end_time); hole_end 1106 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c last = hole_end; hole_end 1247 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c u64 hole_start, u64 hole_end, hole_end 56 drivers/gpu/drm/selftests/test-drm_mm.c u64 hole_start, hole_end; hole_end 60 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_hole(hole, mm, hole_start, hole_end) hole_end 80 drivers/gpu/drm/selftests/test-drm_mm.c u64 hole_start, hole_end; hole_end 88 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_hole(hole, mm, hole_start, hole_end) { hole_end 89 drivers/gpu/drm/selftests/test-drm_mm.c if (start != hole_start || end != hole_end) { hole_end 92 drivers/gpu/drm/selftests/test-drm_mm.c hole_start, hole_end, hole_end 1152 drivers/gpu/drm/selftests/test-drm_mm.c u64 hole_start, hole_end; hole_end 1155 drivers/gpu/drm/selftests/test-drm_mm.c drm_mm_for_each_hole(hole, mm, hole_start, hole_end) { hole_end 1171 drivers/gpu/drm/selftests/test-drm_mm.c hole_start, hole_end, hole_end - hole_start, hole_end 7310 fs/btrfs/inode.c const u64 hole_end = extent_map_end(hole_em); hole_end 7327 fs/btrfs/inode.c if (hole_end <= start || hole_em->start > end) { hole_end 7332 fs/btrfs/inode.c hole_len = hole_end - hole_start; hole_end 536 fs/hugetlbfs/inode.c loff_t hole_start, hole_end; hole_end 543 fs/hugetlbfs/inode.c hole_end = round_down(offset + len, hpage_size); hole_end 545 fs/hugetlbfs/inode.c if (hole_end > hole_start) { hole_end 561 fs/hugetlbfs/inode.c hole_end >> PAGE_SHIFT); hole_end 563 fs/hugetlbfs/inode.c remove_inode_hugepages(inode, hole_start, hole_end); hole_end 1295 fs/proc/task_mmu.c unsigned long hole_end; hole_end 1298 fs/proc/task_mmu.c hole_end = min(end, vma->vm_start); hole_end 1300 fs/proc/task_mmu.c hole_end = end; hole_end 1302 fs/proc/task_mmu.c for (; addr < hole_end; addr += PAGE_SIZE) { hole_end 390 include/drm/drm_mm.h #define drm_mm_for_each_hole(pos, mm, hole_start, hole_end) \ hole_end 395 include/drm/drm_mm.h hole_end = hole_start + pos->hole_size, \ hole_end 440 kernel/kexec_core.c unsigned long hole_start, hole_end, size; hole_end 446 kernel/kexec_core.c hole_end = hole_start + size - 1; hole_end 447 kernel/kexec_core.c while (hole_end <= crashk_res.end) { hole_end 452 kernel/kexec_core.c if (hole_end > KEXEC_CRASH_CONTROL_MEMORY_LIMIT) hole_end 460 kernel/kexec_core.c if ((hole_end >= mstart) && (hole_start <= mend)) { hole_end 463 kernel/kexec_core.c hole_end = hole_start + size - 1; hole_end 470 kernel/kexec_core.c image->control_page = hole_end;