ggtt 172 drivers/gpu/drm/i915/display/intel_fbdev.c struct i915_ggtt *ggtt = &dev_priv->ggtt; ggtt 235 drivers/gpu/drm/i915/display/intel_fbdev.c info->apertures->ranges[0].base = ggtt->gmadr.start; ggtt 236 drivers/gpu/drm/i915/display/intel_fbdev.c info->apertures->ranges[0].size = ggtt->mappable_end; ggtt 240 drivers/gpu/drm/i915/display/intel_fbdev.c (unsigned long)(ggtt->gmadr.start + vma->node.start); ggtt 2235 drivers/gpu/drm/i915/gem/i915_gem_context.c else if (to_i915(dev)->ggtt.alias) ggtt 2236 drivers/gpu/drm/i915/gem/i915_gem_context.c args->value = to_i915(dev)->ggtt.alias->vm.total; ggtt 2238 drivers/gpu/drm/i915/gem/i915_gem_context.c args->value = to_i915(dev)->ggtt.vm.total; ggtt 249 drivers/gpu/drm/i915/gem/i915_gem_domain.c ret = mutex_lock_interruptible(&i915->ggtt.vm.mutex); ggtt 272 drivers/gpu/drm/i915/gem/i915_gem_domain.c mutex_unlock(&i915->ggtt.vm.mutex); ggtt 489 drivers/gpu/drm/i915/gem/i915_gem_domain.c mutex_lock(&i915->ggtt.vm.mutex); ggtt 496 drivers/gpu/drm/i915/gem/i915_gem_domain.c mutex_unlock(&i915->ggtt.vm.mutex); ggtt 929 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c return &i915->ggtt; ggtt 966 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c struct i915_ggtt *ggtt = cache_to_ggtt(cache); ggtt 968 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c intel_gt_flush_ggtt_writes(ggtt->vm.gt); ggtt 972 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c ggtt->vm.clear_range(&ggtt->vm, ggtt 1021 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c struct i915_ggtt *ggtt = cache_to_ggtt(cache); ggtt 1026 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c intel_gt_flush_ggtt_writes(ggtt->vm.gt); ggtt 1051 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c (&ggtt->vm.mm, &cache->node, ggtt 1053 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 0, ggtt->mappable_end, ggtt 1065 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c ggtt->vm.insert_page(&ggtt->vm, ggtt 1072 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c vaddr = (void __force *)io_mapping_map_atomic_wc(&ggtt->iomap, ggtt 1977 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c vm = &dev_priv->ggtt.vm; ggtt 225 drivers/gpu/drm/i915/gem/i915_gem_mman.c struct i915_ggtt *ggtt = &i915->ggtt; ggtt 248 drivers/gpu/drm/i915/gem/i915_gem_mman.c ret = intel_gt_reset_trylock(ggtt->vm.gt, &srcu); ggtt 301 drivers/gpu/drm/i915/gem/i915_gem_mman.c (ggtt->gmadr.start + vma->node.start) >> PAGE_SHIFT, ggtt 303 drivers/gpu/drm/i915/gem/i915_gem_mman.c &ggtt->iomap); ggtt 310 drivers/gpu/drm/i915/gem/i915_gem_mman.c mutex_lock(&i915->ggtt.vm.mutex); ggtt 312 drivers/gpu/drm/i915/gem/i915_gem_mman.c list_add(&obj->userfault_link, &i915->ggtt.userfault_list); ggtt 313 drivers/gpu/drm/i915/gem/i915_gem_mman.c mutex_unlock(&i915->ggtt.vm.mutex); ggtt 316 drivers/gpu/drm/i915/gem/i915_gem_mman.c intel_wakeref_auto(&i915->ggtt.userfault_wakeref, ggtt 332 drivers/gpu/drm/i915/gem/i915_gem_mman.c intel_gt_reset_unlock(ggtt->vm.gt, srcu); ggtt 345 drivers/gpu/drm/i915/gem/i915_gem_mman.c if (!intel_gt_is_wedged(ggtt->vm.gt)) ggtt 418 drivers/gpu/drm/i915/gem/i915_gem_mman.c mutex_lock(&i915->ggtt.vm.mutex); ggtt 435 drivers/gpu/drm/i915/gem/i915_gem_mman.c mutex_unlock(&i915->ggtt.vm.mutex); ggtt 147 drivers/gpu/drm/i915/gem/i915_gem_pm.c intel_wakeref_auto(&i915->ggtt.userfault_wakeref, 0); ggtt 441 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c mutex_lock(&i915->ggtt.vm.mutex); ggtt 443 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c &i915->ggtt.vm.bound_list, vm_link) { ggtt 449 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c mutex_unlock(&i915->ggtt.vm.mutex); ggtt 452 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c mutex_lock(&i915->ggtt.vm.mutex); ggtt 454 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c mutex_unlock(&i915->ggtt.vm.mutex); ggtt 69 drivers/gpu/drm/i915/gem/i915_gem_stolen.c struct i915_ggtt *ggtt = &dev_priv->ggtt; ggtt 96 drivers/gpu/drm/i915/gem/i915_gem_stolen.c ggtt_total_entries(ggtt) * 4); ggtt 615 drivers/gpu/drm/i915/gem/i915_gem_stolen.c struct i915_ggtt *ggtt = &dev_priv->ggtt; ggtt 666 drivers/gpu/drm/i915/gem/i915_gem_stolen.c vma = i915_vma_instance(obj, &ggtt->vm, NULL); ggtt 677 drivers/gpu/drm/i915/gem/i915_gem_stolen.c ret = i915_gem_gtt_reserve(&ggtt->vm, &vma->node, ggtt 691 drivers/gpu/drm/i915/gem/i915_gem_stolen.c mutex_lock(&ggtt->vm.mutex); ggtt 692 drivers/gpu/drm/i915/gem/i915_gem_stolen.c list_move_tail(&vma->vm_link, &ggtt->vm.bound_list); ggtt 693 drivers/gpu/drm/i915/gem/i915_gem_stolen.c mutex_unlock(&ggtt->vm.mutex); ggtt 938 drivers/gpu/drm/i915/gem/selftests/huge_pages.c struct i915_address_space *vm = ctx->vm ?: &engine->gt->ggtt->vm; ggtt 991 drivers/gpu/drm/i915/gem/selftests/huge_pages.c struct i915_address_space *vm = ctx->vm ?: &i915->ggtt.vm; ggtt 1451 drivers/gpu/drm/i915/gem/selftests/huge_pages.c struct i915_address_space *vm = ctx->vm ?: &i915->ggtt.vm; ggtt 1508 drivers/gpu/drm/i915/gem/selftests/huge_pages.c struct i915_address_space *vm = ctx->vm ?: &i915->ggtt.vm; ggtt 174 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c struct i915_address_space *vm = ctx->vm ?: &engine->gt->ggtt->vm; ggtt 313 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c struct i915_address_space *vm = ctx->vm ?: &ctx->i915->ggtt.vm; ggtt 1076 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c vm = ctx->vm ?: &i915->ggtt.alias->vm; ggtt 197 drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c (1 + next_prime_number(i915->ggtt.vm.total >> PAGE_SHIFT)) << PAGE_SHIFT); ggtt 336 drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL); ggtt 46 drivers/gpu/drm/i915/gem/selftests/i915_gem_object.c i915->ggtt.vm.total + PAGE_SIZE); ggtt 111 drivers/gpu/drm/i915/gem/selftests/igt_gem_utils.c struct i915_address_space *vm = ctx->vm ?: &engine->gt->ggtt->vm; ggtt 227 drivers/gpu/drm/i915/gt/intel_context.c ce->vm = i915_vm_get(ctx->vm ?: &engine->gt->ggtt->vm); ggtt 557 drivers/gpu/drm/i915/gt/intel_engine_cs.c vma = i915_vma_instance(obj, &engine->gt->ggtt->vm, NULL); ggtt 29 drivers/gpu/drm/i915/gt/intel_gt.c i915->gt.ggtt = &i915->ggtt; ggtt 240 drivers/gpu/drm/i915/gt/intel_gt.c vma = i915_vma_instance(obj, >->ggtt->vm, NULL); ggtt 39 drivers/gpu/drm/i915/gt/intel_gt_types.h struct i915_ggtt *ggtt; ggtt 2224 drivers/gpu/drm/i915/gt/intel_lrc.c vma = i915_vma_instance(obj, &engine->gt->ggtt->vm, NULL); ggtt 3371 drivers/gpu/drm/i915/gt/intel_lrc.c vma = i915_vma_instance(ctx_obj, &engine->gt->ggtt->vm, NULL); ggtt 196 drivers/gpu/drm/i915/gt/intel_renderstate.c so.vma = i915_vma_instance(so.obj, &engine->gt->ggtt->vm, NULL); ggtt 627 drivers/gpu/drm/i915/gt/intel_reset.c for (i = 0; i < gt->ggtt->num_fences; i++) { ggtt 632 drivers/gpu/drm/i915/gt/intel_reset.c vma = READ_ONCE(gt->ggtt->fence_regs[i].vma); ggtt 639 drivers/gpu/drm/i915/gt/intel_reset.c GEM_BUG_ON(vma->fence != >->ggtt->fence_regs[i]); ggtt 1267 drivers/gpu/drm/i915/gt/intel_ringbuffer.c static struct i915_vma *create_ring_vma(struct i915_ggtt *ggtt, int size) ggtt 1269 drivers/gpu/drm/i915/gt/intel_ringbuffer.c struct i915_address_space *vm = &ggtt->vm; ggtt 1327 drivers/gpu/drm/i915/gt/intel_ringbuffer.c vma = create_ring_vma(engine->gt->ggtt, size); ggtt 1455 drivers/gpu/drm/i915/gt/intel_ringbuffer.c vma = i915_vma_instance(obj, &engine->gt->ggtt->vm, NULL); ggtt 46 drivers/gpu/drm/i915/gt/intel_timeline.c vma = i915_vma_instance(obj, >->ggtt->vm, NULL); ggtt 1514 drivers/gpu/drm/i915/gt/intel_workarounds.c vma = create_scratch(&ce->engine->gt->ggtt->vm, wal->count); ggtt 134 drivers/gpu/drm/i915/gt/selftest_hangcheck.c struct i915_address_space *vm = h->ctx->vm ?: &engine->gt->ggtt->vm; ggtt 1335 drivers/gpu/drm/i915/gt/selftest_hangcheck.c return __igt_reset_evict_vma(gt, >->ggtt->vm, ggtt 1372 drivers/gpu/drm/i915/gt/selftest_hangcheck.c return __igt_reset_evict_vma(gt, >->ggtt->vm, ggtt 252 drivers/gpu/drm/i915/gt/selftest_lrc.c vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL); ggtt 346 drivers/gpu/drm/i915/gt/selftest_lrc.c vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL); ggtt 99 drivers/gpu/drm/i915/gt/selftest_workarounds.c vma = i915_vma_instance(result, &engine->gt->ggtt->vm, NULL); ggtt 289 drivers/gpu/drm/i915/gt/uc/intel_guc.c i915_ggtt_enable_guc(gt->ggtt); ggtt 316 drivers/gpu/drm/i915/gt/uc/intel_guc.c i915_ggtt_disable_guc(gt->ggtt); ggtt 603 drivers/gpu/drm/i915/gt/uc/intel_guc.c vma = i915_vma_instance(obj, >->ggtt->vm, NULL); ggtt 399 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c static u32 uc_fw_ggtt_offset(struct intel_uc_fw *uc_fw, struct i915_ggtt *ggtt) ggtt 401 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c struct drm_mm_node *node = &ggtt->uc_fw; ggtt 414 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c struct i915_ggtt *ggtt = gt->ggtt; ggtt 416 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c .node.start = uc_fw_ggtt_offset(uc_fw, ggtt), ggtt 419 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c .vm = &ggtt->vm, ggtt 423 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c GEM_BUG_ON(dummy.node.size > ggtt->uc_fw.size); ggtt 428 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c ggtt->vm.insert_entries(&ggtt->vm, &dummy, I915_CACHE_NONE, 0); ggtt 435 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c struct i915_ggtt *ggtt = gt->ggtt; ggtt 436 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c u64 start = uc_fw_ggtt_offset(uc_fw, ggtt); ggtt 438 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c ggtt->vm.clear_range(&ggtt->vm, start, obj->base.size); ggtt 455 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c offset = uc_fw_ggtt_offset(uc_fw, gt->ggtt); ggtt 66 drivers/gpu/drm/i915/gvt/aperture_gm.c ret = i915_gem_gtt_insert(&dev_priv->ggtt.vm, node, ggtt 175 drivers/gpu/drm/i915/gvt/aperture_gm.c mutex_lock(&dev_priv->ggtt.vm.mutex); ggtt 182 drivers/gpu/drm/i915/gvt/aperture_gm.c mutex_unlock(&dev_priv->ggtt.vm.mutex); ggtt 198 drivers/gpu/drm/i915/gvt/aperture_gm.c mutex_lock(&dev_priv->ggtt.vm.mutex); ggtt 210 drivers/gpu/drm/i915/gvt/aperture_gm.c mutex_unlock(&dev_priv->ggtt.vm.mutex); ggtt 223 drivers/gpu/drm/i915/gvt/aperture_gm.c mutex_unlock(&dev_priv->ggtt.vm.mutex); ggtt 64 drivers/gpu/drm/i915/gvt/dmabuf.c gtt_entries = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + ggtt 280 drivers/gpu/drm/i915/gvt/gtt.c void __iomem *addr = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + index; ggtt 295 drivers/gpu/drm/i915/gvt/gtt.c void __iomem *addr = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + index; ggtt 380 drivers/gpu/drm/i915/gvt/gvt.h #define gvt_aperture_sz(gvt) (gvt->dev_priv->ggtt.mappable_end) ggtt 381 drivers/gpu/drm/i915/gvt/gvt.h #define gvt_aperture_pa_base(gvt) (gvt->dev_priv->ggtt.gmadr.start) ggtt 383 drivers/gpu/drm/i915/gvt/gvt.h #define gvt_ggtt_gm_sz(gvt) (gvt->dev_priv->ggtt.vm.total) ggtt 385 drivers/gpu/drm/i915/gvt/gvt.h ((gvt->dev_priv->ggtt.vm.total >> PAGE_SHIFT) << 3) ggtt 397 drivers/gpu/drm/i915/gvt/gvt.h #define gvt_fence_sz(gvt) ((gvt)->dev_priv->ggtt.num_fences) ggtt 916 drivers/gpu/drm/i915/gvt/kvmgt.c aperture_va = io_mapping_map_wc(&vgpu->gvt->dev_priv->ggtt.iomap, ggtt 651 drivers/gpu/drm/i915/i915_debugfs.c seq_printf(m, "Total fences = %d\n", i915->ggtt.num_fences); ggtt 654 drivers/gpu/drm/i915/i915_debugfs.c for (i = 0; i < i915->ggtt.num_fences; i++) { ggtt 655 drivers/gpu/drm/i915/i915_debugfs.c struct i915_fence_reg *reg = &i915->ggtt.fence_regs[i]; ggtt 426 drivers/gpu/drm/i915/i915_drv.c struct i915_ggtt *ggtt = &dev_priv->ggtt; ggtt 434 drivers/gpu/drm/i915/i915_drv.c ap->ranges[0].base = ggtt->gmadr.start; ggtt 435 drivers/gpu/drm/i915/i915_drv.c ap->ranges[0].size = ggtt->mappable_end; ggtt 1446 drivers/gpu/drm/i915/i915_drv.h struct i915_ggtt ggtt; /* VM representing the global address space */ ggtt 65 drivers/gpu/drm/i915/i915_gem.c insert_mappable_node(struct i915_ggtt *ggtt, ggtt 69 drivers/gpu/drm/i915/i915_gem.c return drm_mm_insert_node_in_range(&ggtt->vm.mm, node, ggtt 71 drivers/gpu/drm/i915/i915_gem.c 0, ggtt->mappable_end, ggtt 85 drivers/gpu/drm/i915/i915_gem.c struct i915_ggtt *ggtt = &to_i915(dev)->ggtt; ggtt 90 drivers/gpu/drm/i915/i915_gem.c mutex_lock(&ggtt->vm.mutex); ggtt 92 drivers/gpu/drm/i915/i915_gem.c pinned = ggtt->vm.reserved; ggtt 93 drivers/gpu/drm/i915/i915_gem.c list_for_each_entry(vma, &ggtt->vm.bound_list, vm_link) ggtt 97 drivers/gpu/drm/i915/i915_gem.c mutex_unlock(&ggtt->vm.mutex); ggtt 99 drivers/gpu/drm/i915/i915_gem.c args->aper_size = ggtt->vm.total; ggtt 332 drivers/gpu/drm/i915/i915_gem.c struct i915_ggtt *ggtt = &i915->ggtt; ggtt 356 drivers/gpu/drm/i915/i915_gem.c ret = insert_mappable_node(ggtt, &node, PAGE_SIZE); ggtt 397 drivers/gpu/drm/i915/i915_gem.c ggtt->vm.insert_page(&ggtt->vm, ggtt 404 drivers/gpu/drm/i915/i915_gem.c if (gtt_user_read(&ggtt->iomap, page_base, page_offset, ggtt 419 drivers/gpu/drm/i915/i915_gem.c ggtt->vm.clear_range(&ggtt->vm, node.start, node.size); ggtt 524 drivers/gpu/drm/i915/i915_gem.c struct i915_ggtt *ggtt = &i915->ggtt; ggtt 566 drivers/gpu/drm/i915/i915_gem.c ret = insert_mappable_node(ggtt, &node, PAGE_SIZE); ggtt 609 drivers/gpu/drm/i915/i915_gem.c intel_gt_flush_ggtt_writes(ggtt->vm.gt); ggtt 610 drivers/gpu/drm/i915/i915_gem.c ggtt->vm.insert_page(&ggtt->vm, ggtt 623 drivers/gpu/drm/i915/i915_gem.c if (ggtt_write(&ggtt->iomap, page_base, page_offset, ggtt 638 drivers/gpu/drm/i915/i915_gem.c intel_gt_flush_ggtt_writes(ggtt->vm.gt); ggtt 640 drivers/gpu/drm/i915/i915_gem.c ggtt->vm.clear_range(&ggtt->vm, node.start, node.size); ggtt 859 drivers/gpu/drm/i915/i915_gem.c &i915->ggtt.userfault_list, userfault_link) ggtt 867 drivers/gpu/drm/i915/i915_gem.c for (i = 0; i < i915->ggtt.num_fences; i++) { ggtt 868 drivers/gpu/drm/i915/i915_gem.c struct i915_fence_reg *reg = &i915->ggtt.fence_regs[i]; ggtt 966 drivers/gpu/drm/i915/i915_gem.c struct i915_address_space *vm = &dev_priv->ggtt.vm; ggtt 998 drivers/gpu/drm/i915/i915_gem.c if (obj->base.size > dev_priv->ggtt.mappable_end) ggtt 1017 drivers/gpu/drm/i915/i915_gem.c obj->base.size > dev_priv->ggtt.mappable_end / 2) ggtt 1031 drivers/gpu/drm/i915/i915_gem.c vma->fence_size > dev_priv->ggtt.mappable_end / 2) ggtt 1609 drivers/gpu/drm/i915/i915_gem.c intel_wakeref_auto_fini(&dev_priv->ggtt.userfault_wakeref); ggtt 258 drivers/gpu/drm/i915/i915_gem_fence_reg.c list_move(&fence->link, &fence->i915->ggtt.fence_list); ggtt 282 drivers/gpu/drm/i915/i915_gem_fence_reg.c list_move_tail(&fence->link, &fence->i915->ggtt.fence_list); ggtt 318 drivers/gpu/drm/i915/i915_gem_fence_reg.c list_for_each_entry(fence, &i915->ggtt.fence_list, link) { ggtt 336 drivers/gpu/drm/i915/i915_gem_fence_reg.c struct i915_ggtt *ggtt = i915_vm_to_ggtt(vma->vm); ggtt 347 drivers/gpu/drm/i915/i915_gem_fence_reg.c list_move_tail(&fence->link, &ggtt->fence_list); ggtt 425 drivers/gpu/drm/i915/i915_gem_fence_reg.c struct i915_ggtt *ggtt = &i915->ggtt; ggtt 430 drivers/gpu/drm/i915/i915_gem_fence_reg.c lockdep_assert_held(&ggtt->vm.mutex); ggtt 434 drivers/gpu/drm/i915/i915_gem_fence_reg.c list_for_each_entry(fence, &ggtt->fence_list, link) ggtt 463 drivers/gpu/drm/i915/i915_gem_fence_reg.c struct i915_ggtt *ggtt = &fence->i915->ggtt; ggtt 465 drivers/gpu/drm/i915/i915_gem_fence_reg.c lockdep_assert_held(&ggtt->vm.mutex); ggtt 467 drivers/gpu/drm/i915/i915_gem_fence_reg.c list_add(&fence->link, &ggtt->fence_list); ggtt 483 drivers/gpu/drm/i915/i915_gem_fence_reg.c for (i = 0; i < i915->ggtt.num_fences; i++) { ggtt 484 drivers/gpu/drm/i915/i915_gem_fence_reg.c struct i915_fence_reg *reg = &i915->ggtt.fence_regs[i]; ggtt 819 drivers/gpu/drm/i915/i915_gem_fence_reg.c void i915_ggtt_init_fences(struct i915_ggtt *ggtt) ggtt 821 drivers/gpu/drm/i915/i915_gem_fence_reg.c struct drm_i915_private *i915 = ggtt->vm.i915; ggtt 825 drivers/gpu/drm/i915/i915_gem_fence_reg.c INIT_LIST_HEAD(&ggtt->fence_list); ggtt 826 drivers/gpu/drm/i915/i915_gem_fence_reg.c INIT_LIST_HEAD(&ggtt->userfault_list); ggtt 827 drivers/gpu/drm/i915/i915_gem_fence_reg.c intel_wakeref_auto_init(&ggtt->userfault_wakeref, &i915->runtime_pm); ggtt 847 drivers/gpu/drm/i915/i915_gem_fence_reg.c struct i915_fence_reg *fence = &ggtt->fence_regs[i]; ggtt 851 drivers/gpu/drm/i915/i915_gem_fence_reg.c list_add_tail(&fence->link, &ggtt->fence_list); ggtt 853 drivers/gpu/drm/i915/i915_gem_fence_reg.c ggtt->num_fences = num_fences; ggtt 68 drivers/gpu/drm/i915/i915_gem_fence_reg.h void i915_ggtt_init_fences(struct i915_ggtt *ggtt); ggtt 121 drivers/gpu/drm/i915/i915_gem_gtt.c static void gen6_ggtt_invalidate(struct i915_ggtt *ggtt) ggtt 123 drivers/gpu/drm/i915/i915_gem_gtt.c struct intel_uncore *uncore = ggtt->vm.gt->uncore; ggtt 132 drivers/gpu/drm/i915/i915_gem_gtt.c static void guc_ggtt_invalidate(struct i915_ggtt *ggtt) ggtt 134 drivers/gpu/drm/i915/i915_gem_gtt.c struct intel_uncore *uncore = ggtt->vm.gt->uncore; ggtt 136 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_ggtt_invalidate(ggtt); ggtt 140 drivers/gpu/drm/i915/i915_gem_gtt.c static void gmch_ggtt_invalidate(struct i915_ggtt *ggtt) ggtt 1736 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_ggtt_invalidate(vm->gt->ggtt); ggtt 1820 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_ggtt *ggtt = i915_vm_to_ggtt(vma->vm); ggtt 1827 drivers/gpu/drm/i915/i915_gem_gtt.c ppgtt->pd_addr = (gen6_pte_t __iomem *)ggtt->gsm + ggtt_offset; ggtt 1833 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_ggtt_invalidate(ggtt); ggtt 1872 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_ggtt *ggtt = ppgtt->base.vm.gt->ggtt; ggtt 1876 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(size > ggtt->vm.total); ggtt 1884 drivers/gpu/drm/i915/i915_gem_gtt.c vma->vm = &ggtt->vm; ggtt 1961 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_ggtt * const ggtt = &i915->ggtt; ggtt 1977 drivers/gpu/drm/i915/i915_gem_gtt.c ppgtt->base.vm.pte_encode = ggtt->vm.pte_encode; ggtt 2121 drivers/gpu/drm/i915/i915_gem_gtt.c static void ggtt_suspend_mappings(struct i915_ggtt *ggtt) ggtt 2123 drivers/gpu/drm/i915/i915_gem_gtt.c struct drm_i915_private *i915 = ggtt->vm.i915; ggtt 2131 drivers/gpu/drm/i915/i915_gem_gtt.c intel_gt_check_and_clear_faults(ggtt->vm.gt); ggtt 2133 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.clear_range(&ggtt->vm, 0, ggtt->vm.total); ggtt 2135 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->invalidate(ggtt); ggtt 2140 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt_suspend_mappings(&i915->ggtt); ggtt 2180 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm); ggtt 2182 drivers/gpu/drm/i915/i915_gem_gtt.c (gen8_pte_t __iomem *)ggtt->gsm + offset / I915_GTT_PAGE_SIZE; ggtt 2186 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->invalidate(ggtt); ggtt 2194 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm); ggtt 2205 drivers/gpu/drm/i915/i915_gem_gtt.c gtt_entries = (gen8_pte_t __iomem *)ggtt->gsm; ggtt 2214 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->invalidate(ggtt); ggtt 2223 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm); ggtt 2225 drivers/gpu/drm/i915/i915_gem_gtt.c (gen6_pte_t __iomem *)ggtt->gsm + offset / I915_GTT_PAGE_SIZE; ggtt 2229 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->invalidate(ggtt); ggtt 2243 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm); ggtt 2244 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_pte_t __iomem *entries = (gen6_pte_t __iomem *)ggtt->gsm; ggtt 2255 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->invalidate(ggtt); ggtt 2266 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm); ggtt 2271 drivers/gpu/drm/i915/i915_gem_gtt.c (gen8_pte_t __iomem *)ggtt->gsm + first_entry; ggtt 2272 drivers/gpu/drm/i915/i915_gem_gtt.c const int max_entries = ggtt_total_entries(ggtt) - first_entry; ggtt 2381 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm); ggtt 2385 drivers/gpu/drm/i915/i915_gem_gtt.c (gen6_pte_t __iomem *)ggtt->gsm + first_entry; ggtt 2386 drivers/gpu/drm/i915/i915_gem_gtt.c const int max_entries = ggtt_total_entries(ggtt) - first_entry; ggtt 2532 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_ggtt *ggtt = &dev_priv->ggtt; ggtt 2534 drivers/gpu/drm/i915/i915_gem_gtt.c if (unlikely(ggtt->do_idle_maps)) { ggtt 2578 drivers/gpu/drm/i915/i915_gem_gtt.c static int init_aliasing_ppgtt(struct i915_ggtt *ggtt) ggtt 2583 drivers/gpu/drm/i915/i915_gem_gtt.c ppgtt = i915_ppgtt_create(ggtt->vm.i915); ggtt 2587 drivers/gpu/drm/i915/i915_gem_gtt.c if (GEM_WARN_ON(ppgtt->vm.total < ggtt->vm.total)) { ggtt 2598 drivers/gpu/drm/i915/i915_gem_gtt.c err = ppgtt->vm.allocate_va_range(&ppgtt->vm, 0, ggtt->vm.total); ggtt 2602 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->alias = ppgtt; ggtt 2604 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(ggtt->vm.vma_ops.bind_vma != ggtt_bind_vma); ggtt 2605 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.bind_vma = aliasing_gtt_bind_vma; ggtt 2607 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(ggtt->vm.vma_ops.unbind_vma != ggtt_unbind_vma); ggtt 2608 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.unbind_vma = aliasing_gtt_unbind_vma; ggtt 2617 drivers/gpu/drm/i915/i915_gem_gtt.c static void fini_aliasing_ppgtt(struct i915_ggtt *ggtt) ggtt 2619 drivers/gpu/drm/i915/i915_gem_gtt.c struct drm_i915_private *i915 = ggtt->vm.i915; ggtt 2624 drivers/gpu/drm/i915/i915_gem_gtt.c ppgtt = fetch_and_zero(&ggtt->alias); ggtt 2630 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.bind_vma = ggtt_bind_vma; ggtt 2631 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.unbind_vma = ggtt_unbind_vma; ggtt 2637 drivers/gpu/drm/i915/i915_gem_gtt.c static int ggtt_reserve_guc_top(struct i915_ggtt *ggtt) ggtt 2642 drivers/gpu/drm/i915/i915_gem_gtt.c if (!USES_GUC(ggtt->vm.i915)) ggtt 2645 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(ggtt->vm.total <= GUC_GGTT_TOP); ggtt 2646 drivers/gpu/drm/i915/i915_gem_gtt.c size = ggtt->vm.total - GUC_GGTT_TOP; ggtt 2648 drivers/gpu/drm/i915/i915_gem_gtt.c ret = i915_gem_gtt_reserve(&ggtt->vm, &ggtt->uc_fw, size, ggtt 2657 drivers/gpu/drm/i915/i915_gem_gtt.c static void ggtt_release_guc_top(struct i915_ggtt *ggtt) ggtt 2659 drivers/gpu/drm/i915/i915_gem_gtt.c if (drm_mm_node_allocated(&ggtt->uc_fw)) ggtt 2660 drivers/gpu/drm/i915/i915_gem_gtt.c drm_mm_remove_node(&ggtt->uc_fw); ggtt 2663 drivers/gpu/drm/i915/i915_gem_gtt.c static void cleanup_init_ggtt(struct i915_ggtt *ggtt) ggtt 2665 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt_release_guc_top(ggtt); ggtt 2666 drivers/gpu/drm/i915/i915_gem_gtt.c drm_mm_remove_node(&ggtt->error_capture); ggtt 2669 drivers/gpu/drm/i915/i915_gem_gtt.c static int init_ggtt(struct i915_ggtt *ggtt) ggtt 2690 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->pin_bias = max_t(u32, I915_GTT_PAGE_SIZE, ggtt 2691 drivers/gpu/drm/i915/i915_gem_gtt.c intel_wopcm_guc_size(&ggtt->vm.i915->wopcm)); ggtt 2693 drivers/gpu/drm/i915/i915_gem_gtt.c ret = intel_vgt_balloon(ggtt); ggtt 2698 drivers/gpu/drm/i915/i915_gem_gtt.c ret = drm_mm_insert_node_in_range(&ggtt->vm.mm, &ggtt->error_capture, ggtt 2700 drivers/gpu/drm/i915/i915_gem_gtt.c 0, ggtt->mappable_end, ggtt 2710 drivers/gpu/drm/i915/i915_gem_gtt.c ret = ggtt_reserve_guc_top(ggtt); ggtt 2715 drivers/gpu/drm/i915/i915_gem_gtt.c drm_mm_for_each_hole(entry, &ggtt->vm.mm, hole_start, hole_end) { ggtt 2718 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.clear_range(&ggtt->vm, hole_start, ggtt 2723 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.clear_range(&ggtt->vm, ggtt->vm.total - PAGE_SIZE, PAGE_SIZE); ggtt 2728 drivers/gpu/drm/i915/i915_gem_gtt.c cleanup_init_ggtt(ggtt); ggtt 2736 drivers/gpu/drm/i915/i915_gem_gtt.c ret = init_ggtt(&i915->ggtt); ggtt 2741 drivers/gpu/drm/i915/i915_gem_gtt.c ret = init_aliasing_ppgtt(&i915->ggtt); ggtt 2743 drivers/gpu/drm/i915/i915_gem_gtt.c cleanup_init_ggtt(&i915->ggtt); ggtt 2749 drivers/gpu/drm/i915/i915_gem_gtt.c static void ggtt_cleanup_hw(struct i915_ggtt *ggtt) ggtt 2751 drivers/gpu/drm/i915/i915_gem_gtt.c struct drm_i915_private *i915 = ggtt->vm.i915; ggtt 2754 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.closed = true; ggtt 2761 drivers/gpu/drm/i915/i915_gem_gtt.c list_for_each_entry_safe(vma, vn, &ggtt->vm.bound_list, vm_link) ggtt 2764 drivers/gpu/drm/i915/i915_gem_gtt.c if (drm_mm_node_allocated(&ggtt->error_capture)) ggtt 2765 drivers/gpu/drm/i915/i915_gem_gtt.c drm_mm_remove_node(&ggtt->error_capture); ggtt 2767 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt_release_guc_top(ggtt); ggtt 2769 drivers/gpu/drm/i915/i915_gem_gtt.c if (drm_mm_initialized(&ggtt->vm.mm)) { ggtt 2770 drivers/gpu/drm/i915/i915_gem_gtt.c intel_vgt_deballoon(ggtt); ggtt 2771 drivers/gpu/drm/i915/i915_gem_gtt.c i915_address_space_fini(&ggtt->vm); ggtt 2774 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.cleanup(&ggtt->vm); ggtt 2778 drivers/gpu/drm/i915/i915_gem_gtt.c arch_phys_wc_del(ggtt->mtrr); ggtt 2779 drivers/gpu/drm/i915/i915_gem_gtt.c io_mapping_fini(&ggtt->iomap); ggtt 2790 drivers/gpu/drm/i915/i915_gem_gtt.c fini_aliasing_ppgtt(&i915->ggtt); ggtt 2792 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt_cleanup_hw(&i915->ggtt); ggtt 2837 drivers/gpu/drm/i915/i915_gem_gtt.c static int ggtt_probe_common(struct i915_ggtt *ggtt, u64 size) ggtt 2839 drivers/gpu/drm/i915/i915_gem_gtt.c struct drm_i915_private *dev_priv = ggtt->vm.i915; ggtt 2855 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->gsm = ioremap_nocache(phys_addr, size); ggtt 2857 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->gsm = ioremap_wc(phys_addr, size); ggtt 2858 drivers/gpu/drm/i915/i915_gem_gtt.c if (!ggtt->gsm) { ggtt 2863 drivers/gpu/drm/i915/i915_gem_gtt.c ret = setup_scratch_page(&ggtt->vm, GFP_DMA32); ggtt 2867 drivers/gpu/drm/i915/i915_gem_gtt.c iounmap(ggtt->gsm); ggtt 2871 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.scratch[0].encode = ggtt 2872 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.pte_encode(px_dma(&ggtt->vm.scratch[0]), ggtt 2961 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm); ggtt 2963 drivers/gpu/drm/i915/i915_gem_gtt.c iounmap(ggtt->gsm); ggtt 2981 drivers/gpu/drm/i915/i915_gem_gtt.c static int gen8_gmch_probe(struct i915_ggtt *ggtt) ggtt 2983 drivers/gpu/drm/i915/i915_gem_gtt.c struct drm_i915_private *dev_priv = ggtt->vm.i915; ggtt 2990 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->gmadr = ggtt 2993 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->mappable_end = resource_size(&ggtt->gmadr); ggtt 3007 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.total = (size / sizeof(gen8_pte_t)) * I915_GTT_PAGE_SIZE; ggtt 3008 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.cleanup = gen6_gmch_remove; ggtt 3009 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.insert_page = gen8_ggtt_insert_page; ggtt 3010 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.clear_range = nop_clear_range; ggtt 3012 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.clear_range = gen8_ggtt_clear_range; ggtt 3014 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.insert_entries = gen8_ggtt_insert_entries; ggtt 3019 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.insert_entries = bxt_vtd_ggtt_insert_entries__BKL; ggtt 3020 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.insert_page = bxt_vtd_ggtt_insert_page__BKL; ggtt 3021 drivers/gpu/drm/i915/i915_gem_gtt.c if (ggtt->vm.clear_range != nop_clear_range) ggtt 3022 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.clear_range = bxt_vtd_ggtt_clear_range__BKL; ggtt 3025 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->invalidate = gen6_ggtt_invalidate; ggtt 3027 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.bind_vma = ggtt_bind_vma; ggtt 3028 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.unbind_vma = ggtt_unbind_vma; ggtt 3029 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.set_pages = ggtt_set_pages; ggtt 3030 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.clear_pages = clear_pages; ggtt 3032 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.pte_encode = gen8_pte_encode; ggtt 3036 drivers/gpu/drm/i915/i915_gem_gtt.c return ggtt_probe_common(ggtt, size); ggtt 3039 drivers/gpu/drm/i915/i915_gem_gtt.c static int gen6_gmch_probe(struct i915_ggtt *ggtt) ggtt 3041 drivers/gpu/drm/i915/i915_gem_gtt.c struct drm_i915_private *dev_priv = ggtt->vm.i915; ggtt 3047 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->gmadr = ggtt 3050 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->mappable_end = resource_size(&ggtt->gmadr); ggtt 3055 drivers/gpu/drm/i915/i915_gem_gtt.c if (ggtt->mappable_end < (64<<20) || ggtt->mappable_end > (512<<20)) { ggtt 3056 drivers/gpu/drm/i915/i915_gem_gtt.c DRM_ERROR("Unknown GMADR size (%pa)\n", &ggtt->mappable_end); ggtt 3068 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.total = (size / sizeof(gen6_pte_t)) * I915_GTT_PAGE_SIZE; ggtt 3070 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.clear_range = nop_clear_range; ggtt 3072 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.clear_range = gen6_ggtt_clear_range; ggtt 3073 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.insert_page = gen6_ggtt_insert_page; ggtt 3074 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.insert_entries = gen6_ggtt_insert_entries; ggtt 3075 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.cleanup = gen6_gmch_remove; ggtt 3077 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->invalidate = gen6_ggtt_invalidate; ggtt 3080 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.pte_encode = iris_pte_encode; ggtt 3082 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.pte_encode = hsw_pte_encode; ggtt 3084 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.pte_encode = byt_pte_encode; ggtt 3086 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.pte_encode = ivb_pte_encode; ggtt 3088 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.pte_encode = snb_pte_encode; ggtt 3090 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.bind_vma = ggtt_bind_vma; ggtt 3091 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.unbind_vma = ggtt_unbind_vma; ggtt 3092 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.set_pages = ggtt_set_pages; ggtt 3093 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.clear_pages = clear_pages; ggtt 3095 drivers/gpu/drm/i915/i915_gem_gtt.c return ggtt_probe_common(ggtt, size); ggtt 3103 drivers/gpu/drm/i915/i915_gem_gtt.c static int i915_gmch_probe(struct i915_ggtt *ggtt) ggtt 3105 drivers/gpu/drm/i915/i915_gem_gtt.c struct drm_i915_private *dev_priv = ggtt->vm.i915; ggtt 3115 drivers/gpu/drm/i915/i915_gem_gtt.c intel_gtt_get(&ggtt->vm.total, &gmadr_base, &ggtt->mappable_end); ggtt 3117 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->gmadr = ggtt 3119 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->mappable_end); ggtt 3121 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->do_idle_maps = needs_idle_maps(dev_priv); ggtt 3122 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.insert_page = i915_ggtt_insert_page; ggtt 3123 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.insert_entries = i915_ggtt_insert_entries; ggtt 3124 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.clear_range = i915_ggtt_clear_range; ggtt 3125 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.cleanup = i915_gmch_remove; ggtt 3127 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->invalidate = gmch_ggtt_invalidate; ggtt 3129 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.bind_vma = ggtt_bind_vma; ggtt 3130 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.unbind_vma = ggtt_unbind_vma; ggtt 3131 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.set_pages = ggtt_set_pages; ggtt 3132 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.vma_ops.clear_pages = clear_pages; ggtt 3134 drivers/gpu/drm/i915/i915_gem_gtt.c if (unlikely(ggtt->do_idle_maps)) ggtt 3141 drivers/gpu/drm/i915/i915_gem_gtt.c static int ggtt_probe_hw(struct i915_ggtt *ggtt, struct intel_gt *gt) ggtt 3146 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.gt = gt; ggtt 3147 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.i915 = i915; ggtt 3148 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.dma = &i915->drm.pdev->dev; ggtt 3151 drivers/gpu/drm/i915/i915_gem_gtt.c ret = i915_gmch_probe(ggtt); ggtt 3153 drivers/gpu/drm/i915/i915_gem_gtt.c ret = gen6_gmch_probe(ggtt); ggtt 3155 drivers/gpu/drm/i915/i915_gem_gtt.c ret = gen8_gmch_probe(ggtt); ggtt 3159 drivers/gpu/drm/i915/i915_gem_gtt.c if ((ggtt->vm.total - 1) >> 32) { ggtt 3162 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.total >> 20); ggtt 3163 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.total = 1ULL << 32; ggtt 3164 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->mappable_end = ggtt 3165 drivers/gpu/drm/i915/i915_gem_gtt.c min_t(u64, ggtt->mappable_end, ggtt->vm.total); ggtt 3168 drivers/gpu/drm/i915/i915_gem_gtt.c if (ggtt->mappable_end > ggtt->vm.total) { ggtt 3171 drivers/gpu/drm/i915/i915_gem_gtt.c &ggtt->mappable_end, ggtt->vm.total); ggtt 3172 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->mappable_end = ggtt->vm.total; ggtt 3176 drivers/gpu/drm/i915/i915_gem_gtt.c DRM_DEBUG_DRIVER("GGTT size = %lluM\n", ggtt->vm.total >> 20); ggtt 3177 drivers/gpu/drm/i915/i915_gem_gtt.c DRM_DEBUG_DRIVER("GMADR size = %lluM\n", (u64)ggtt->mappable_end >> 20); ggtt 3192 drivers/gpu/drm/i915/i915_gem_gtt.c ret = ggtt_probe_hw(&i915->ggtt, &i915->gt); ggtt 3202 drivers/gpu/drm/i915/i915_gem_gtt.c static int ggtt_init_hw(struct i915_ggtt *ggtt) ggtt 3204 drivers/gpu/drm/i915/i915_gem_gtt.c struct drm_i915_private *i915 = ggtt->vm.i915; ggtt 3209 drivers/gpu/drm/i915/i915_gem_gtt.c i915_address_space_init(&ggtt->vm, VM_CLASS_GGTT); ggtt 3211 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.is_ggtt = true; ggtt 3214 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.has_read_only = IS_VALLEYVIEW(i915); ggtt 3217 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.mm.color_adjust = i915_gtt_color_adjust; ggtt 3219 drivers/gpu/drm/i915/i915_gem_gtt.c if (!io_mapping_init_wc(&ggtt->iomap, ggtt 3220 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->gmadr.start, ggtt 3221 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->mappable_end)) { ggtt 3222 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.cleanup(&ggtt->vm); ggtt 3227 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->mtrr = arch_phys_wc_add(ggtt->gmadr.start, ggtt->mappable_end); ggtt 3229 drivers/gpu/drm/i915/i915_gem_gtt.c i915_ggtt_init_fences(ggtt); ggtt 3252 drivers/gpu/drm/i915/i915_gem_gtt.c ret = ggtt_init_hw(&dev_priv->ggtt); ggtt 3267 drivers/gpu/drm/i915/i915_gem_gtt.c dev_priv->ggtt.vm.cleanup(&dev_priv->ggtt.vm); ggtt 3279 drivers/gpu/drm/i915/i915_gem_gtt.c void i915_ggtt_enable_guc(struct i915_ggtt *ggtt) ggtt 3281 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(ggtt->invalidate != gen6_ggtt_invalidate); ggtt 3283 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->invalidate = guc_ggtt_invalidate; ggtt 3285 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->invalidate(ggtt); ggtt 3288 drivers/gpu/drm/i915/i915_gem_gtt.c void i915_ggtt_disable_guc(struct i915_ggtt *ggtt) ggtt 3291 drivers/gpu/drm/i915/i915_gem_gtt.c if (ggtt->invalidate == gen6_ggtt_invalidate) ggtt 3295 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(ggtt->invalidate != guc_ggtt_invalidate); ggtt 3297 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->invalidate = gen6_ggtt_invalidate; ggtt 3299 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->invalidate(ggtt); ggtt 3302 drivers/gpu/drm/i915/i915_gem_gtt.c static void ggtt_restore_mappings(struct i915_ggtt *ggtt) ggtt 3307 drivers/gpu/drm/i915/i915_gem_gtt.c intel_gt_check_and_clear_faults(ggtt->vm.gt); ggtt 3309 drivers/gpu/drm/i915/i915_gem_gtt.c mutex_lock(&ggtt->vm.mutex); ggtt 3312 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.clear_range(&ggtt->vm, 0, ggtt->vm.total); ggtt 3313 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.closed = true; /* skip rewriting PTE on VMA unbind */ ggtt 3316 drivers/gpu/drm/i915/i915_gem_gtt.c list_for_each_entry_safe(vma, vn, &ggtt->vm.bound_list, vm_link) { ggtt 3322 drivers/gpu/drm/i915/i915_gem_gtt.c mutex_unlock(&ggtt->vm.mutex); ggtt 3336 drivers/gpu/drm/i915/i915_gem_gtt.c mutex_lock(&ggtt->vm.mutex); ggtt 3339 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->vm.closed = false; ggtt 3340 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt->invalidate(ggtt); ggtt 3342 drivers/gpu/drm/i915/i915_gem_gtt.c mutex_unlock(&ggtt->vm.mutex); ggtt 3350 drivers/gpu/drm/i915/i915_gem_gtt.c ggtt_restore_mappings(&i915->ggtt); ggtt 3642 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(vm == &vm->i915->ggtt.alias->vm); ggtt 3739 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(vm == &vm->i915->ggtt.alias->vm); ggtt 76 drivers/gpu/drm/i915/i915_gem_gtt.h #define ggtt_total_entries(ggtt) ((ggtt)->vm.total >> PAGE_SHIFT) ggtt 395 drivers/gpu/drm/i915/i915_gem_gtt.h void (*invalidate)(struct i915_ggtt *ggtt); ggtt 557 drivers/gpu/drm/i915/i915_gem_gtt.h void i915_ggtt_enable_guc(struct i915_ggtt *ggtt); ggtt 558 drivers/gpu/drm/i915/i915_gem_gtt.h void i915_ggtt_disable_guc(struct i915_ggtt *ggtt); ggtt 31 drivers/gpu/drm/i915/i915_getparam.c value = i915->ggtt.num_fences; ggtt 961 drivers/gpu/drm/i915/i915_gpu_error.c struct i915_ggtt *ggtt = &i915->ggtt; ggtt 962 drivers/gpu/drm/i915/i915_gpu_error.c const u64 slot = ggtt->error_capture.start; ggtt 995 drivers/gpu/drm/i915/i915_gpu_error.c ggtt->vm.insert_page(&ggtt->vm, dma, slot, I915_CACHE_NONE, 0); ggtt 997 drivers/gpu/drm/i915/i915_gpu_error.c s = io_mapping_map_wc(&ggtt->iomap, slot, PAGE_SIZE); ggtt 1045 drivers/gpu/drm/i915/i915_gpu_error.c for (i = 0; i < dev_priv->ggtt.num_fences; i++) ggtt 1050 drivers/gpu/drm/i915/i915_gpu_error.c for (i = 0; i < dev_priv->ggtt.num_fences; i++) ggtt 1055 drivers/gpu/drm/i915/i915_gpu_error.c for (i = 0; i < dev_priv->ggtt.num_fences; i++) ggtt 1666 drivers/gpu/drm/i915/i915_gpu_error.c struct i915_ggtt *ggtt = &error->i915->ggtt; ggtt 1667 drivers/gpu/drm/i915/i915_gpu_error.c const u64 slot = ggtt->error_capture.start; ggtt 1669 drivers/gpu/drm/i915/i915_gpu_error.c ggtt->vm.clear_range(&ggtt->vm, slot, PAGE_SIZE); ggtt 120 drivers/gpu/drm/i915/i915_vgpu.c static void vgt_deballoon_space(struct i915_ggtt *ggtt, ggtt 131 drivers/gpu/drm/i915/i915_vgpu.c ggtt->vm.reserved -= node->size; ggtt 142 drivers/gpu/drm/i915/i915_vgpu.c void intel_vgt_deballoon(struct i915_ggtt *ggtt) ggtt 146 drivers/gpu/drm/i915/i915_vgpu.c if (!intel_vgpu_active(ggtt->vm.i915)) ggtt 152 drivers/gpu/drm/i915/i915_vgpu.c vgt_deballoon_space(ggtt, &bl_info.space[i]); ggtt 155 drivers/gpu/drm/i915/i915_vgpu.c static int vgt_balloon_space(struct i915_ggtt *ggtt, ggtt 167 drivers/gpu/drm/i915/i915_vgpu.c ret = i915_gem_gtt_reserve(&ggtt->vm, node, ggtt 171 drivers/gpu/drm/i915/i915_vgpu.c ggtt->vm.reserved += size; ggtt 220 drivers/gpu/drm/i915/i915_vgpu.c int intel_vgt_balloon(struct i915_ggtt *ggtt) ggtt 222 drivers/gpu/drm/i915/i915_vgpu.c struct intel_uncore *uncore = &ggtt->vm.i915->uncore; ggtt 223 drivers/gpu/drm/i915/i915_vgpu.c unsigned long ggtt_end = ggtt->vm.total; ggtt 229 drivers/gpu/drm/i915/i915_vgpu.c if (!intel_vgpu_active(ggtt->vm.i915)) ggtt 250 drivers/gpu/drm/i915/i915_vgpu.c if (mappable_end > ggtt->mappable_end || ggtt 251 drivers/gpu/drm/i915/i915_vgpu.c unmappable_base < ggtt->mappable_end || ggtt 258 drivers/gpu/drm/i915/i915_vgpu.c if (unmappable_base > ggtt->mappable_end) { ggtt 259 drivers/gpu/drm/i915/i915_vgpu.c ret = vgt_balloon_space(ggtt, &bl_info.space[2], ggtt 260 drivers/gpu/drm/i915/i915_vgpu.c ggtt->mappable_end, unmappable_base); ggtt 267 drivers/gpu/drm/i915/i915_vgpu.c ret = vgt_balloon_space(ggtt, &bl_info.space[3], ggtt 275 drivers/gpu/drm/i915/i915_vgpu.c ret = vgt_balloon_space(ggtt, &bl_info.space[0], ggtt 282 drivers/gpu/drm/i915/i915_vgpu.c if (mappable_end < ggtt->mappable_end) { ggtt 283 drivers/gpu/drm/i915/i915_vgpu.c ret = vgt_balloon_space(ggtt, &bl_info.space[1], ggtt 284 drivers/gpu/drm/i915/i915_vgpu.c mappable_end, ggtt->mappable_end); ggtt 294 drivers/gpu/drm/i915/i915_vgpu.c vgt_deballoon_space(ggtt, &bl_info.space[0]); ggtt 296 drivers/gpu/drm/i915/i915_vgpu.c vgt_deballoon_space(ggtt, &bl_info.space[3]); ggtt 298 drivers/gpu/drm/i915/i915_vgpu.c vgt_deballoon_space(ggtt, &bl_info.space[2]); ggtt 46 drivers/gpu/drm/i915/i915_vgpu.h int intel_vgt_balloon(struct i915_ggtt *ggtt); ggtt 47 drivers/gpu/drm/i915/i915_vgpu.h void intel_vgt_deballoon(struct i915_ggtt *ggtt); ggtt 107 drivers/gpu/drm/i915/i915_vma.c GEM_BUG_ON(vm == &vm->i915->ggtt.alias->vm); ggtt 570 drivers/gpu/drm/i915/i915_vma.c end = min_t(u64, end, dev_priv->ggtt.mappable_end); ggtt 39 drivers/gpu/drm/i915/selftests/i915_gem.c struct i915_ggtt *ggtt = &i915->ggtt; ggtt 40 drivers/gpu/drm/i915/selftests/i915_gem.c const u64 slot = ggtt->error_capture.start; ggtt 50 drivers/gpu/drm/i915/selftests/i915_gem.c ggtt->vm.insert_page(&ggtt->vm, dma, slot, I915_CACHE_NONE, 0); ggtt 52 drivers/gpu/drm/i915/selftests/i915_gem.c s = io_mapping_map_atomic_wc(&ggtt->iomap, slot); ggtt 60 drivers/gpu/drm/i915/selftests/i915_gem.c ggtt->vm.clear_range(&ggtt->vm, slot, PAGE_SIZE); ggtt 73 drivers/gpu/drm/i915/selftests/i915_gem_evict.c count, i915->ggtt.vm.total / PAGE_SIZE); ggtt 99 drivers/gpu/drm/i915/selftests/i915_gem_evict.c if (list_empty(&i915->ggtt.vm.bound_list)) { ggtt 109 drivers/gpu/drm/i915/selftests/i915_gem_evict.c struct i915_ggtt *ggtt = &i915->ggtt; ggtt 112 drivers/gpu/drm/i915/selftests/i915_gem_evict.c mutex_lock(&ggtt->vm.mutex); ggtt 113 drivers/gpu/drm/i915/selftests/i915_gem_evict.c list_for_each_entry(vma, &i915->ggtt.vm.bound_list, vm_link) ggtt 116 drivers/gpu/drm/i915/selftests/i915_gem_evict.c mutex_unlock(&ggtt->vm.mutex); ggtt 140 drivers/gpu/drm/i915/selftests/i915_gem_evict.c struct i915_ggtt *ggtt = &i915->ggtt; ggtt 151 drivers/gpu/drm/i915/selftests/i915_gem_evict.c err = i915_gem_evict_something(&ggtt->vm, ggtt 164 drivers/gpu/drm/i915/selftests/i915_gem_evict.c err = i915_gem_evict_something(&ggtt->vm, ggtt 218 drivers/gpu/drm/i915/selftests/i915_gem_evict.c struct i915_ggtt *ggtt = &i915->ggtt; ggtt 233 drivers/gpu/drm/i915/selftests/i915_gem_evict.c err = i915_gem_evict_for_node(&ggtt->vm, &target, 0); ggtt 243 drivers/gpu/drm/i915/selftests/i915_gem_evict.c err = i915_gem_evict_for_node(&ggtt->vm, &target, 0); ggtt 265 drivers/gpu/drm/i915/selftests/i915_gem_evict.c struct i915_ggtt *ggtt = &i915->ggtt; ggtt 282 drivers/gpu/drm/i915/selftests/i915_gem_evict.c ggtt->vm.mm.color_adjust = mock_color_adjust; ggtt 320 drivers/gpu/drm/i915/selftests/i915_gem_evict.c err = i915_gem_evict_for_node(&ggtt->vm, &target, 0); ggtt 331 drivers/gpu/drm/i915/selftests/i915_gem_evict.c err = i915_gem_evict_for_node(&ggtt->vm, &target, 0); ggtt 343 drivers/gpu/drm/i915/selftests/i915_gem_evict.c ggtt->vm.mm.color_adjust = NULL; ggtt 350 drivers/gpu/drm/i915/selftests/i915_gem_evict.c struct i915_ggtt *ggtt = &i915->ggtt; ggtt 361 drivers/gpu/drm/i915/selftests/i915_gem_evict.c err = i915_gem_evict_vm(&ggtt->vm); ggtt 370 drivers/gpu/drm/i915/selftests/i915_gem_evict.c err = i915_gem_evict_vm(&ggtt->vm); ggtt 416 drivers/gpu/drm/i915/selftests/i915_gem_evict.c err = i915_gem_gtt_insert(&i915->ggtt.vm, &hole, ggtt 418 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 0, i915->ggtt.vm.total, ggtt 434 drivers/gpu/drm/i915/selftests/i915_gem_evict.c if (i915_gem_gtt_insert(&i915->ggtt.vm, &r->node, ggtt 436 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 0, i915->ggtt.vm.total, ggtt 1082 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c struct i915_ggtt *ggtt = &i915->ggtt; ggtt 1090 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c list_sort(NULL, &ggtt->vm.mm.hole_stack, sort_holes); ggtt 1091 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c drm_mm_for_each_hole(node, &ggtt->vm.mm, hole_start, hole_end) { ggtt 1095 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (ggtt->vm.mm.color_adjust) ggtt 1096 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c ggtt->vm.mm.color_adjust(node, 0, ggtt 1101 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = func(i915, &ggtt->vm, hole_start, hole_end, end_time); ggtt 1144 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c struct i915_ggtt *ggtt = &i915->ggtt; ggtt 1164 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = drm_mm_insert_node_in_range(&ggtt->vm.mm, &tmp, ggtt 1167 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 0, ggtt->mappable_end, ggtt 1177 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c ggtt->vm.insert_page(&ggtt->vm, ggtt 1192 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c vaddr = io_mapping_map_atomic_wc(&ggtt->iomap, offset); ggtt 1196 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c intel_gt_flush_ggtt_writes(ggtt->vm.gt); ggtt 1204 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c vaddr = io_mapping_map_atomic_wc(&ggtt->iomap, offset); ggtt 1218 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c ggtt->vm.clear_range(&ggtt->vm, tmp.start, tmp.size); ggtt 1267 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c struct i915_ggtt *ggtt = arg; ggtt 1269 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c return exercise_mock(ggtt->vm.i915, fill_hole); ggtt 1274 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c struct i915_ggtt *ggtt = arg; ggtt 1276 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c return exercise_mock(ggtt->vm.i915, walk_hole); ggtt 1281 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c struct i915_ggtt *ggtt = arg; ggtt 1283 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c return exercise_mock(ggtt->vm.i915, pot_hole); ggtt 1288 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c struct i915_ggtt *ggtt = arg; ggtt 1290 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c return exercise_mock(ggtt->vm.i915, drunk_hole); ggtt 1295 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c struct i915_ggtt *ggtt = arg; ggtt 1308 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c total + 2 * I915_GTT_PAGE_SIZE <= ggtt->vm.total; ggtt 1312 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c obj = i915_gem_object_create_internal(ggtt->vm.i915, ggtt 1327 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c vma = i915_vma_instance(obj, &ggtt->vm, NULL); ggtt 1333 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = i915_gem_gtt_reserve(&ggtt->vm, &vma->node, ggtt 1340 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c total, ggtt->vm.total, err); ggtt 1358 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c total + 2 * I915_GTT_PAGE_SIZE <= ggtt->vm.total; ggtt 1362 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c obj = i915_gem_object_create_internal(ggtt->vm.i915, ggtt 1377 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c vma = i915_vma_instance(obj, &ggtt->vm, NULL); ggtt 1383 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = i915_gem_gtt_reserve(&ggtt->vm, &vma->node, ggtt 1390 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c total, ggtt->vm.total, err); ggtt 1411 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c vma = i915_vma_instance(obj, &ggtt->vm, NULL); ggtt 1423 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c offset = random_offset(0, ggtt->vm.total, ggtt 1427 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = i915_gem_gtt_reserve(&ggtt->vm, &vma->node, ggtt 1434 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c total, ggtt->vm.total, err); ggtt 1460 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c struct i915_ggtt *ggtt = arg; ggtt 1469 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c ggtt->vm.total + I915_GTT_PAGE_SIZE, 0, ggtt 1470 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 0, ggtt->vm.total, ggtt 1500 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = i915_gem_gtt_insert(&ggtt->vm, &tmp, ggtt 1515 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c total + I915_GTT_PAGE_SIZE <= ggtt->vm.total; ggtt 1519 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c obj = i915_gem_object_create_internal(ggtt->vm.i915, ggtt 1534 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c vma = i915_vma_instance(obj, &ggtt->vm, NULL); ggtt 1540 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = i915_gem_gtt_insert(&ggtt->vm, &vma->node, ggtt 1542 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 0, ggtt->vm.total, ggtt 1551 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c total, ggtt->vm.total, err); ggtt 1563 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c vma = i915_vma_instance(obj, &ggtt->vm, NULL); ggtt 1583 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c vma = i915_vma_instance(obj, &ggtt->vm, NULL); ggtt 1598 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = i915_gem_gtt_insert(&ggtt->vm, &vma->node, ggtt 1600 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 0, ggtt->vm.total, ggtt 1604 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c total, ggtt->vm.total, err); ggtt 1620 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c total + 2 * I915_GTT_PAGE_SIZE <= ggtt->vm.total; ggtt 1624 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c obj = i915_gem_object_create_internal(ggtt->vm.i915, ggtt 1639 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c vma = i915_vma_instance(obj, &ggtt->vm, NULL); ggtt 1645 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = i915_gem_gtt_insert(&ggtt->vm, &vma->node, ggtt 1647 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 0, ggtt->vm.total, ggtt 1651 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c total, ggtt->vm.total, err); ggtt 1678 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c struct i915_ggtt *ggtt; ggtt 1685 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c ggtt = kmalloc(sizeof(*ggtt), GFP_KERNEL); ggtt 1686 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (!ggtt) { ggtt 1690 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c mock_init_ggtt(i915, ggtt); ggtt 1693 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c err = i915_subtests(tests, ggtt); ggtt 1699 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c mock_fini_ggtt(ggtt); ggtt 1700 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c kfree(ggtt); ggtt 1725 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c GEM_BUG_ON(offset_in_page(i915->ggtt.vm.total)); ggtt 640 drivers/gpu/drm/i915/selftests/i915_request.c vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL); ggtt 764 drivers/gpu/drm/i915/selftests/i915_request.c struct i915_address_space *vm = ctx->vm ?: &i915->ggtt.vm; ggtt 147 drivers/gpu/drm/i915/selftests/i915_vma.c struct i915_ggtt *ggtt = arg; ggtt 148 drivers/gpu/drm/i915/selftests/i915_vma.c struct drm_i915_private *i915 = ggtt->vm.i915; ggtt 254 drivers/gpu/drm/i915/selftests/i915_vma.c struct i915_ggtt *ggtt = arg; ggtt 265 drivers/gpu/drm/i915/selftests/i915_vma.c VALID(0, PIN_GLOBAL | PIN_OFFSET_BIAS | (ggtt->mappable_end - 4096)), ggtt 266 drivers/gpu/drm/i915/selftests/i915_vma.c VALID(0, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_BIAS | (ggtt->mappable_end - 4096)), ggtt 267 drivers/gpu/drm/i915/selftests/i915_vma.c VALID(0, PIN_GLOBAL | PIN_OFFSET_BIAS | (ggtt->vm.total - 4096)), ggtt 269 drivers/gpu/drm/i915/selftests/i915_vma.c VALID(0, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_FIXED | (ggtt->mappable_end - 4096)), ggtt 270 drivers/gpu/drm/i915/selftests/i915_vma.c INVALID(0, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_FIXED | ggtt->mappable_end), ggtt 271 drivers/gpu/drm/i915/selftests/i915_vma.c VALID(0, PIN_GLOBAL | PIN_OFFSET_FIXED | (ggtt->vm.total - 4096)), ggtt 272 drivers/gpu/drm/i915/selftests/i915_vma.c INVALID(0, PIN_GLOBAL | PIN_OFFSET_FIXED | ggtt->vm.total), ggtt 277 drivers/gpu/drm/i915/selftests/i915_vma.c VALID(ggtt->mappable_end - 4096, PIN_GLOBAL | PIN_MAPPABLE), ggtt 278 drivers/gpu/drm/i915/selftests/i915_vma.c VALID(ggtt->mappable_end, PIN_GLOBAL | PIN_MAPPABLE), ggtt 279 drivers/gpu/drm/i915/selftests/i915_vma.c NOSPACE(ggtt->mappable_end + 4096, PIN_GLOBAL | PIN_MAPPABLE), ggtt 280 drivers/gpu/drm/i915/selftests/i915_vma.c VALID(ggtt->vm.total - 4096, PIN_GLOBAL), ggtt 281 drivers/gpu/drm/i915/selftests/i915_vma.c VALID(ggtt->vm.total, PIN_GLOBAL), ggtt 282 drivers/gpu/drm/i915/selftests/i915_vma.c NOSPACE(ggtt->vm.total + 4096, PIN_GLOBAL), ggtt 284 drivers/gpu/drm/i915/selftests/i915_vma.c INVALID(8192, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_FIXED | (ggtt->mappable_end - 4096)), ggtt 285 drivers/gpu/drm/i915/selftests/i915_vma.c INVALID(8192, PIN_GLOBAL | PIN_OFFSET_FIXED | (ggtt->vm.total - 4096)), ggtt 288 drivers/gpu/drm/i915/selftests/i915_vma.c VALID(8192, PIN_GLOBAL | PIN_OFFSET_BIAS | (ggtt->mappable_end - 4096)), ggtt 296 drivers/gpu/drm/i915/selftests/i915_vma.c NOSPACE(0, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_BIAS | ggtt->mappable_end), ggtt 297 drivers/gpu/drm/i915/selftests/i915_vma.c NOSPACE(0, PIN_GLOBAL | PIN_OFFSET_BIAS | ggtt->vm.total), ggtt 298 drivers/gpu/drm/i915/selftests/i915_vma.c NOSPACE(8192, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_BIAS | (ggtt->mappable_end - 4096)), ggtt 299 drivers/gpu/drm/i915/selftests/i915_vma.c NOSPACE(8192, PIN_GLOBAL | PIN_OFFSET_BIAS | (ggtt->vm.total - 4096)), ggtt 315 drivers/gpu/drm/i915/selftests/i915_vma.c GEM_BUG_ON(!drm_mm_clean(&ggtt->vm.mm)); ggtt 317 drivers/gpu/drm/i915/selftests/i915_vma.c obj = i915_gem_object_create_internal(ggtt->vm.i915, PAGE_SIZE); ggtt 321 drivers/gpu/drm/i915/selftests/i915_vma.c vma = checked_vma_instance(obj, &ggtt->vm, NULL); ggtt 475 drivers/gpu/drm/i915/selftests/i915_vma.c struct i915_ggtt *ggtt = arg; ggtt 476 drivers/gpu/drm/i915/selftests/i915_vma.c struct i915_address_space *vm = &ggtt->vm; ggtt 697 drivers/gpu/drm/i915/selftests/i915_vma.c struct i915_ggtt *ggtt = arg; ggtt 698 drivers/gpu/drm/i915/selftests/i915_vma.c struct i915_address_space *vm = &ggtt->vm; ggtt 820 drivers/gpu/drm/i915/selftests/i915_vma.c struct i915_ggtt *ggtt; ggtt 827 drivers/gpu/drm/i915/selftests/i915_vma.c ggtt = kmalloc(sizeof(*ggtt), GFP_KERNEL); ggtt 828 drivers/gpu/drm/i915/selftests/i915_vma.c if (!ggtt) { ggtt 832 drivers/gpu/drm/i915/selftests/i915_vma.c mock_init_ggtt(i915, ggtt); ggtt 835 drivers/gpu/drm/i915/selftests/i915_vma.c err = i915_subtests(tests, ggtt); ggtt 841 drivers/gpu/drm/i915/selftests/i915_vma.c mock_fini_ggtt(ggtt); ggtt 842 drivers/gpu/drm/i915/selftests/i915_vma.c kfree(ggtt); ggtt 77 drivers/gpu/drm/i915/selftests/mock_gem_device.c mock_fini_ggtt(&i915->ggtt); ggtt 201 drivers/gpu/drm/i915/selftests/mock_gem_device.c mock_init_ggtt(i915, &i915->ggtt); ggtt 97 drivers/gpu/drm/i915/selftests/mock_gtt.c void mock_init_ggtt(struct drm_i915_private *i915, struct i915_ggtt *ggtt) ggtt 99 drivers/gpu/drm/i915/selftests/mock_gtt.c memset(ggtt, 0, sizeof(*ggtt)); ggtt 101 drivers/gpu/drm/i915/selftests/mock_gtt.c ggtt->vm.gt = &i915->gt; ggtt 102 drivers/gpu/drm/i915/selftests/mock_gtt.c ggtt->vm.i915 = i915; ggtt 103 drivers/gpu/drm/i915/selftests/mock_gtt.c ggtt->vm.is_ggtt = true; ggtt 105 drivers/gpu/drm/i915/selftests/mock_gtt.c ggtt->gmadr = (struct resource) DEFINE_RES_MEM(0, 2048 * PAGE_SIZE); ggtt 106 drivers/gpu/drm/i915/selftests/mock_gtt.c ggtt->mappable_end = resource_size(&ggtt->gmadr); ggtt 107 drivers/gpu/drm/i915/selftests/mock_gtt.c ggtt->vm.total = 4096 * PAGE_SIZE; ggtt 109 drivers/gpu/drm/i915/selftests/mock_gtt.c ggtt->vm.clear_range = nop_clear_range; ggtt 110 drivers/gpu/drm/i915/selftests/mock_gtt.c ggtt->vm.insert_page = mock_insert_page; ggtt 111 drivers/gpu/drm/i915/selftests/mock_gtt.c ggtt->vm.insert_entries = mock_insert_entries; ggtt 112 drivers/gpu/drm/i915/selftests/mock_gtt.c ggtt->vm.cleanup = mock_cleanup; ggtt 114 drivers/gpu/drm/i915/selftests/mock_gtt.c ggtt->vm.vma_ops.bind_vma = mock_bind_ggtt; ggtt 115 drivers/gpu/drm/i915/selftests/mock_gtt.c ggtt->vm.vma_ops.unbind_vma = mock_unbind_ggtt; ggtt 116 drivers/gpu/drm/i915/selftests/mock_gtt.c ggtt->vm.vma_ops.set_pages = ggtt_set_pages; ggtt 117 drivers/gpu/drm/i915/selftests/mock_gtt.c ggtt->vm.vma_ops.clear_pages = clear_pages; ggtt 119 drivers/gpu/drm/i915/selftests/mock_gtt.c i915_address_space_init(&ggtt->vm, VM_CLASS_GGTT); ggtt 124 drivers/gpu/drm/i915/selftests/mock_gtt.c void mock_fini_ggtt(struct i915_ggtt *ggtt) ggtt 126 drivers/gpu/drm/i915/selftests/mock_gtt.c i915_address_space_fini(&ggtt->vm); ggtt 28 drivers/gpu/drm/i915/selftests/mock_gtt.h void mock_init_ggtt(struct drm_i915_private *i915, struct i915_ggtt *ggtt); ggtt 29 drivers/gpu/drm/i915/selftests/mock_gtt.h void mock_fini_ggtt(struct i915_ggtt *ggtt);