ggtt              172 drivers/gpu/drm/i915/display/intel_fbdev.c 	struct i915_ggtt *ggtt = &dev_priv->ggtt;
ggtt              235 drivers/gpu/drm/i915/display/intel_fbdev.c 	info->apertures->ranges[0].base = ggtt->gmadr.start;
ggtt              236 drivers/gpu/drm/i915/display/intel_fbdev.c 	info->apertures->ranges[0].size = ggtt->mappable_end;
ggtt              240 drivers/gpu/drm/i915/display/intel_fbdev.c 		(unsigned long)(ggtt->gmadr.start + vma->node.start);
ggtt             2235 drivers/gpu/drm/i915/gem/i915_gem_context.c 		else if (to_i915(dev)->ggtt.alias)
ggtt             2236 drivers/gpu/drm/i915/gem/i915_gem_context.c 			args->value = to_i915(dev)->ggtt.alias->vm.total;
ggtt             2238 drivers/gpu/drm/i915/gem/i915_gem_context.c 			args->value = to_i915(dev)->ggtt.vm.total;
ggtt              249 drivers/gpu/drm/i915/gem/i915_gem_domain.c 			ret = mutex_lock_interruptible(&i915->ggtt.vm.mutex);
ggtt              272 drivers/gpu/drm/i915/gem/i915_gem_domain.c 			mutex_unlock(&i915->ggtt.vm.mutex);
ggtt              489 drivers/gpu/drm/i915/gem/i915_gem_domain.c 	mutex_lock(&i915->ggtt.vm.mutex);
ggtt              496 drivers/gpu/drm/i915/gem/i915_gem_domain.c 	mutex_unlock(&i915->ggtt.vm.mutex);
ggtt              929 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	return &i915->ggtt;
ggtt              966 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		struct i915_ggtt *ggtt = cache_to_ggtt(cache);
ggtt              968 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		intel_gt_flush_ggtt_writes(ggtt->vm.gt);
ggtt              972 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 			ggtt->vm.clear_range(&ggtt->vm,
ggtt             1021 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	struct i915_ggtt *ggtt = cache_to_ggtt(cache);
ggtt             1026 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		intel_gt_flush_ggtt_writes(ggtt->vm.gt);
ggtt             1051 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 				(&ggtt->vm.mm, &cache->node,
ggtt             1053 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 				 0, ggtt->mappable_end,
ggtt             1065 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		ggtt->vm.insert_page(&ggtt->vm,
ggtt             1072 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 	vaddr = (void __force *)io_mapping_map_atomic_wc(&ggtt->iomap,
ggtt             1977 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c 		vm = &dev_priv->ggtt.vm;
ggtt              225 drivers/gpu/drm/i915/gem/i915_gem_mman.c 	struct i915_ggtt *ggtt = &i915->ggtt;
ggtt              248 drivers/gpu/drm/i915/gem/i915_gem_mman.c 	ret = intel_gt_reset_trylock(ggtt->vm.gt, &srcu);
ggtt              301 drivers/gpu/drm/i915/gem/i915_gem_mman.c 			       (ggtt->gmadr.start + vma->node.start) >> PAGE_SHIFT,
ggtt              303 drivers/gpu/drm/i915/gem/i915_gem_mman.c 			       &ggtt->iomap);
ggtt              310 drivers/gpu/drm/i915/gem/i915_gem_mman.c 	mutex_lock(&i915->ggtt.vm.mutex);
ggtt              312 drivers/gpu/drm/i915/gem/i915_gem_mman.c 		list_add(&obj->userfault_link, &i915->ggtt.userfault_list);
ggtt              313 drivers/gpu/drm/i915/gem/i915_gem_mman.c 	mutex_unlock(&i915->ggtt.vm.mutex);
ggtt              316 drivers/gpu/drm/i915/gem/i915_gem_mman.c 		intel_wakeref_auto(&i915->ggtt.userfault_wakeref,
ggtt              332 drivers/gpu/drm/i915/gem/i915_gem_mman.c 	intel_gt_reset_unlock(ggtt->vm.gt, srcu);
ggtt              345 drivers/gpu/drm/i915/gem/i915_gem_mman.c 		if (!intel_gt_is_wedged(ggtt->vm.gt))
ggtt              418 drivers/gpu/drm/i915/gem/i915_gem_mman.c 	mutex_lock(&i915->ggtt.vm.mutex);
ggtt              435 drivers/gpu/drm/i915/gem/i915_gem_mman.c 	mutex_unlock(&i915->ggtt.vm.mutex);
ggtt              147 drivers/gpu/drm/i915/gem/i915_gem_pm.c 	intel_wakeref_auto(&i915->ggtt.userfault_wakeref, 0);
ggtt              441 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c 	mutex_lock(&i915->ggtt.vm.mutex);
ggtt              443 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c 				 &i915->ggtt.vm.bound_list, vm_link) {
ggtt              449 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c 		mutex_unlock(&i915->ggtt.vm.mutex);
ggtt              452 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c 		mutex_lock(&i915->ggtt.vm.mutex);
ggtt              454 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c 	mutex_unlock(&i915->ggtt.vm.mutex);
ggtt               69 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 	struct i915_ggtt *ggtt = &dev_priv->ggtt;
ggtt               96 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 							 ggtt_total_entries(ggtt) * 4);
ggtt              615 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 	struct i915_ggtt *ggtt = &dev_priv->ggtt;
ggtt              666 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 	vma = i915_vma_instance(obj, &ggtt->vm, NULL);
ggtt              677 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 	ret = i915_gem_gtt_reserve(&ggtt->vm, &vma->node,
ggtt              691 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 	mutex_lock(&ggtt->vm.mutex);
ggtt              692 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 	list_move_tail(&vma->vm_link, &ggtt->vm.bound_list);
ggtt              693 drivers/gpu/drm/i915/gem/i915_gem_stolen.c 	mutex_unlock(&ggtt->vm.mutex);
ggtt              938 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 	struct i915_address_space *vm = ctx->vm ?: &engine->gt->ggtt->vm;
ggtt              991 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 	struct i915_address_space *vm = ctx->vm ?: &i915->ggtt.vm;
ggtt             1451 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 	struct i915_address_space *vm = ctx->vm ?: &i915->ggtt.vm;
ggtt             1508 drivers/gpu/drm/i915/gem/selftests/huge_pages.c 	struct i915_address_space *vm = ctx->vm ?: &i915->ggtt.vm;
ggtt              174 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c 	struct i915_address_space *vm = ctx->vm ?: &engine->gt->ggtt->vm;
ggtt              313 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c 	struct i915_address_space *vm = ctx->vm ?: &ctx->i915->ggtt.vm;
ggtt             1076 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c 	vm = ctx->vm ?: &i915->ggtt.alias->vm;
ggtt              197 drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c 			      (1 + next_prime_number(i915->ggtt.vm.total >> PAGE_SHIFT)) << PAGE_SHIFT);
ggtt              336 drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c 	vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
ggtt               46 drivers/gpu/drm/i915/gem/selftests/i915_gem_object.c 			      i915->ggtt.vm.total + PAGE_SIZE);
ggtt              111 drivers/gpu/drm/i915/gem/selftests/igt_gem_utils.c 	struct i915_address_space *vm = ctx->vm ?: &engine->gt->ggtt->vm;
ggtt              227 drivers/gpu/drm/i915/gt/intel_context.c 	ce->vm = i915_vm_get(ctx->vm ?: &engine->gt->ggtt->vm);
ggtt              557 drivers/gpu/drm/i915/gt/intel_engine_cs.c 	vma = i915_vma_instance(obj, &engine->gt->ggtt->vm, NULL);
ggtt               29 drivers/gpu/drm/i915/gt/intel_gt.c 	i915->gt.ggtt = &i915->ggtt;
ggtt              240 drivers/gpu/drm/i915/gt/intel_gt.c 	vma = i915_vma_instance(obj, &gt->ggtt->vm, NULL);
ggtt               39 drivers/gpu/drm/i915/gt/intel_gt_types.h 	struct i915_ggtt *ggtt;
ggtt             2224 drivers/gpu/drm/i915/gt/intel_lrc.c 	vma = i915_vma_instance(obj, &engine->gt->ggtt->vm, NULL);
ggtt             3371 drivers/gpu/drm/i915/gt/intel_lrc.c 	vma = i915_vma_instance(ctx_obj, &engine->gt->ggtt->vm, NULL);
ggtt              196 drivers/gpu/drm/i915/gt/intel_renderstate.c 	so.vma = i915_vma_instance(so.obj, &engine->gt->ggtt->vm, NULL);
ggtt              627 drivers/gpu/drm/i915/gt/intel_reset.c 	for (i = 0; i < gt->ggtt->num_fences; i++) {
ggtt              632 drivers/gpu/drm/i915/gt/intel_reset.c 		vma = READ_ONCE(gt->ggtt->fence_regs[i].vma);
ggtt              639 drivers/gpu/drm/i915/gt/intel_reset.c 		GEM_BUG_ON(vma->fence != &gt->ggtt->fence_regs[i]);
ggtt             1267 drivers/gpu/drm/i915/gt/intel_ringbuffer.c static struct i915_vma *create_ring_vma(struct i915_ggtt *ggtt, int size)
ggtt             1269 drivers/gpu/drm/i915/gt/intel_ringbuffer.c 	struct i915_address_space *vm = &ggtt->vm;
ggtt             1327 drivers/gpu/drm/i915/gt/intel_ringbuffer.c 	vma = create_ring_vma(engine->gt->ggtt, size);
ggtt             1455 drivers/gpu/drm/i915/gt/intel_ringbuffer.c 	vma = i915_vma_instance(obj, &engine->gt->ggtt->vm, NULL);
ggtt               46 drivers/gpu/drm/i915/gt/intel_timeline.c 	vma = i915_vma_instance(obj, &gt->ggtt->vm, NULL);
ggtt             1514 drivers/gpu/drm/i915/gt/intel_workarounds.c 	vma = create_scratch(&ce->engine->gt->ggtt->vm, wal->count);
ggtt              134 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 	struct i915_address_space *vm = h->ctx->vm ?: &engine->gt->ggtt->vm;
ggtt             1335 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 	return __igt_reset_evict_vma(gt, &gt->ggtt->vm,
ggtt             1372 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 	return __igt_reset_evict_vma(gt, &gt->ggtt->vm,
ggtt              252 drivers/gpu/drm/i915/gt/selftest_lrc.c 	vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
ggtt              346 drivers/gpu/drm/i915/gt/selftest_lrc.c 	vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
ggtt               99 drivers/gpu/drm/i915/gt/selftest_workarounds.c 	vma = i915_vma_instance(result, &engine->gt->ggtt->vm, NULL);
ggtt              289 drivers/gpu/drm/i915/gt/uc/intel_guc.c 	i915_ggtt_enable_guc(gt->ggtt);
ggtt              316 drivers/gpu/drm/i915/gt/uc/intel_guc.c 	i915_ggtt_disable_guc(gt->ggtt);
ggtt              603 drivers/gpu/drm/i915/gt/uc/intel_guc.c 	vma = i915_vma_instance(obj, &gt->ggtt->vm, NULL);
ggtt              399 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c static u32 uc_fw_ggtt_offset(struct intel_uc_fw *uc_fw, struct i915_ggtt *ggtt)
ggtt              401 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 	struct drm_mm_node *node = &ggtt->uc_fw;
ggtt              414 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 	struct i915_ggtt *ggtt = gt->ggtt;
ggtt              416 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 		.node.start = uc_fw_ggtt_offset(uc_fw, ggtt),
ggtt              419 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 		.vm = &ggtt->vm,
ggtt              423 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 	GEM_BUG_ON(dummy.node.size > ggtt->uc_fw.size);
ggtt              428 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 	ggtt->vm.insert_entries(&ggtt->vm, &dummy, I915_CACHE_NONE, 0);
ggtt              435 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 	struct i915_ggtt *ggtt = gt->ggtt;
ggtt              436 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 	u64 start = uc_fw_ggtt_offset(uc_fw, ggtt);
ggtt              438 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 	ggtt->vm.clear_range(&ggtt->vm, start, obj->base.size);
ggtt              455 drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c 	offset = uc_fw_ggtt_offset(uc_fw, gt->ggtt);
ggtt               66 drivers/gpu/drm/i915/gvt/aperture_gm.c 	ret = i915_gem_gtt_insert(&dev_priv->ggtt.vm, node,
ggtt              175 drivers/gpu/drm/i915/gvt/aperture_gm.c 	mutex_lock(&dev_priv->ggtt.vm.mutex);
ggtt              182 drivers/gpu/drm/i915/gvt/aperture_gm.c 	mutex_unlock(&dev_priv->ggtt.vm.mutex);
ggtt              198 drivers/gpu/drm/i915/gvt/aperture_gm.c 	mutex_lock(&dev_priv->ggtt.vm.mutex);
ggtt              210 drivers/gpu/drm/i915/gvt/aperture_gm.c 	mutex_unlock(&dev_priv->ggtt.vm.mutex);
ggtt              223 drivers/gpu/drm/i915/gvt/aperture_gm.c 	mutex_unlock(&dev_priv->ggtt.vm.mutex);
ggtt               64 drivers/gpu/drm/i915/gvt/dmabuf.c 	gtt_entries = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm +
ggtt              280 drivers/gpu/drm/i915/gvt/gtt.c 	void __iomem *addr = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + index;
ggtt              295 drivers/gpu/drm/i915/gvt/gtt.c 	void __iomem *addr = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + index;
ggtt              380 drivers/gpu/drm/i915/gvt/gvt.h #define gvt_aperture_sz(gvt)	  (gvt->dev_priv->ggtt.mappable_end)
ggtt              381 drivers/gpu/drm/i915/gvt/gvt.h #define gvt_aperture_pa_base(gvt) (gvt->dev_priv->ggtt.gmadr.start)
ggtt              383 drivers/gpu/drm/i915/gvt/gvt.h #define gvt_ggtt_gm_sz(gvt)	  (gvt->dev_priv->ggtt.vm.total)
ggtt              385 drivers/gpu/drm/i915/gvt/gvt.h 	((gvt->dev_priv->ggtt.vm.total >> PAGE_SHIFT) << 3)
ggtt              397 drivers/gpu/drm/i915/gvt/gvt.h #define gvt_fence_sz(gvt) ((gvt)->dev_priv->ggtt.num_fences)
ggtt              916 drivers/gpu/drm/i915/gvt/kvmgt.c 	aperture_va = io_mapping_map_wc(&vgpu->gvt->dev_priv->ggtt.iomap,
ggtt              651 drivers/gpu/drm/i915/i915_debugfs.c 	seq_printf(m, "Total fences = %d\n", i915->ggtt.num_fences);
ggtt              654 drivers/gpu/drm/i915/i915_debugfs.c 	for (i = 0; i < i915->ggtt.num_fences; i++) {
ggtt              655 drivers/gpu/drm/i915/i915_debugfs.c 		struct i915_fence_reg *reg = &i915->ggtt.fence_regs[i];
ggtt              426 drivers/gpu/drm/i915/i915_drv.c 	struct i915_ggtt *ggtt = &dev_priv->ggtt;
ggtt              434 drivers/gpu/drm/i915/i915_drv.c 	ap->ranges[0].base = ggtt->gmadr.start;
ggtt              435 drivers/gpu/drm/i915/i915_drv.c 	ap->ranges[0].size = ggtt->mappable_end;
ggtt             1446 drivers/gpu/drm/i915/i915_drv.h 	struct i915_ggtt ggtt; /* VM representing the global address space */
ggtt               65 drivers/gpu/drm/i915/i915_gem.c insert_mappable_node(struct i915_ggtt *ggtt,
ggtt               69 drivers/gpu/drm/i915/i915_gem.c 	return drm_mm_insert_node_in_range(&ggtt->vm.mm, node,
ggtt               71 drivers/gpu/drm/i915/i915_gem.c 					   0, ggtt->mappable_end,
ggtt               85 drivers/gpu/drm/i915/i915_gem.c 	struct i915_ggtt *ggtt = &to_i915(dev)->ggtt;
ggtt               90 drivers/gpu/drm/i915/i915_gem.c 	mutex_lock(&ggtt->vm.mutex);
ggtt               92 drivers/gpu/drm/i915/i915_gem.c 	pinned = ggtt->vm.reserved;
ggtt               93 drivers/gpu/drm/i915/i915_gem.c 	list_for_each_entry(vma, &ggtt->vm.bound_list, vm_link)
ggtt               97 drivers/gpu/drm/i915/i915_gem.c 	mutex_unlock(&ggtt->vm.mutex);
ggtt               99 drivers/gpu/drm/i915/i915_gem.c 	args->aper_size = ggtt->vm.total;
ggtt              332 drivers/gpu/drm/i915/i915_gem.c 	struct i915_ggtt *ggtt = &i915->ggtt;
ggtt              356 drivers/gpu/drm/i915/i915_gem.c 		ret = insert_mappable_node(ggtt, &node, PAGE_SIZE);
ggtt              397 drivers/gpu/drm/i915/i915_gem.c 			ggtt->vm.insert_page(&ggtt->vm,
ggtt              404 drivers/gpu/drm/i915/i915_gem.c 		if (gtt_user_read(&ggtt->iomap, page_base, page_offset,
ggtt              419 drivers/gpu/drm/i915/i915_gem.c 		ggtt->vm.clear_range(&ggtt->vm, node.start, node.size);
ggtt              524 drivers/gpu/drm/i915/i915_gem.c 	struct i915_ggtt *ggtt = &i915->ggtt;
ggtt              566 drivers/gpu/drm/i915/i915_gem.c 		ret = insert_mappable_node(ggtt, &node, PAGE_SIZE);
ggtt              609 drivers/gpu/drm/i915/i915_gem.c 			intel_gt_flush_ggtt_writes(ggtt->vm.gt);
ggtt              610 drivers/gpu/drm/i915/i915_gem.c 			ggtt->vm.insert_page(&ggtt->vm,
ggtt              623 drivers/gpu/drm/i915/i915_gem.c 		if (ggtt_write(&ggtt->iomap, page_base, page_offset,
ggtt              638 drivers/gpu/drm/i915/i915_gem.c 	intel_gt_flush_ggtt_writes(ggtt->vm.gt);
ggtt              640 drivers/gpu/drm/i915/i915_gem.c 		ggtt->vm.clear_range(&ggtt->vm, node.start, node.size);
ggtt              859 drivers/gpu/drm/i915/i915_gem.c 				 &i915->ggtt.userfault_list, userfault_link)
ggtt              867 drivers/gpu/drm/i915/i915_gem.c 	for (i = 0; i < i915->ggtt.num_fences; i++) {
ggtt              868 drivers/gpu/drm/i915/i915_gem.c 		struct i915_fence_reg *reg = &i915->ggtt.fence_regs[i];
ggtt              966 drivers/gpu/drm/i915/i915_gem.c 	struct i915_address_space *vm = &dev_priv->ggtt.vm;
ggtt              998 drivers/gpu/drm/i915/i915_gem.c 		if (obj->base.size > dev_priv->ggtt.mappable_end)
ggtt             1017 drivers/gpu/drm/i915/i915_gem.c 		    obj->base.size > dev_priv->ggtt.mappable_end / 2)
ggtt             1031 drivers/gpu/drm/i915/i915_gem.c 			    vma->fence_size > dev_priv->ggtt.mappable_end / 2)
ggtt             1609 drivers/gpu/drm/i915/i915_gem.c 	intel_wakeref_auto_fini(&dev_priv->ggtt.userfault_wakeref);
ggtt              258 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		list_move(&fence->link, &fence->i915->ggtt.fence_list);
ggtt              282 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		list_move_tail(&fence->link, &fence->i915->ggtt.fence_list);
ggtt              318 drivers/gpu/drm/i915/i915_gem_fence_reg.c 	list_for_each_entry(fence, &i915->ggtt.fence_list, link) {
ggtt              336 drivers/gpu/drm/i915/i915_gem_fence_reg.c 	struct i915_ggtt *ggtt = i915_vm_to_ggtt(vma->vm);
ggtt              347 drivers/gpu/drm/i915/i915_gem_fence_reg.c 			list_move_tail(&fence->link, &ggtt->fence_list);
ggtt              425 drivers/gpu/drm/i915/i915_gem_fence_reg.c 	struct i915_ggtt *ggtt = &i915->ggtt;
ggtt              430 drivers/gpu/drm/i915/i915_gem_fence_reg.c 	lockdep_assert_held(&ggtt->vm.mutex);
ggtt              434 drivers/gpu/drm/i915/i915_gem_fence_reg.c 	list_for_each_entry(fence, &ggtt->fence_list, link)
ggtt              463 drivers/gpu/drm/i915/i915_gem_fence_reg.c 	struct i915_ggtt *ggtt = &fence->i915->ggtt;
ggtt              465 drivers/gpu/drm/i915/i915_gem_fence_reg.c 	lockdep_assert_held(&ggtt->vm.mutex);
ggtt              467 drivers/gpu/drm/i915/i915_gem_fence_reg.c 	list_add(&fence->link, &ggtt->fence_list);
ggtt              483 drivers/gpu/drm/i915/i915_gem_fence_reg.c 	for (i = 0; i < i915->ggtt.num_fences; i++) {
ggtt              484 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		struct i915_fence_reg *reg = &i915->ggtt.fence_regs[i];
ggtt              819 drivers/gpu/drm/i915/i915_gem_fence_reg.c void i915_ggtt_init_fences(struct i915_ggtt *ggtt)
ggtt              821 drivers/gpu/drm/i915/i915_gem_fence_reg.c 	struct drm_i915_private *i915 = ggtt->vm.i915;
ggtt              825 drivers/gpu/drm/i915/i915_gem_fence_reg.c 	INIT_LIST_HEAD(&ggtt->fence_list);
ggtt              826 drivers/gpu/drm/i915/i915_gem_fence_reg.c 	INIT_LIST_HEAD(&ggtt->userfault_list);
ggtt              827 drivers/gpu/drm/i915/i915_gem_fence_reg.c 	intel_wakeref_auto_init(&ggtt->userfault_wakeref, &i915->runtime_pm);
ggtt              847 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		struct i915_fence_reg *fence = &ggtt->fence_regs[i];
ggtt              851 drivers/gpu/drm/i915/i915_gem_fence_reg.c 		list_add_tail(&fence->link, &ggtt->fence_list);
ggtt              853 drivers/gpu/drm/i915/i915_gem_fence_reg.c 	ggtt->num_fences = num_fences;
ggtt               68 drivers/gpu/drm/i915/i915_gem_fence_reg.h void i915_ggtt_init_fences(struct i915_ggtt *ggtt);
ggtt              121 drivers/gpu/drm/i915/i915_gem_gtt.c static void gen6_ggtt_invalidate(struct i915_ggtt *ggtt)
ggtt              123 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct intel_uncore *uncore = ggtt->vm.gt->uncore;
ggtt              132 drivers/gpu/drm/i915/i915_gem_gtt.c static void guc_ggtt_invalidate(struct i915_ggtt *ggtt)
ggtt              134 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct intel_uncore *uncore = ggtt->vm.gt->uncore;
ggtt              136 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen6_ggtt_invalidate(ggtt);
ggtt              140 drivers/gpu/drm/i915/i915_gem_gtt.c static void gmch_ggtt_invalidate(struct i915_ggtt *ggtt)
ggtt             1736 drivers/gpu/drm/i915/i915_gem_gtt.c 		gen6_ggtt_invalidate(vm->gt->ggtt);
ggtt             1820 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_ggtt *ggtt = i915_vm_to_ggtt(vma->vm);
ggtt             1827 drivers/gpu/drm/i915/i915_gem_gtt.c 	ppgtt->pd_addr = (gen6_pte_t __iomem *)ggtt->gsm + ggtt_offset;
ggtt             1833 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen6_ggtt_invalidate(ggtt);
ggtt             1872 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_ggtt *ggtt = ppgtt->base.vm.gt->ggtt;
ggtt             1876 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(size > ggtt->vm.total);
ggtt             1884 drivers/gpu/drm/i915/i915_gem_gtt.c 	vma->vm = &ggtt->vm;
ggtt             1961 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_ggtt * const ggtt = &i915->ggtt;
ggtt             1977 drivers/gpu/drm/i915/i915_gem_gtt.c 	ppgtt->base.vm.pte_encode = ggtt->vm.pte_encode;
ggtt             2121 drivers/gpu/drm/i915/i915_gem_gtt.c static void ggtt_suspend_mappings(struct i915_ggtt *ggtt)
ggtt             2123 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct drm_i915_private *i915 = ggtt->vm.i915;
ggtt             2131 drivers/gpu/drm/i915/i915_gem_gtt.c 	intel_gt_check_and_clear_faults(ggtt->vm.gt);
ggtt             2133 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.clear_range(&ggtt->vm, 0, ggtt->vm.total);
ggtt             2135 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->invalidate(ggtt);
ggtt             2140 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt_suspend_mappings(&i915->ggtt);
ggtt             2180 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
ggtt             2182 drivers/gpu/drm/i915/i915_gem_gtt.c 		(gen8_pte_t __iomem *)ggtt->gsm + offset / I915_GTT_PAGE_SIZE;
ggtt             2186 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->invalidate(ggtt);
ggtt             2194 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
ggtt             2205 drivers/gpu/drm/i915/i915_gem_gtt.c 	gtt_entries = (gen8_pte_t __iomem *)ggtt->gsm;
ggtt             2214 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->invalidate(ggtt);
ggtt             2223 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
ggtt             2225 drivers/gpu/drm/i915/i915_gem_gtt.c 		(gen6_pte_t __iomem *)ggtt->gsm + offset / I915_GTT_PAGE_SIZE;
ggtt             2229 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->invalidate(ggtt);
ggtt             2243 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
ggtt             2244 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen6_pte_t __iomem *entries = (gen6_pte_t __iomem *)ggtt->gsm;
ggtt             2255 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->invalidate(ggtt);
ggtt             2266 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
ggtt             2271 drivers/gpu/drm/i915/i915_gem_gtt.c 		(gen8_pte_t __iomem *)ggtt->gsm + first_entry;
ggtt             2272 drivers/gpu/drm/i915/i915_gem_gtt.c 	const int max_entries = ggtt_total_entries(ggtt) - first_entry;
ggtt             2381 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
ggtt             2385 drivers/gpu/drm/i915/i915_gem_gtt.c 		(gen6_pte_t __iomem *)ggtt->gsm + first_entry;
ggtt             2386 drivers/gpu/drm/i915/i915_gem_gtt.c 	const int max_entries = ggtt_total_entries(ggtt) - first_entry;
ggtt             2532 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_ggtt *ggtt = &dev_priv->ggtt;
ggtt             2534 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (unlikely(ggtt->do_idle_maps)) {
ggtt             2578 drivers/gpu/drm/i915/i915_gem_gtt.c static int init_aliasing_ppgtt(struct i915_ggtt *ggtt)
ggtt             2583 drivers/gpu/drm/i915/i915_gem_gtt.c 	ppgtt = i915_ppgtt_create(ggtt->vm.i915);
ggtt             2587 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (GEM_WARN_ON(ppgtt->vm.total < ggtt->vm.total)) {
ggtt             2598 drivers/gpu/drm/i915/i915_gem_gtt.c 	err = ppgtt->vm.allocate_va_range(&ppgtt->vm, 0, ggtt->vm.total);
ggtt             2602 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->alias = ppgtt;
ggtt             2604 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(ggtt->vm.vma_ops.bind_vma != ggtt_bind_vma);
ggtt             2605 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.bind_vma = aliasing_gtt_bind_vma;
ggtt             2607 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(ggtt->vm.vma_ops.unbind_vma != ggtt_unbind_vma);
ggtt             2608 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.unbind_vma = aliasing_gtt_unbind_vma;
ggtt             2617 drivers/gpu/drm/i915/i915_gem_gtt.c static void fini_aliasing_ppgtt(struct i915_ggtt *ggtt)
ggtt             2619 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct drm_i915_private *i915 = ggtt->vm.i915;
ggtt             2624 drivers/gpu/drm/i915/i915_gem_gtt.c 	ppgtt = fetch_and_zero(&ggtt->alias);
ggtt             2630 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.bind_vma   = ggtt_bind_vma;
ggtt             2631 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.unbind_vma = ggtt_unbind_vma;
ggtt             2637 drivers/gpu/drm/i915/i915_gem_gtt.c static int ggtt_reserve_guc_top(struct i915_ggtt *ggtt)
ggtt             2642 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (!USES_GUC(ggtt->vm.i915))
ggtt             2645 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(ggtt->vm.total <= GUC_GGTT_TOP);
ggtt             2646 drivers/gpu/drm/i915/i915_gem_gtt.c 	size = ggtt->vm.total - GUC_GGTT_TOP;
ggtt             2648 drivers/gpu/drm/i915/i915_gem_gtt.c 	ret = i915_gem_gtt_reserve(&ggtt->vm, &ggtt->uc_fw, size,
ggtt             2657 drivers/gpu/drm/i915/i915_gem_gtt.c static void ggtt_release_guc_top(struct i915_ggtt *ggtt)
ggtt             2659 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (drm_mm_node_allocated(&ggtt->uc_fw))
ggtt             2660 drivers/gpu/drm/i915/i915_gem_gtt.c 		drm_mm_remove_node(&ggtt->uc_fw);
ggtt             2663 drivers/gpu/drm/i915/i915_gem_gtt.c static void cleanup_init_ggtt(struct i915_ggtt *ggtt)
ggtt             2665 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt_release_guc_top(ggtt);
ggtt             2666 drivers/gpu/drm/i915/i915_gem_gtt.c 	drm_mm_remove_node(&ggtt->error_capture);
ggtt             2669 drivers/gpu/drm/i915/i915_gem_gtt.c static int init_ggtt(struct i915_ggtt *ggtt)
ggtt             2690 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->pin_bias = max_t(u32, I915_GTT_PAGE_SIZE,
ggtt             2691 drivers/gpu/drm/i915/i915_gem_gtt.c 			       intel_wopcm_guc_size(&ggtt->vm.i915->wopcm));
ggtt             2693 drivers/gpu/drm/i915/i915_gem_gtt.c 	ret = intel_vgt_balloon(ggtt);
ggtt             2698 drivers/gpu/drm/i915/i915_gem_gtt.c 	ret = drm_mm_insert_node_in_range(&ggtt->vm.mm, &ggtt->error_capture,
ggtt             2700 drivers/gpu/drm/i915/i915_gem_gtt.c 					  0, ggtt->mappable_end,
ggtt             2710 drivers/gpu/drm/i915/i915_gem_gtt.c 	ret = ggtt_reserve_guc_top(ggtt);
ggtt             2715 drivers/gpu/drm/i915/i915_gem_gtt.c 	drm_mm_for_each_hole(entry, &ggtt->vm.mm, hole_start, hole_end) {
ggtt             2718 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->vm.clear_range(&ggtt->vm, hole_start,
ggtt             2723 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.clear_range(&ggtt->vm, ggtt->vm.total - PAGE_SIZE, PAGE_SIZE);
ggtt             2728 drivers/gpu/drm/i915/i915_gem_gtt.c 	cleanup_init_ggtt(ggtt);
ggtt             2736 drivers/gpu/drm/i915/i915_gem_gtt.c 	ret = init_ggtt(&i915->ggtt);
ggtt             2741 drivers/gpu/drm/i915/i915_gem_gtt.c 		ret = init_aliasing_ppgtt(&i915->ggtt);
ggtt             2743 drivers/gpu/drm/i915/i915_gem_gtt.c 			cleanup_init_ggtt(&i915->ggtt);
ggtt             2749 drivers/gpu/drm/i915/i915_gem_gtt.c static void ggtt_cleanup_hw(struct i915_ggtt *ggtt)
ggtt             2751 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct drm_i915_private *i915 = ggtt->vm.i915;
ggtt             2754 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.closed = true;
ggtt             2761 drivers/gpu/drm/i915/i915_gem_gtt.c 	list_for_each_entry_safe(vma, vn, &ggtt->vm.bound_list, vm_link)
ggtt             2764 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (drm_mm_node_allocated(&ggtt->error_capture))
ggtt             2765 drivers/gpu/drm/i915/i915_gem_gtt.c 		drm_mm_remove_node(&ggtt->error_capture);
ggtt             2767 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt_release_guc_top(ggtt);
ggtt             2769 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (drm_mm_initialized(&ggtt->vm.mm)) {
ggtt             2770 drivers/gpu/drm/i915/i915_gem_gtt.c 		intel_vgt_deballoon(ggtt);
ggtt             2771 drivers/gpu/drm/i915/i915_gem_gtt.c 		i915_address_space_fini(&ggtt->vm);
ggtt             2774 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.cleanup(&ggtt->vm);
ggtt             2778 drivers/gpu/drm/i915/i915_gem_gtt.c 	arch_phys_wc_del(ggtt->mtrr);
ggtt             2779 drivers/gpu/drm/i915/i915_gem_gtt.c 	io_mapping_fini(&ggtt->iomap);
ggtt             2790 drivers/gpu/drm/i915/i915_gem_gtt.c 	fini_aliasing_ppgtt(&i915->ggtt);
ggtt             2792 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt_cleanup_hw(&i915->ggtt);
ggtt             2837 drivers/gpu/drm/i915/i915_gem_gtt.c static int ggtt_probe_common(struct i915_ggtt *ggtt, u64 size)
ggtt             2839 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct drm_i915_private *dev_priv = ggtt->vm.i915;
ggtt             2855 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->gsm = ioremap_nocache(phys_addr, size);
ggtt             2857 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->gsm = ioremap_wc(phys_addr, size);
ggtt             2858 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (!ggtt->gsm) {
ggtt             2863 drivers/gpu/drm/i915/i915_gem_gtt.c 	ret = setup_scratch_page(&ggtt->vm, GFP_DMA32);
ggtt             2867 drivers/gpu/drm/i915/i915_gem_gtt.c 		iounmap(ggtt->gsm);
ggtt             2871 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.scratch[0].encode =
ggtt             2872 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->vm.pte_encode(px_dma(&ggtt->vm.scratch[0]),
ggtt             2961 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_ggtt *ggtt = i915_vm_to_ggtt(vm);
ggtt             2963 drivers/gpu/drm/i915/i915_gem_gtt.c 	iounmap(ggtt->gsm);
ggtt             2981 drivers/gpu/drm/i915/i915_gem_gtt.c static int gen8_gmch_probe(struct i915_ggtt *ggtt)
ggtt             2983 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct drm_i915_private *dev_priv = ggtt->vm.i915;
ggtt             2990 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->gmadr =
ggtt             2993 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->mappable_end = resource_size(&ggtt->gmadr);
ggtt             3007 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.total = (size / sizeof(gen8_pte_t)) * I915_GTT_PAGE_SIZE;
ggtt             3008 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.cleanup = gen6_gmch_remove;
ggtt             3009 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.insert_page = gen8_ggtt_insert_page;
ggtt             3010 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.clear_range = nop_clear_range;
ggtt             3012 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->vm.clear_range = gen8_ggtt_clear_range;
ggtt             3014 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.insert_entries = gen8_ggtt_insert_entries;
ggtt             3019 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->vm.insert_entries = bxt_vtd_ggtt_insert_entries__BKL;
ggtt             3020 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->vm.insert_page    = bxt_vtd_ggtt_insert_page__BKL;
ggtt             3021 drivers/gpu/drm/i915/i915_gem_gtt.c 		if (ggtt->vm.clear_range != nop_clear_range)
ggtt             3022 drivers/gpu/drm/i915/i915_gem_gtt.c 			ggtt->vm.clear_range = bxt_vtd_ggtt_clear_range__BKL;
ggtt             3025 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->invalidate = gen6_ggtt_invalidate;
ggtt             3027 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.bind_vma    = ggtt_bind_vma;
ggtt             3028 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.unbind_vma  = ggtt_unbind_vma;
ggtt             3029 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.set_pages   = ggtt_set_pages;
ggtt             3030 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.clear_pages = clear_pages;
ggtt             3032 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.pte_encode = gen8_pte_encode;
ggtt             3036 drivers/gpu/drm/i915/i915_gem_gtt.c 	return ggtt_probe_common(ggtt, size);
ggtt             3039 drivers/gpu/drm/i915/i915_gem_gtt.c static int gen6_gmch_probe(struct i915_ggtt *ggtt)
ggtt             3041 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct drm_i915_private *dev_priv = ggtt->vm.i915;
ggtt             3047 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->gmadr =
ggtt             3050 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->mappable_end = resource_size(&ggtt->gmadr);
ggtt             3055 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (ggtt->mappable_end < (64<<20) || ggtt->mappable_end > (512<<20)) {
ggtt             3056 drivers/gpu/drm/i915/i915_gem_gtt.c 		DRM_ERROR("Unknown GMADR size (%pa)\n", &ggtt->mappable_end);
ggtt             3068 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.total = (size / sizeof(gen6_pte_t)) * I915_GTT_PAGE_SIZE;
ggtt             3070 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.clear_range = nop_clear_range;
ggtt             3072 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->vm.clear_range = gen6_ggtt_clear_range;
ggtt             3073 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.insert_page = gen6_ggtt_insert_page;
ggtt             3074 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.insert_entries = gen6_ggtt_insert_entries;
ggtt             3075 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.cleanup = gen6_gmch_remove;
ggtt             3077 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->invalidate = gen6_ggtt_invalidate;
ggtt             3080 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->vm.pte_encode = iris_pte_encode;
ggtt             3082 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->vm.pte_encode = hsw_pte_encode;
ggtt             3084 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->vm.pte_encode = byt_pte_encode;
ggtt             3086 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->vm.pte_encode = ivb_pte_encode;
ggtt             3088 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->vm.pte_encode = snb_pte_encode;
ggtt             3090 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.bind_vma    = ggtt_bind_vma;
ggtt             3091 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.unbind_vma  = ggtt_unbind_vma;
ggtt             3092 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.set_pages   = ggtt_set_pages;
ggtt             3093 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.clear_pages = clear_pages;
ggtt             3095 drivers/gpu/drm/i915/i915_gem_gtt.c 	return ggtt_probe_common(ggtt, size);
ggtt             3103 drivers/gpu/drm/i915/i915_gem_gtt.c static int i915_gmch_probe(struct i915_ggtt *ggtt)
ggtt             3105 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct drm_i915_private *dev_priv = ggtt->vm.i915;
ggtt             3115 drivers/gpu/drm/i915/i915_gem_gtt.c 	intel_gtt_get(&ggtt->vm.total, &gmadr_base, &ggtt->mappable_end);
ggtt             3117 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->gmadr =
ggtt             3119 drivers/gpu/drm/i915/i915_gem_gtt.c 						 ggtt->mappable_end);
ggtt             3121 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->do_idle_maps = needs_idle_maps(dev_priv);
ggtt             3122 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.insert_page = i915_ggtt_insert_page;
ggtt             3123 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.insert_entries = i915_ggtt_insert_entries;
ggtt             3124 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.clear_range = i915_ggtt_clear_range;
ggtt             3125 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.cleanup = i915_gmch_remove;
ggtt             3127 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->invalidate = gmch_ggtt_invalidate;
ggtt             3129 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.bind_vma    = ggtt_bind_vma;
ggtt             3130 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.unbind_vma  = ggtt_unbind_vma;
ggtt             3131 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.set_pages   = ggtt_set_pages;
ggtt             3132 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.vma_ops.clear_pages = clear_pages;
ggtt             3134 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (unlikely(ggtt->do_idle_maps))
ggtt             3141 drivers/gpu/drm/i915/i915_gem_gtt.c static int ggtt_probe_hw(struct i915_ggtt *ggtt, struct intel_gt *gt)
ggtt             3146 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.gt = gt;
ggtt             3147 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.i915 = i915;
ggtt             3148 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.dma = &i915->drm.pdev->dev;
ggtt             3151 drivers/gpu/drm/i915/i915_gem_gtt.c 		ret = i915_gmch_probe(ggtt);
ggtt             3153 drivers/gpu/drm/i915/i915_gem_gtt.c 		ret = gen6_gmch_probe(ggtt);
ggtt             3155 drivers/gpu/drm/i915/i915_gem_gtt.c 		ret = gen8_gmch_probe(ggtt);
ggtt             3159 drivers/gpu/drm/i915/i915_gem_gtt.c 	if ((ggtt->vm.total - 1) >> 32) {
ggtt             3162 drivers/gpu/drm/i915/i915_gem_gtt.c 			  ggtt->vm.total >> 20);
ggtt             3163 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->vm.total = 1ULL << 32;
ggtt             3164 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->mappable_end =
ggtt             3165 drivers/gpu/drm/i915/i915_gem_gtt.c 			min_t(u64, ggtt->mappable_end, ggtt->vm.total);
ggtt             3168 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (ggtt->mappable_end > ggtt->vm.total) {
ggtt             3171 drivers/gpu/drm/i915/i915_gem_gtt.c 			  &ggtt->mappable_end, ggtt->vm.total);
ggtt             3172 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->mappable_end = ggtt->vm.total;
ggtt             3176 drivers/gpu/drm/i915/i915_gem_gtt.c 	DRM_DEBUG_DRIVER("GGTT size = %lluM\n", ggtt->vm.total >> 20);
ggtt             3177 drivers/gpu/drm/i915/i915_gem_gtt.c 	DRM_DEBUG_DRIVER("GMADR size = %lluM\n", (u64)ggtt->mappable_end >> 20);
ggtt             3192 drivers/gpu/drm/i915/i915_gem_gtt.c 	ret = ggtt_probe_hw(&i915->ggtt, &i915->gt);
ggtt             3202 drivers/gpu/drm/i915/i915_gem_gtt.c static int ggtt_init_hw(struct i915_ggtt *ggtt)
ggtt             3204 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct drm_i915_private *i915 = ggtt->vm.i915;
ggtt             3209 drivers/gpu/drm/i915/i915_gem_gtt.c 	i915_address_space_init(&ggtt->vm, VM_CLASS_GGTT);
ggtt             3211 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.is_ggtt = true;
ggtt             3214 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.has_read_only = IS_VALLEYVIEW(i915);
ggtt             3217 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->vm.mm.color_adjust = i915_gtt_color_adjust;
ggtt             3219 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (!io_mapping_init_wc(&ggtt->iomap,
ggtt             3220 drivers/gpu/drm/i915/i915_gem_gtt.c 				ggtt->gmadr.start,
ggtt             3221 drivers/gpu/drm/i915/i915_gem_gtt.c 				ggtt->mappable_end)) {
ggtt             3222 drivers/gpu/drm/i915/i915_gem_gtt.c 		ggtt->vm.cleanup(&ggtt->vm);
ggtt             3227 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->mtrr = arch_phys_wc_add(ggtt->gmadr.start, ggtt->mappable_end);
ggtt             3229 drivers/gpu/drm/i915/i915_gem_gtt.c 	i915_ggtt_init_fences(ggtt);
ggtt             3252 drivers/gpu/drm/i915/i915_gem_gtt.c 	ret = ggtt_init_hw(&dev_priv->ggtt);
ggtt             3267 drivers/gpu/drm/i915/i915_gem_gtt.c 	dev_priv->ggtt.vm.cleanup(&dev_priv->ggtt.vm);
ggtt             3279 drivers/gpu/drm/i915/i915_gem_gtt.c void i915_ggtt_enable_guc(struct i915_ggtt *ggtt)
ggtt             3281 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(ggtt->invalidate != gen6_ggtt_invalidate);
ggtt             3283 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->invalidate = guc_ggtt_invalidate;
ggtt             3285 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->invalidate(ggtt);
ggtt             3288 drivers/gpu/drm/i915/i915_gem_gtt.c void i915_ggtt_disable_guc(struct i915_ggtt *ggtt)
ggtt             3291 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (ggtt->invalidate == gen6_ggtt_invalidate)
ggtt             3295 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(ggtt->invalidate != guc_ggtt_invalidate);
ggtt             3297 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->invalidate = gen6_ggtt_invalidate;
ggtt             3299 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->invalidate(ggtt);
ggtt             3302 drivers/gpu/drm/i915/i915_gem_gtt.c static void ggtt_restore_mappings(struct i915_ggtt *ggtt)
ggtt             3307 drivers/gpu/drm/i915/i915_gem_gtt.c 	intel_gt_check_and_clear_faults(ggtt->vm.gt);
ggtt             3309 drivers/gpu/drm/i915/i915_gem_gtt.c 	mutex_lock(&ggtt->vm.mutex);
ggtt             3312 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.clear_range(&ggtt->vm, 0, ggtt->vm.total);
ggtt             3313 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.closed = true; /* skip rewriting PTE on VMA unbind */
ggtt             3316 drivers/gpu/drm/i915/i915_gem_gtt.c 	list_for_each_entry_safe(vma, vn, &ggtt->vm.bound_list, vm_link) {
ggtt             3322 drivers/gpu/drm/i915/i915_gem_gtt.c 		mutex_unlock(&ggtt->vm.mutex);
ggtt             3336 drivers/gpu/drm/i915/i915_gem_gtt.c 		mutex_lock(&ggtt->vm.mutex);
ggtt             3339 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->vm.closed = false;
ggtt             3340 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt->invalidate(ggtt);
ggtt             3342 drivers/gpu/drm/i915/i915_gem_gtt.c 	mutex_unlock(&ggtt->vm.mutex);
ggtt             3350 drivers/gpu/drm/i915/i915_gem_gtt.c 	ggtt_restore_mappings(&i915->ggtt);
ggtt             3642 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(vm == &vm->i915->ggtt.alias->vm);
ggtt             3739 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(vm == &vm->i915->ggtt.alias->vm);
ggtt               76 drivers/gpu/drm/i915/i915_gem_gtt.h #define ggtt_total_entries(ggtt) ((ggtt)->vm.total >> PAGE_SHIFT)
ggtt              395 drivers/gpu/drm/i915/i915_gem_gtt.h 	void (*invalidate)(struct i915_ggtt *ggtt);
ggtt              557 drivers/gpu/drm/i915/i915_gem_gtt.h void i915_ggtt_enable_guc(struct i915_ggtt *ggtt);
ggtt              558 drivers/gpu/drm/i915/i915_gem_gtt.h void i915_ggtt_disable_guc(struct i915_ggtt *ggtt);
ggtt               31 drivers/gpu/drm/i915/i915_getparam.c 		value = i915->ggtt.num_fences;
ggtt              961 drivers/gpu/drm/i915/i915_gpu_error.c 	struct i915_ggtt *ggtt = &i915->ggtt;
ggtt              962 drivers/gpu/drm/i915/i915_gpu_error.c 	const u64 slot = ggtt->error_capture.start;
ggtt              995 drivers/gpu/drm/i915/i915_gpu_error.c 		ggtt->vm.insert_page(&ggtt->vm, dma, slot, I915_CACHE_NONE, 0);
ggtt              997 drivers/gpu/drm/i915/i915_gpu_error.c 		s = io_mapping_map_wc(&ggtt->iomap, slot, PAGE_SIZE);
ggtt             1045 drivers/gpu/drm/i915/i915_gpu_error.c 		for (i = 0; i < dev_priv->ggtt.num_fences; i++)
ggtt             1050 drivers/gpu/drm/i915/i915_gpu_error.c 		for (i = 0; i < dev_priv->ggtt.num_fences; i++)
ggtt             1055 drivers/gpu/drm/i915/i915_gpu_error.c 		for (i = 0; i < dev_priv->ggtt.num_fences; i++)
ggtt             1666 drivers/gpu/drm/i915/i915_gpu_error.c 	struct i915_ggtt *ggtt = &error->i915->ggtt;
ggtt             1667 drivers/gpu/drm/i915/i915_gpu_error.c 	const u64 slot = ggtt->error_capture.start;
ggtt             1669 drivers/gpu/drm/i915/i915_gpu_error.c 	ggtt->vm.clear_range(&ggtt->vm, slot, PAGE_SIZE);
ggtt              120 drivers/gpu/drm/i915/i915_vgpu.c static void vgt_deballoon_space(struct i915_ggtt *ggtt,
ggtt              131 drivers/gpu/drm/i915/i915_vgpu.c 	ggtt->vm.reserved -= node->size;
ggtt              142 drivers/gpu/drm/i915/i915_vgpu.c void intel_vgt_deballoon(struct i915_ggtt *ggtt)
ggtt              146 drivers/gpu/drm/i915/i915_vgpu.c 	if (!intel_vgpu_active(ggtt->vm.i915))
ggtt              152 drivers/gpu/drm/i915/i915_vgpu.c 		vgt_deballoon_space(ggtt, &bl_info.space[i]);
ggtt              155 drivers/gpu/drm/i915/i915_vgpu.c static int vgt_balloon_space(struct i915_ggtt *ggtt,
ggtt              167 drivers/gpu/drm/i915/i915_vgpu.c 	ret = i915_gem_gtt_reserve(&ggtt->vm, node,
ggtt              171 drivers/gpu/drm/i915/i915_vgpu.c 		ggtt->vm.reserved += size;
ggtt              220 drivers/gpu/drm/i915/i915_vgpu.c int intel_vgt_balloon(struct i915_ggtt *ggtt)
ggtt              222 drivers/gpu/drm/i915/i915_vgpu.c 	struct intel_uncore *uncore = &ggtt->vm.i915->uncore;
ggtt              223 drivers/gpu/drm/i915/i915_vgpu.c 	unsigned long ggtt_end = ggtt->vm.total;
ggtt              229 drivers/gpu/drm/i915/i915_vgpu.c 	if (!intel_vgpu_active(ggtt->vm.i915))
ggtt              250 drivers/gpu/drm/i915/i915_vgpu.c 	if (mappable_end > ggtt->mappable_end ||
ggtt              251 drivers/gpu/drm/i915/i915_vgpu.c 	    unmappable_base < ggtt->mappable_end ||
ggtt              258 drivers/gpu/drm/i915/i915_vgpu.c 	if (unmappable_base > ggtt->mappable_end) {
ggtt              259 drivers/gpu/drm/i915/i915_vgpu.c 		ret = vgt_balloon_space(ggtt, &bl_info.space[2],
ggtt              260 drivers/gpu/drm/i915/i915_vgpu.c 					ggtt->mappable_end, unmappable_base);
ggtt              267 drivers/gpu/drm/i915/i915_vgpu.c 		ret = vgt_balloon_space(ggtt, &bl_info.space[3],
ggtt              275 drivers/gpu/drm/i915/i915_vgpu.c 		ret = vgt_balloon_space(ggtt, &bl_info.space[0],
ggtt              282 drivers/gpu/drm/i915/i915_vgpu.c 	if (mappable_end < ggtt->mappable_end) {
ggtt              283 drivers/gpu/drm/i915/i915_vgpu.c 		ret = vgt_balloon_space(ggtt, &bl_info.space[1],
ggtt              284 drivers/gpu/drm/i915/i915_vgpu.c 					mappable_end, ggtt->mappable_end);
ggtt              294 drivers/gpu/drm/i915/i915_vgpu.c 	vgt_deballoon_space(ggtt, &bl_info.space[0]);
ggtt              296 drivers/gpu/drm/i915/i915_vgpu.c 	vgt_deballoon_space(ggtt, &bl_info.space[3]);
ggtt              298 drivers/gpu/drm/i915/i915_vgpu.c 	vgt_deballoon_space(ggtt, &bl_info.space[2]);
ggtt               46 drivers/gpu/drm/i915/i915_vgpu.h int intel_vgt_balloon(struct i915_ggtt *ggtt);
ggtt               47 drivers/gpu/drm/i915/i915_vgpu.h void intel_vgt_deballoon(struct i915_ggtt *ggtt);
ggtt              107 drivers/gpu/drm/i915/i915_vma.c 	GEM_BUG_ON(vm == &vm->i915->ggtt.alias->vm);
ggtt              570 drivers/gpu/drm/i915/i915_vma.c 		end = min_t(u64, end, dev_priv->ggtt.mappable_end);
ggtt               39 drivers/gpu/drm/i915/selftests/i915_gem.c 	struct i915_ggtt *ggtt = &i915->ggtt;
ggtt               40 drivers/gpu/drm/i915/selftests/i915_gem.c 	const u64 slot = ggtt->error_capture.start;
ggtt               50 drivers/gpu/drm/i915/selftests/i915_gem.c 		ggtt->vm.insert_page(&ggtt->vm, dma, slot, I915_CACHE_NONE, 0);
ggtt               52 drivers/gpu/drm/i915/selftests/i915_gem.c 		s = io_mapping_map_atomic_wc(&ggtt->iomap, slot);
ggtt               60 drivers/gpu/drm/i915/selftests/i915_gem.c 	ggtt->vm.clear_range(&ggtt->vm, slot, PAGE_SIZE);
ggtt               73 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 		 count, i915->ggtt.vm.total / PAGE_SIZE);
ggtt               99 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	if (list_empty(&i915->ggtt.vm.bound_list)) {
ggtt              109 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	struct i915_ggtt *ggtt = &i915->ggtt;
ggtt              112 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	mutex_lock(&ggtt->vm.mutex);
ggtt              113 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	list_for_each_entry(vma, &i915->ggtt.vm.bound_list, vm_link)
ggtt              116 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	mutex_unlock(&ggtt->vm.mutex);
ggtt              140 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	struct i915_ggtt *ggtt = &i915->ggtt;
ggtt              151 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	err = i915_gem_evict_something(&ggtt->vm,
ggtt              164 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	err = i915_gem_evict_something(&ggtt->vm,
ggtt              218 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	struct i915_ggtt *ggtt = &i915->ggtt;
ggtt              233 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	err = i915_gem_evict_for_node(&ggtt->vm, &target, 0);
ggtt              243 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	err = i915_gem_evict_for_node(&ggtt->vm, &target, 0);
ggtt              265 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	struct i915_ggtt *ggtt = &i915->ggtt;
ggtt              282 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	ggtt->vm.mm.color_adjust = mock_color_adjust;
ggtt              320 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	err = i915_gem_evict_for_node(&ggtt->vm, &target, 0);
ggtt              331 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	err = i915_gem_evict_for_node(&ggtt->vm, &target, 0);
ggtt              343 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	ggtt->vm.mm.color_adjust = NULL;
ggtt              350 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	struct i915_ggtt *ggtt = &i915->ggtt;
ggtt              361 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	err = i915_gem_evict_vm(&ggtt->vm);
ggtt              370 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	err = i915_gem_evict_vm(&ggtt->vm);
ggtt              416 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 	err = i915_gem_gtt_insert(&i915->ggtt.vm, &hole,
ggtt              418 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 				  0, i915->ggtt.vm.total,
ggtt              434 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 		if (i915_gem_gtt_insert(&i915->ggtt.vm, &r->node,
ggtt              436 drivers/gpu/drm/i915/selftests/i915_gem_evict.c 					0, i915->ggtt.vm.total,
ggtt             1082 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	struct i915_ggtt *ggtt = &i915->ggtt;
ggtt             1090 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	list_sort(NULL, &ggtt->vm.mm.hole_stack, sort_holes);
ggtt             1091 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	drm_mm_for_each_hole(node, &ggtt->vm.mm, hole_start, hole_end) {
ggtt             1095 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		if (ggtt->vm.mm.color_adjust)
ggtt             1096 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			ggtt->vm.mm.color_adjust(node, 0,
ggtt             1101 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		err = func(i915, &ggtt->vm, hole_start, hole_end, end_time);
ggtt             1144 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	struct i915_ggtt *ggtt = &i915->ggtt;
ggtt             1164 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	err = drm_mm_insert_node_in_range(&ggtt->vm.mm, &tmp,
ggtt             1167 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 					  0, ggtt->mappable_end,
ggtt             1177 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		ggtt->vm.insert_page(&ggtt->vm,
ggtt             1192 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		vaddr = io_mapping_map_atomic_wc(&ggtt->iomap, offset);
ggtt             1196 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	intel_gt_flush_ggtt_writes(ggtt->vm.gt);
ggtt             1204 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		vaddr = io_mapping_map_atomic_wc(&ggtt->iomap, offset);
ggtt             1218 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	ggtt->vm.clear_range(&ggtt->vm, tmp.start, tmp.size);
ggtt             1267 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	struct i915_ggtt *ggtt = arg;
ggtt             1269 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	return exercise_mock(ggtt->vm.i915, fill_hole);
ggtt             1274 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	struct i915_ggtt *ggtt = arg;
ggtt             1276 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	return exercise_mock(ggtt->vm.i915, walk_hole);
ggtt             1281 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	struct i915_ggtt *ggtt = arg;
ggtt             1283 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	return exercise_mock(ggtt->vm.i915, pot_hole);
ggtt             1288 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	struct i915_ggtt *ggtt = arg;
ggtt             1290 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	return exercise_mock(ggtt->vm.i915, drunk_hole);
ggtt             1295 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	struct i915_ggtt *ggtt = arg;
ggtt             1308 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	     total + 2 * I915_GTT_PAGE_SIZE <= ggtt->vm.total;
ggtt             1312 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		obj = i915_gem_object_create_internal(ggtt->vm.i915,
ggtt             1327 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		vma = i915_vma_instance(obj, &ggtt->vm, NULL);
ggtt             1333 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		err = i915_gem_gtt_reserve(&ggtt->vm, &vma->node,
ggtt             1340 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			       total, ggtt->vm.total, err);
ggtt             1358 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	     total + 2 * I915_GTT_PAGE_SIZE <= ggtt->vm.total;
ggtt             1362 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		obj = i915_gem_object_create_internal(ggtt->vm.i915,
ggtt             1377 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		vma = i915_vma_instance(obj, &ggtt->vm, NULL);
ggtt             1383 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		err = i915_gem_gtt_reserve(&ggtt->vm, &vma->node,
ggtt             1390 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			       total, ggtt->vm.total, err);
ggtt             1411 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		vma = i915_vma_instance(obj, &ggtt->vm, NULL);
ggtt             1423 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		offset = random_offset(0, ggtt->vm.total,
ggtt             1427 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		err = i915_gem_gtt_reserve(&ggtt->vm, &vma->node,
ggtt             1434 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			       total, ggtt->vm.total, err);
ggtt             1460 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	struct i915_ggtt *ggtt = arg;
ggtt             1469 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			ggtt->vm.total + I915_GTT_PAGE_SIZE, 0,
ggtt             1470 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			0, ggtt->vm.total,
ggtt             1500 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		err = i915_gem_gtt_insert(&ggtt->vm, &tmp,
ggtt             1515 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	     total + I915_GTT_PAGE_SIZE <= ggtt->vm.total;
ggtt             1519 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		obj = i915_gem_object_create_internal(ggtt->vm.i915,
ggtt             1534 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		vma = i915_vma_instance(obj, &ggtt->vm, NULL);
ggtt             1540 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		err = i915_gem_gtt_insert(&ggtt->vm, &vma->node,
ggtt             1542 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 					  0, ggtt->vm.total,
ggtt             1551 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			       total, ggtt->vm.total, err);
ggtt             1563 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		vma = i915_vma_instance(obj, &ggtt->vm, NULL);
ggtt             1583 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		vma = i915_vma_instance(obj, &ggtt->vm, NULL);
ggtt             1598 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		err = i915_gem_gtt_insert(&ggtt->vm, &vma->node,
ggtt             1600 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 					  0, ggtt->vm.total,
ggtt             1604 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			       total, ggtt->vm.total, err);
ggtt             1620 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	     total + 2 * I915_GTT_PAGE_SIZE <= ggtt->vm.total;
ggtt             1624 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		obj = i915_gem_object_create_internal(ggtt->vm.i915,
ggtt             1639 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		vma = i915_vma_instance(obj, &ggtt->vm, NULL);
ggtt             1645 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 		err = i915_gem_gtt_insert(&ggtt->vm, &vma->node,
ggtt             1647 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 					  0, ggtt->vm.total,
ggtt             1651 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 			       total, ggtt->vm.total, err);
ggtt             1678 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	struct i915_ggtt *ggtt;
ggtt             1685 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	ggtt = kmalloc(sizeof(*ggtt), GFP_KERNEL);
ggtt             1686 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	if (!ggtt) {
ggtt             1690 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	mock_init_ggtt(i915, ggtt);
ggtt             1693 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	err = i915_subtests(tests, ggtt);
ggtt             1699 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	mock_fini_ggtt(ggtt);
ggtt             1700 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	kfree(ggtt);
ggtt             1725 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c 	GEM_BUG_ON(offset_in_page(i915->ggtt.vm.total));
ggtt              640 drivers/gpu/drm/i915/selftests/i915_request.c 	vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
ggtt              764 drivers/gpu/drm/i915/selftests/i915_request.c 	struct i915_address_space *vm = ctx->vm ?: &i915->ggtt.vm;
ggtt              147 drivers/gpu/drm/i915/selftests/i915_vma.c 	struct i915_ggtt *ggtt = arg;
ggtt              148 drivers/gpu/drm/i915/selftests/i915_vma.c 	struct drm_i915_private *i915 = ggtt->vm.i915;
ggtt              254 drivers/gpu/drm/i915/selftests/i915_vma.c 	struct i915_ggtt *ggtt = arg;
ggtt              265 drivers/gpu/drm/i915/selftests/i915_vma.c 		VALID(0, PIN_GLOBAL | PIN_OFFSET_BIAS | (ggtt->mappable_end - 4096)),
ggtt              266 drivers/gpu/drm/i915/selftests/i915_vma.c 		VALID(0, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_BIAS | (ggtt->mappable_end - 4096)),
ggtt              267 drivers/gpu/drm/i915/selftests/i915_vma.c 		VALID(0, PIN_GLOBAL | PIN_OFFSET_BIAS | (ggtt->vm.total - 4096)),
ggtt              269 drivers/gpu/drm/i915/selftests/i915_vma.c 		VALID(0, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_FIXED | (ggtt->mappable_end - 4096)),
ggtt              270 drivers/gpu/drm/i915/selftests/i915_vma.c 		INVALID(0, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_FIXED | ggtt->mappable_end),
ggtt              271 drivers/gpu/drm/i915/selftests/i915_vma.c 		VALID(0, PIN_GLOBAL | PIN_OFFSET_FIXED | (ggtt->vm.total - 4096)),
ggtt              272 drivers/gpu/drm/i915/selftests/i915_vma.c 		INVALID(0, PIN_GLOBAL | PIN_OFFSET_FIXED | ggtt->vm.total),
ggtt              277 drivers/gpu/drm/i915/selftests/i915_vma.c 		VALID(ggtt->mappable_end - 4096, PIN_GLOBAL | PIN_MAPPABLE),
ggtt              278 drivers/gpu/drm/i915/selftests/i915_vma.c 		VALID(ggtt->mappable_end, PIN_GLOBAL | PIN_MAPPABLE),
ggtt              279 drivers/gpu/drm/i915/selftests/i915_vma.c 		NOSPACE(ggtt->mappable_end + 4096, PIN_GLOBAL | PIN_MAPPABLE),
ggtt              280 drivers/gpu/drm/i915/selftests/i915_vma.c 		VALID(ggtt->vm.total - 4096, PIN_GLOBAL),
ggtt              281 drivers/gpu/drm/i915/selftests/i915_vma.c 		VALID(ggtt->vm.total, PIN_GLOBAL),
ggtt              282 drivers/gpu/drm/i915/selftests/i915_vma.c 		NOSPACE(ggtt->vm.total + 4096, PIN_GLOBAL),
ggtt              284 drivers/gpu/drm/i915/selftests/i915_vma.c 		INVALID(8192, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_FIXED | (ggtt->mappable_end - 4096)),
ggtt              285 drivers/gpu/drm/i915/selftests/i915_vma.c 		INVALID(8192, PIN_GLOBAL | PIN_OFFSET_FIXED | (ggtt->vm.total - 4096)),
ggtt              288 drivers/gpu/drm/i915/selftests/i915_vma.c 		VALID(8192, PIN_GLOBAL | PIN_OFFSET_BIAS | (ggtt->mappable_end - 4096)),
ggtt              296 drivers/gpu/drm/i915/selftests/i915_vma.c 		NOSPACE(0, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_BIAS | ggtt->mappable_end),
ggtt              297 drivers/gpu/drm/i915/selftests/i915_vma.c 		NOSPACE(0, PIN_GLOBAL | PIN_OFFSET_BIAS | ggtt->vm.total),
ggtt              298 drivers/gpu/drm/i915/selftests/i915_vma.c 		NOSPACE(8192, PIN_GLOBAL | PIN_MAPPABLE | PIN_OFFSET_BIAS | (ggtt->mappable_end - 4096)),
ggtt              299 drivers/gpu/drm/i915/selftests/i915_vma.c 		NOSPACE(8192, PIN_GLOBAL | PIN_OFFSET_BIAS | (ggtt->vm.total - 4096)),
ggtt              315 drivers/gpu/drm/i915/selftests/i915_vma.c 	GEM_BUG_ON(!drm_mm_clean(&ggtt->vm.mm));
ggtt              317 drivers/gpu/drm/i915/selftests/i915_vma.c 	obj = i915_gem_object_create_internal(ggtt->vm.i915, PAGE_SIZE);
ggtt              321 drivers/gpu/drm/i915/selftests/i915_vma.c 	vma = checked_vma_instance(obj, &ggtt->vm, NULL);
ggtt              475 drivers/gpu/drm/i915/selftests/i915_vma.c 	struct i915_ggtt *ggtt = arg;
ggtt              476 drivers/gpu/drm/i915/selftests/i915_vma.c 	struct i915_address_space *vm = &ggtt->vm;
ggtt              697 drivers/gpu/drm/i915/selftests/i915_vma.c 	struct i915_ggtt *ggtt = arg;
ggtt              698 drivers/gpu/drm/i915/selftests/i915_vma.c 	struct i915_address_space *vm = &ggtt->vm;
ggtt              820 drivers/gpu/drm/i915/selftests/i915_vma.c 	struct i915_ggtt *ggtt;
ggtt              827 drivers/gpu/drm/i915/selftests/i915_vma.c 	ggtt = kmalloc(sizeof(*ggtt), GFP_KERNEL);
ggtt              828 drivers/gpu/drm/i915/selftests/i915_vma.c 	if (!ggtt) {
ggtt              832 drivers/gpu/drm/i915/selftests/i915_vma.c 	mock_init_ggtt(i915, ggtt);
ggtt              835 drivers/gpu/drm/i915/selftests/i915_vma.c 	err = i915_subtests(tests, ggtt);
ggtt              841 drivers/gpu/drm/i915/selftests/i915_vma.c 	mock_fini_ggtt(ggtt);
ggtt              842 drivers/gpu/drm/i915/selftests/i915_vma.c 	kfree(ggtt);
ggtt               77 drivers/gpu/drm/i915/selftests/mock_gem_device.c 	mock_fini_ggtt(&i915->ggtt);
ggtt              201 drivers/gpu/drm/i915/selftests/mock_gem_device.c 	mock_init_ggtt(i915, &i915->ggtt);
ggtt               97 drivers/gpu/drm/i915/selftests/mock_gtt.c void mock_init_ggtt(struct drm_i915_private *i915, struct i915_ggtt *ggtt)
ggtt               99 drivers/gpu/drm/i915/selftests/mock_gtt.c 	memset(ggtt, 0, sizeof(*ggtt));
ggtt              101 drivers/gpu/drm/i915/selftests/mock_gtt.c 	ggtt->vm.gt = &i915->gt;
ggtt              102 drivers/gpu/drm/i915/selftests/mock_gtt.c 	ggtt->vm.i915 = i915;
ggtt              103 drivers/gpu/drm/i915/selftests/mock_gtt.c 	ggtt->vm.is_ggtt = true;
ggtt              105 drivers/gpu/drm/i915/selftests/mock_gtt.c 	ggtt->gmadr = (struct resource) DEFINE_RES_MEM(0, 2048 * PAGE_SIZE);
ggtt              106 drivers/gpu/drm/i915/selftests/mock_gtt.c 	ggtt->mappable_end = resource_size(&ggtt->gmadr);
ggtt              107 drivers/gpu/drm/i915/selftests/mock_gtt.c 	ggtt->vm.total = 4096 * PAGE_SIZE;
ggtt              109 drivers/gpu/drm/i915/selftests/mock_gtt.c 	ggtt->vm.clear_range = nop_clear_range;
ggtt              110 drivers/gpu/drm/i915/selftests/mock_gtt.c 	ggtt->vm.insert_page = mock_insert_page;
ggtt              111 drivers/gpu/drm/i915/selftests/mock_gtt.c 	ggtt->vm.insert_entries = mock_insert_entries;
ggtt              112 drivers/gpu/drm/i915/selftests/mock_gtt.c 	ggtt->vm.cleanup = mock_cleanup;
ggtt              114 drivers/gpu/drm/i915/selftests/mock_gtt.c 	ggtt->vm.vma_ops.bind_vma    = mock_bind_ggtt;
ggtt              115 drivers/gpu/drm/i915/selftests/mock_gtt.c 	ggtt->vm.vma_ops.unbind_vma  = mock_unbind_ggtt;
ggtt              116 drivers/gpu/drm/i915/selftests/mock_gtt.c 	ggtt->vm.vma_ops.set_pages   = ggtt_set_pages;
ggtt              117 drivers/gpu/drm/i915/selftests/mock_gtt.c 	ggtt->vm.vma_ops.clear_pages = clear_pages;
ggtt              119 drivers/gpu/drm/i915/selftests/mock_gtt.c 	i915_address_space_init(&ggtt->vm, VM_CLASS_GGTT);
ggtt              124 drivers/gpu/drm/i915/selftests/mock_gtt.c void mock_fini_ggtt(struct i915_ggtt *ggtt)
ggtt              126 drivers/gpu/drm/i915/selftests/mock_gtt.c 	i915_address_space_fini(&ggtt->vm);
ggtt               28 drivers/gpu/drm/i915/selftests/mock_gtt.h void mock_init_ggtt(struct drm_i915_private *i915, struct i915_ggtt *ggtt);
ggtt               29 drivers/gpu/drm/i915/selftests/mock_gtt.h void mock_fini_ggtt(struct i915_ggtt *ggtt);