spt               718 arch/powerpc/include/asm/book3s/64/mmu-hash.h 	struct subpage_prot_table *spt;
spt               192 arch/powerpc/include/asm/book3s/64/mmu.h 	return ctx->hash_context->spt;
spt              1166 arch/powerpc/mm/book3s64/hash_utils.c 	struct subpage_prot_table *spt = mm_ctx_subpage_prot(&mm->context);
spt              1170 arch/powerpc/mm/book3s64/hash_utils.c 	if (!spt)
spt              1173 arch/powerpc/mm/book3s64/hash_utils.c 	if (ea >= spt->maxaddr)
spt              1177 arch/powerpc/mm/book3s64/hash_utils.c 		sbpm = spt->low_prot;
spt              1179 arch/powerpc/mm/book3s64/hash_utils.c 		sbpm = spt->protptrs[ea >> SBP_L3_SHIFT];
spt               122 arch/powerpc/mm/book3s64/mmu_context.c 		if (current->mm->context.hash_context->spt) {
spt               123 arch/powerpc/mm/book3s64/mmu_context.c 			mm->context.hash_context->spt = kmalloc(sizeof(struct subpage_prot_table),
spt               125 arch/powerpc/mm/book3s64/mmu_context.c 			if (!mm->context.hash_context->spt) {
spt               136 arch/powerpc/mm/book3s64/mmu_context.c 		kfree(mm->context.hash_context->spt);
spt                24 arch/powerpc/mm/book3s64/subpage_prot.c 	struct subpage_prot_table *spt = mm_ctx_subpage_prot(&mm->context);
spt                28 arch/powerpc/mm/book3s64/subpage_prot.c 	if (!spt)
spt                32 arch/powerpc/mm/book3s64/subpage_prot.c 		if (spt->low_prot[i]) {
spt                33 arch/powerpc/mm/book3s64/subpage_prot.c 			free_page((unsigned long)spt->low_prot[i]);
spt                34 arch/powerpc/mm/book3s64/subpage_prot.c 			spt->low_prot[i] = NULL;
spt                39 arch/powerpc/mm/book3s64/subpage_prot.c 		p = spt->protptrs[i];
spt                42 arch/powerpc/mm/book3s64/subpage_prot.c 		spt->protptrs[i] = NULL;
spt                43 arch/powerpc/mm/book3s64/subpage_prot.c 		for (j = 0; j < SBP_L2_COUNT && addr < spt->maxaddr;
spt                49 arch/powerpc/mm/book3s64/subpage_prot.c 	spt->maxaddr = 0;
spt                50 arch/powerpc/mm/book3s64/subpage_prot.c 	kfree(spt);
spt                89 arch/powerpc/mm/book3s64/subpage_prot.c 	struct subpage_prot_table *spt;
spt                97 arch/powerpc/mm/book3s64/subpage_prot.c 	spt = mm_ctx_subpage_prot(&mm->context);
spt                98 arch/powerpc/mm/book3s64/subpage_prot.c 	if (!spt)
spt               102 arch/powerpc/mm/book3s64/subpage_prot.c 	if (limit > spt->maxaddr)
spt               103 arch/powerpc/mm/book3s64/subpage_prot.c 		limit = spt->maxaddr;
spt               107 arch/powerpc/mm/book3s64/subpage_prot.c 			spm = spt->low_prot;
spt               109 arch/powerpc/mm/book3s64/subpage_prot.c 			spm = spt->protptrs[addr >> SBP_L3_SHIFT];
spt               192 arch/powerpc/mm/book3s64/subpage_prot.c 	struct subpage_prot_table *spt;
spt               222 arch/powerpc/mm/book3s64/subpage_prot.c 	spt = mm_ctx_subpage_prot(&mm->context);
spt               223 arch/powerpc/mm/book3s64/subpage_prot.c 	if (!spt) {
spt               228 arch/powerpc/mm/book3s64/subpage_prot.c 		spt = kzalloc(sizeof(struct subpage_prot_table), GFP_KERNEL);
spt               229 arch/powerpc/mm/book3s64/subpage_prot.c 		if (!spt) {
spt               233 arch/powerpc/mm/book3s64/subpage_prot.c 		mm->context.hash_context->spt = spt;
spt               241 arch/powerpc/mm/book3s64/subpage_prot.c 			spm = spt->low_prot;
spt               243 arch/powerpc/mm/book3s64/subpage_prot.c 			spm = spt->protptrs[addr >> SBP_L3_SHIFT];
spt               248 arch/powerpc/mm/book3s64/subpage_prot.c 				spt->protptrs[addr >> SBP_L3_SHIFT] = spm;
spt               279 arch/powerpc/mm/book3s64/subpage_prot.c 	if (limit > spt->maxaddr)
spt               280 arch/powerpc/mm/book3s64/subpage_prot.c 		spt->maxaddr = limit;
spt               329 arch/x86/include/asm/kvm_host.h 	u64 *spt;
spt              1503 arch/x86/kvm/mmu.c 	kvm_mmu_page_set_gfn(sp, spte - sp->spt, gfn);
spt              1515 arch/x86/kvm/mmu.c 	gfn = kvm_mmu_page_get_gfn(sp, spte - sp->spt);
spt              2113 arch/x86/kvm/mmu.c static int is_empty_shadow_page(u64 *spt)
spt              2118 arch/x86/kvm/mmu.c 	for (pos = spt, end = pos + PAGE_SIZE / sizeof(u64); pos != end; pos++)
spt              2142 arch/x86/kvm/mmu.c 	MMU_WARN_ON(!is_empty_shadow_page(sp->spt));
spt              2145 arch/x86/kvm/mmu.c 	free_page((unsigned long)sp->spt);
spt              2183 arch/x86/kvm/mmu.c 	sp->spt = mmu_memory_cache_alloc(&vcpu->arch.mmu_page_cache);
spt              2186 arch/x86/kvm/mmu.c 	set_page_private(virt_to_page(sp->spt), (unsigned long)sp);
spt              2216 arch/x86/kvm/mmu.c 	index = spte - sp->spt;
spt              2281 arch/x86/kvm/mmu.c 		u64 ent = sp->spt[i];
spt              2630 arch/x86/kvm/mmu.c 	clear_page(sp->spt);
spt              2710 arch/x86/kvm/mmu.c 	spte = __pa(sp->spt) | shadow_present_mask | PT_WRITABLE_MASK |
spt              2779 arch/x86/kvm/mmu.c 		mmu_page_zap_pte(kvm, sp, sp->spt + i);
spt              3241 arch/x86/kvm/mmu.c 	gfn = kvm_mmu_page_get_gfn(sp, start - sp->spt);
spt              3267 arch/x86/kvm/mmu.c 	i = (sptep - sp->spt) & ~(PTE_PREFETCH_NUM - 1);
spt              3268 arch/x86/kvm/mmu.c 	spte = sp->spt + i;
spt              3512 arch/x86/kvm/mmu.c 		gfn = kvm_mmu_page_get_gfn(sp, sptep - sp->spt);
spt              3792 arch/x86/kvm/mmu.c 		vcpu->arch.mmu->root_hpa = __pa(sp->spt);
spt              3805 arch/x86/kvm/mmu.c 			root = __pa(sp->spt);
spt              3847 arch/x86/kvm/mmu.c 		root = __pa(sp->spt);
spt              3884 arch/x86/kvm/mmu.c 		root = __pa(sp->spt);
spt              5427 arch/x86/kvm/mmu.c 	spte = &sp->spt[page_offset / sizeof(*spte)];
spt                40 arch/x86/kvm/mmu_audit.c 		u64 *ent = sp->spt;
spt               113 arch/x86/kvm/mmu_audit.c 	gfn = kvm_mmu_page_get_gfn(sp, sptep - sp->spt);
spt               136 arch/x86/kvm/mmu_audit.c 	gfn = kvm_mmu_page_get_gfn(rev_sp, sptep - rev_sp->spt);
spt               145 arch/x86/kvm/mmu_audit.c 		       (long int)(sptep - rev_sp->spt), rev_sp->gfn);
spt               183 arch/x86/kvm/mmu_audit.c 		if (!is_shadow_present_pte(sp->spt[i]))
spt               186 arch/x86/kvm/mmu_audit.c 		inspect_spte_has_rmap(kvm, sp->spt + i);
spt               594 arch/x86/kvm/paging_tmpl.h 	i = (sptep - sp->spt) & ~(PTE_PREFETCH_NUM - 1);
spt               595 arch/x86/kvm/paging_tmpl.h 	spte = sp->spt + i;
spt               926 arch/x86/kvm/paging_tmpl.h 			pte_gpa += (sptep - sp->spt) * sizeof(pt_element_t);
spt              1025 arch/x86/kvm/paging_tmpl.h 		if (!sp->spt[i])
spt              1034 arch/x86/kvm/paging_tmpl.h 		if (FNAME(prefetch_invalid_gpte)(vcpu, sp, &sp->spt[i], gpte)) {
spt              1050 arch/x86/kvm/paging_tmpl.h 		if (sync_mmio_spte(vcpu, &sp->spt[i], gfn, pte_access,
spt              1055 arch/x86/kvm/paging_tmpl.h 			drop_spte(vcpu->kvm, &sp->spt[i]);
spt              1067 arch/x86/kvm/paging_tmpl.h 		host_writable = sp->spt[i] & SPTE_HOST_WRITEABLE;
spt              1069 arch/x86/kvm/paging_tmpl.h 		set_spte_ret |= set_spte(vcpu, &sp->spt[i],
spt              1071 arch/x86/kvm/paging_tmpl.h 					 gfn, spte_to_pfn(sp->spt[i]),
spt               170 drivers/block/ataflop.c 	unsigned	spt;		/* sectors per track */
spt               750 drivers/block/ataflop.c 	if (!UDT || desc->track >= UDT->blocks/UDT->spt/2 || desc->head >= 2) {
spt               755 drivers/block/ataflop.c 	nsect = UDT->spt;
spt              1018 drivers/block/ataflop.c 	dma_wd.fdc_acces_seccount = read_track ? SUDT->spt : 1;
spt              1083 drivers/block/ataflop.c 	if (addr >= PhysTrackBuffer + SUDT->spt*512) {
spt              1151 drivers/block/ataflop.c 	    !(read_track && FDC_READ(FDCREG_SECTOR) > SUDT->spt)) {
spt              1460 drivers/block/ataflop.c 	ReqTrack = block / UDT->spt;
spt              1461 drivers/block/ataflop.c 	ReqSector = block - ReqTrack * UDT->spt + 1;
spt              1594 drivers/block/ataflop.c 		getprm.sect = dtp->spt;
spt              1596 drivers/block/ataflop.c 		getprm.track = dtp->blocks/dtp->spt/2;
spt              1655 drivers/block/ataflop.c 			    && dtp->spt     == setprm.sect
spt              1666 drivers/block/ataflop.c 				  default_params[drive].spt     = dtp->spt;
spt              1688 drivers/block/ataflop.c 		dtp->spt    = setprm.sect;
spt              1697 drivers/block/ataflop.c 				drive, dtp->blocks, dtp->spt, dtp->stretch);
spt              1700 drivers/block/ataflop.c 		if (setprm.track != dtp->blocks/dtp->spt/2 ||
spt               646 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_vgpu_ppgtt_spt *spt,
spt               651 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_gvt *gvt = spt->vgpu->gvt;
spt               661 drivers/gpu/drm/i915/gvt/gtt.c 			spt->guest_page.gfn << I915_GTT_PAGE_SHIFT,
spt               662 drivers/gpu/drm/i915/gvt/gtt.c 			spt->vgpu);
spt               667 drivers/gpu/drm/i915/gvt/gtt.c 				   spt->guest_page.pde_ips : false);
spt               675 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_vgpu_ppgtt_spt *spt,
spt               680 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_gvt *gvt = spt->vgpu->gvt;
spt               690 drivers/gpu/drm/i915/gvt/gtt.c 			spt->guest_page.gfn << I915_GTT_PAGE_SHIFT,
spt               691 drivers/gpu/drm/i915/gvt/gtt.c 			spt->vgpu);
spt               694 drivers/gpu/drm/i915/gvt/gtt.c #define ppgtt_get_guest_entry(spt, e, index) \
spt               695 drivers/gpu/drm/i915/gvt/gtt.c 	ppgtt_spt_get_entry(spt, NULL, \
spt               696 drivers/gpu/drm/i915/gvt/gtt.c 		spt->guest_page.type, e, index, true)
spt               698 drivers/gpu/drm/i915/gvt/gtt.c #define ppgtt_set_guest_entry(spt, e, index) \
spt               699 drivers/gpu/drm/i915/gvt/gtt.c 	ppgtt_spt_set_entry(spt, NULL, \
spt               700 drivers/gpu/drm/i915/gvt/gtt.c 		spt->guest_page.type, e, index, true)
spt               702 drivers/gpu/drm/i915/gvt/gtt.c #define ppgtt_get_shadow_entry(spt, e, index) \
spt               703 drivers/gpu/drm/i915/gvt/gtt.c 	ppgtt_spt_get_entry(spt, spt->shadow_page.vaddr, \
spt               704 drivers/gpu/drm/i915/gvt/gtt.c 		spt->shadow_page.type, e, index, false)
spt               706 drivers/gpu/drm/i915/gvt/gtt.c #define ppgtt_set_shadow_entry(spt, e, index) \
spt               707 drivers/gpu/drm/i915/gvt/gtt.c 	ppgtt_spt_set_entry(spt, spt->shadow_page.vaddr, \
spt               708 drivers/gpu/drm/i915/gvt/gtt.c 		spt->shadow_page.type, e, index, false)
spt               712 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_ppgtt_spt *spt;
spt               714 drivers/gpu/drm/i915/gvt/gtt.c 	spt = kzalloc(sizeof(*spt), gfp_mask);
spt               715 drivers/gpu/drm/i915/gvt/gtt.c 	if (!spt)
spt               718 drivers/gpu/drm/i915/gvt/gtt.c 	spt->shadow_page.page = alloc_page(gfp_mask);
spt               719 drivers/gpu/drm/i915/gvt/gtt.c 	if (!spt->shadow_page.page) {
spt               720 drivers/gpu/drm/i915/gvt/gtt.c 		kfree(spt);
spt               723 drivers/gpu/drm/i915/gvt/gtt.c 	return spt;
spt               726 drivers/gpu/drm/i915/gvt/gtt.c static void free_spt(struct intel_vgpu_ppgtt_spt *spt)
spt               728 drivers/gpu/drm/i915/gvt/gtt.c 	__free_page(spt->shadow_page.page);
spt               729 drivers/gpu/drm/i915/gvt/gtt.c 	kfree(spt);
spt               735 drivers/gpu/drm/i915/gvt/gtt.c static void ppgtt_free_spt(struct intel_vgpu_ppgtt_spt *spt)
spt               737 drivers/gpu/drm/i915/gvt/gtt.c 	struct device *kdev = &spt->vgpu->gvt->dev_priv->drm.pdev->dev;
spt               739 drivers/gpu/drm/i915/gvt/gtt.c 	trace_spt_free(spt->vgpu->id, spt, spt->guest_page.type);
spt               741 drivers/gpu/drm/i915/gvt/gtt.c 	dma_unmap_page(kdev, spt->shadow_page.mfn << I915_GTT_PAGE_SHIFT, 4096,
spt               744 drivers/gpu/drm/i915/gvt/gtt.c 	radix_tree_delete(&spt->vgpu->gtt.spt_tree, spt->shadow_page.mfn);
spt               746 drivers/gpu/drm/i915/gvt/gtt.c 	if (spt->guest_page.gfn) {
spt               747 drivers/gpu/drm/i915/gvt/gtt.c 		if (spt->guest_page.oos_page)
spt               748 drivers/gpu/drm/i915/gvt/gtt.c 			detach_oos_page(spt->vgpu, spt->guest_page.oos_page);
spt               750 drivers/gpu/drm/i915/gvt/gtt.c 		intel_vgpu_unregister_page_track(spt->vgpu, spt->guest_page.gfn);
spt               753 drivers/gpu/drm/i915/gvt/gtt.c 	list_del_init(&spt->post_shadow_list);
spt               754 drivers/gpu/drm/i915/gvt/gtt.c 	free_spt(spt);
spt               759 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_ppgtt_spt *spt, *spn;
spt               766 drivers/gpu/drm/i915/gvt/gtt.c 		spt = radix_tree_deref_slot(slot);
spt               767 drivers/gpu/drm/i915/gvt/gtt.c 		list_move(&spt->post_shadow_list, &all_spt);
spt               771 drivers/gpu/drm/i915/gvt/gtt.c 	list_for_each_entry_safe(spt, spn, &all_spt, post_shadow_list)
spt               772 drivers/gpu/drm/i915/gvt/gtt.c 		ppgtt_free_spt(spt);
spt               776 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_vgpu_ppgtt_spt *spt,
spt               783 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_ppgtt_spt *spt = page_track->priv_data;
spt               790 drivers/gpu/drm/i915/gvt/gtt.c 	ret = ppgtt_handle_guest_write_page_table_bytes(spt, gpa, data, bytes);
spt               823 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_ppgtt_spt *spt = NULL;
spt               828 drivers/gpu/drm/i915/gvt/gtt.c 	spt = alloc_spt(GFP_KERNEL | __GFP_ZERO);
spt               829 drivers/gpu/drm/i915/gvt/gtt.c 	if (!spt) {
spt               837 drivers/gpu/drm/i915/gvt/gtt.c 	spt->vgpu = vgpu;
spt               838 drivers/gpu/drm/i915/gvt/gtt.c 	atomic_set(&spt->refcount, 1);
spt               839 drivers/gpu/drm/i915/gvt/gtt.c 	INIT_LIST_HEAD(&spt->post_shadow_list);
spt               844 drivers/gpu/drm/i915/gvt/gtt.c 	spt->shadow_page.type = type;
spt               845 drivers/gpu/drm/i915/gvt/gtt.c 	daddr = dma_map_page(kdev, spt->shadow_page.page,
spt               852 drivers/gpu/drm/i915/gvt/gtt.c 	spt->shadow_page.vaddr = page_address(spt->shadow_page.page);
spt               853 drivers/gpu/drm/i915/gvt/gtt.c 	spt->shadow_page.mfn = daddr >> I915_GTT_PAGE_SHIFT;
spt               855 drivers/gpu/drm/i915/gvt/gtt.c 	ret = radix_tree_insert(&vgpu->gtt.spt_tree, spt->shadow_page.mfn, spt);
spt               859 drivers/gpu/drm/i915/gvt/gtt.c 	return spt;
spt               864 drivers/gpu/drm/i915/gvt/gtt.c 	free_spt(spt);
spt               873 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_ppgtt_spt *spt;
spt               876 drivers/gpu/drm/i915/gvt/gtt.c 	spt = ppgtt_alloc_spt(vgpu, type);
spt               877 drivers/gpu/drm/i915/gvt/gtt.c 	if (IS_ERR(spt))
spt               878 drivers/gpu/drm/i915/gvt/gtt.c 		return spt;
spt               884 drivers/gpu/drm/i915/gvt/gtt.c 			ppgtt_write_protection_handler, spt);
spt               886 drivers/gpu/drm/i915/gvt/gtt.c 		ppgtt_free_spt(spt);
spt               890 drivers/gpu/drm/i915/gvt/gtt.c 	spt->guest_page.type = type;
spt               891 drivers/gpu/drm/i915/gvt/gtt.c 	spt->guest_page.gfn = gfn;
spt               892 drivers/gpu/drm/i915/gvt/gtt.c 	spt->guest_page.pde_ips = guest_pde_ips;
spt               894 drivers/gpu/drm/i915/gvt/gtt.c 	trace_spt_alloc(vgpu->id, spt, type, spt->shadow_page.mfn, gfn);
spt               896 drivers/gpu/drm/i915/gvt/gtt.c 	return spt;
spt               899 drivers/gpu/drm/i915/gvt/gtt.c #define pt_entry_size_shift(spt) \
spt               900 drivers/gpu/drm/i915/gvt/gtt.c 	((spt)->vgpu->gvt->device_info.gtt_entry_size_shift)
spt               902 drivers/gpu/drm/i915/gvt/gtt.c #define pt_entries(spt) \
spt               903 drivers/gpu/drm/i915/gvt/gtt.c 	(I915_GTT_PAGE_SIZE >> pt_entry_size_shift(spt))
spt               905 drivers/gpu/drm/i915/gvt/gtt.c #define for_each_present_guest_entry(spt, e, i) \
spt               906 drivers/gpu/drm/i915/gvt/gtt.c 	for (i = 0; i < pt_entries(spt); \
spt               907 drivers/gpu/drm/i915/gvt/gtt.c 	     i += spt->guest_page.pde_ips ? GTT_64K_PTE_STRIDE : 1) \
spt               908 drivers/gpu/drm/i915/gvt/gtt.c 		if (!ppgtt_get_guest_entry(spt, e, i) && \
spt               909 drivers/gpu/drm/i915/gvt/gtt.c 		    spt->vgpu->gvt->gtt.pte_ops->test_present(e))
spt               911 drivers/gpu/drm/i915/gvt/gtt.c #define for_each_present_shadow_entry(spt, e, i) \
spt               912 drivers/gpu/drm/i915/gvt/gtt.c 	for (i = 0; i < pt_entries(spt); \
spt               913 drivers/gpu/drm/i915/gvt/gtt.c 	     i += spt->shadow_page.pde_ips ? GTT_64K_PTE_STRIDE : 1) \
spt               914 drivers/gpu/drm/i915/gvt/gtt.c 		if (!ppgtt_get_shadow_entry(spt, e, i) && \
spt               915 drivers/gpu/drm/i915/gvt/gtt.c 		    spt->vgpu->gvt->gtt.pte_ops->test_present(e))
spt               917 drivers/gpu/drm/i915/gvt/gtt.c #define for_each_shadow_entry(spt, e, i) \
spt               918 drivers/gpu/drm/i915/gvt/gtt.c 	for (i = 0; i < pt_entries(spt); \
spt               919 drivers/gpu/drm/i915/gvt/gtt.c 	     i += (spt->shadow_page.pde_ips ? GTT_64K_PTE_STRIDE : 1)) \
spt               920 drivers/gpu/drm/i915/gvt/gtt.c 		if (!ppgtt_get_shadow_entry(spt, e, i))
spt               922 drivers/gpu/drm/i915/gvt/gtt.c static inline void ppgtt_get_spt(struct intel_vgpu_ppgtt_spt *spt)
spt               924 drivers/gpu/drm/i915/gvt/gtt.c 	int v = atomic_read(&spt->refcount);
spt               926 drivers/gpu/drm/i915/gvt/gtt.c 	trace_spt_refcount(spt->vgpu->id, "inc", spt, v, (v + 1));
spt               927 drivers/gpu/drm/i915/gvt/gtt.c 	atomic_inc(&spt->refcount);
spt               930 drivers/gpu/drm/i915/gvt/gtt.c static inline int ppgtt_put_spt(struct intel_vgpu_ppgtt_spt *spt)
spt               932 drivers/gpu/drm/i915/gvt/gtt.c 	int v = atomic_read(&spt->refcount);
spt               934 drivers/gpu/drm/i915/gvt/gtt.c 	trace_spt_refcount(spt->vgpu->id, "dec", spt, v, (v - 1));
spt               935 drivers/gpu/drm/i915/gvt/gtt.c 	return atomic_dec_return(&spt->refcount);
spt               938 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_invalidate_spt(struct intel_vgpu_ppgtt_spt *spt);
spt               974 drivers/gpu/drm/i915/gvt/gtt.c static inline void ppgtt_invalidate_pte(struct intel_vgpu_ppgtt_spt *spt,
spt               977 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu *vgpu = spt->vgpu;
spt               983 drivers/gpu/drm/i915/gvt/gtt.c 	type = spt->shadow_page.type;
spt               992 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_invalidate_spt(struct intel_vgpu_ppgtt_spt *spt)
spt               994 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu *vgpu = spt->vgpu;
spt               999 drivers/gpu/drm/i915/gvt/gtt.c 	trace_spt_change(spt->vgpu->id, "die", spt,
spt              1000 drivers/gpu/drm/i915/gvt/gtt.c 			spt->guest_page.gfn, spt->shadow_page.type);
spt              1002 drivers/gpu/drm/i915/gvt/gtt.c 	if (ppgtt_put_spt(spt) > 0)
spt              1005 drivers/gpu/drm/i915/gvt/gtt.c 	for_each_present_shadow_entry(spt, &e, index) {
spt              1009 drivers/gpu/drm/i915/gvt/gtt.c 			ppgtt_invalidate_pte(spt, &e);
spt              1026 drivers/gpu/drm/i915/gvt/gtt.c 					spt->vgpu, &e);
spt              1035 drivers/gpu/drm/i915/gvt/gtt.c 	trace_spt_change(spt->vgpu->id, "release", spt,
spt              1036 drivers/gpu/drm/i915/gvt/gtt.c 			 spt->guest_page.gfn, spt->shadow_page.type);
spt              1037 drivers/gpu/drm/i915/gvt/gtt.c 	ppgtt_free_spt(spt);
spt              1041 drivers/gpu/drm/i915/gvt/gtt.c 			spt, e.val64, e.type);
spt              1061 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_populate_spt(struct intel_vgpu_ppgtt_spt *spt);
spt              1067 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_ppgtt_spt *spt = NULL;
spt              1076 drivers/gpu/drm/i915/gvt/gtt.c 	spt = intel_vgpu_find_spt_by_gfn(vgpu, ops->get_pfn(we));
spt              1077 drivers/gpu/drm/i915/gvt/gtt.c 	if (spt) {
spt              1078 drivers/gpu/drm/i915/gvt/gtt.c 		ppgtt_get_spt(spt);
spt              1080 drivers/gpu/drm/i915/gvt/gtt.c 		if (ips != spt->guest_page.pde_ips) {
spt              1081 drivers/gpu/drm/i915/gvt/gtt.c 			spt->guest_page.pde_ips = ips;
spt              1084 drivers/gpu/drm/i915/gvt/gtt.c 			clear_page(spt->shadow_page.vaddr);
spt              1085 drivers/gpu/drm/i915/gvt/gtt.c 			ret = ppgtt_populate_spt(spt);
spt              1087 drivers/gpu/drm/i915/gvt/gtt.c 				ppgtt_put_spt(spt);
spt              1099 drivers/gpu/drm/i915/gvt/gtt.c 		spt = ppgtt_alloc_spt_gfn(vgpu, type, ops->get_pfn(we), ips);
spt              1100 drivers/gpu/drm/i915/gvt/gtt.c 		if (IS_ERR(spt)) {
spt              1101 drivers/gpu/drm/i915/gvt/gtt.c 			ret = PTR_ERR(spt);
spt              1105 drivers/gpu/drm/i915/gvt/gtt.c 		ret = intel_vgpu_enable_page_track(vgpu, spt->guest_page.gfn);
spt              1109 drivers/gpu/drm/i915/gvt/gtt.c 		ret = ppgtt_populate_spt(spt);
spt              1113 drivers/gpu/drm/i915/gvt/gtt.c 		trace_spt_change(vgpu->id, "new", spt, spt->guest_page.gfn,
spt              1114 drivers/gpu/drm/i915/gvt/gtt.c 				 spt->shadow_page.type);
spt              1116 drivers/gpu/drm/i915/gvt/gtt.c 	return spt;
spt              1119 drivers/gpu/drm/i915/gvt/gtt.c 	ppgtt_free_spt(spt);
spt              1120 drivers/gpu/drm/i915/gvt/gtt.c 	spt = NULL;
spt              1123 drivers/gpu/drm/i915/gvt/gtt.c 		     spt, we->val64, we->type);
spt              1167 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_ppgtt_spt *spt, unsigned long index,
spt              1190 drivers/gpu/drm/i915/gvt/gtt.c 			ppgtt_invalidate_spt(spt);
spt              1209 drivers/gpu/drm/i915/gvt/gtt.c 	ppgtt_set_shadow_entry(spt, se, index);
spt              1214 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_ppgtt_spt *spt, unsigned long index,
spt              1239 drivers/gpu/drm/i915/gvt/gtt.c 		ppgtt_set_shadow_entry(spt, &entry, index + i);
spt              1245 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_ppgtt_spt *spt, unsigned long index,
spt              1270 drivers/gpu/drm/i915/gvt/gtt.c 		return split_64KB_gtt_entry(vgpu, spt, index, &se);
spt              1275 drivers/gpu/drm/i915/gvt/gtt.c 			return split_2MB_gtt_entry(vgpu, spt, index, &se);
spt              1294 drivers/gpu/drm/i915/gvt/gtt.c 	ppgtt_set_shadow_entry(spt, &se, index);
spt              1298 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_populate_spt(struct intel_vgpu_ppgtt_spt *spt)
spt              1300 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu *vgpu = spt->vgpu;
spt              1308 drivers/gpu/drm/i915/gvt/gtt.c 	trace_spt_change(spt->vgpu->id, "born", spt,
spt              1309 drivers/gpu/drm/i915/gvt/gtt.c 			 spt->guest_page.gfn, spt->shadow_page.type);
spt              1311 drivers/gpu/drm/i915/gvt/gtt.c 	for_each_present_guest_entry(spt, &ge, i) {
spt              1318 drivers/gpu/drm/i915/gvt/gtt.c 			ppgtt_get_shadow_entry(spt, &se, i);
spt              1320 drivers/gpu/drm/i915/gvt/gtt.c 			ppgtt_set_shadow_entry(spt, &se, i);
spt              1325 drivers/gpu/drm/i915/gvt/gtt.c 				ppgtt_set_shadow_entry(spt, &se, i);
spt              1329 drivers/gpu/drm/i915/gvt/gtt.c 			ret = ppgtt_populate_shadow_entry(vgpu, spt, i, &ge);
spt              1337 drivers/gpu/drm/i915/gvt/gtt.c 			spt, ge.val64, ge.type);
spt              1341 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_handle_guest_entry_removal(struct intel_vgpu_ppgtt_spt *spt,
spt              1344 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu *vgpu = spt->vgpu;
spt              1348 drivers/gpu/drm/i915/gvt/gtt.c 	trace_spt_guest_change(spt->vgpu->id, "remove", spt,
spt              1349 drivers/gpu/drm/i915/gvt/gtt.c 			       spt->shadow_page.type, se->val64, index);
spt              1358 drivers/gpu/drm/i915/gvt/gtt.c 	    vgpu->gtt.scratch_pt[spt->shadow_page.type].page_mfn)
spt              1376 drivers/gpu/drm/i915/gvt/gtt.c 		ppgtt_invalidate_pte(spt, se);
spt              1382 drivers/gpu/drm/i915/gvt/gtt.c 			spt, se->val64, se->type);
spt              1386 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_handle_guest_entry_add(struct intel_vgpu_ppgtt_spt *spt,
spt              1389 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu *vgpu = spt->vgpu;
spt              1394 drivers/gpu/drm/i915/gvt/gtt.c 	trace_spt_guest_change(spt->vgpu->id, "add", spt, spt->shadow_page.type,
spt              1406 drivers/gpu/drm/i915/gvt/gtt.c 		ppgtt_get_shadow_entry(spt, &m, index);
spt              1408 drivers/gpu/drm/i915/gvt/gtt.c 		ppgtt_set_shadow_entry(spt, &m, index);
spt              1410 drivers/gpu/drm/i915/gvt/gtt.c 		ret = ppgtt_populate_shadow_entry(vgpu, spt, index, we);
spt              1417 drivers/gpu/drm/i915/gvt/gtt.c 		spt, we->val64, we->type);
spt              1427 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_ppgtt_spt *spt = oos_page->spt;
spt              1433 drivers/gpu/drm/i915/gvt/gtt.c 			 spt, spt->guest_page.type);
spt              1435 drivers/gpu/drm/i915/gvt/gtt.c 	old.type = new.type = get_entry_type(spt->guest_page.type);
spt              1442 drivers/gpu/drm/i915/gvt/gtt.c 			       spt->guest_page.gfn << PAGE_SHIFT, vgpu);
spt              1445 drivers/gpu/drm/i915/gvt/gtt.c 			&& !test_and_clear_bit(index, spt->post_shadow_bitmap))
spt              1449 drivers/gpu/drm/i915/gvt/gtt.c 				spt, spt->guest_page.type,
spt              1452 drivers/gpu/drm/i915/gvt/gtt.c 		ret = ppgtt_populate_shadow_entry(vgpu, spt, index, &new);
spt              1459 drivers/gpu/drm/i915/gvt/gtt.c 	spt->guest_page.write_cnt = 0;
spt              1460 drivers/gpu/drm/i915/gvt/gtt.c 	list_del_init(&spt->post_shadow_list);
spt              1468 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_ppgtt_spt *spt = oos_page->spt;
spt              1471 drivers/gpu/drm/i915/gvt/gtt.c 			 spt, spt->guest_page.type);
spt              1473 drivers/gpu/drm/i915/gvt/gtt.c 	spt->guest_page.write_cnt = 0;
spt              1474 drivers/gpu/drm/i915/gvt/gtt.c 	spt->guest_page.oos_page = NULL;
spt              1475 drivers/gpu/drm/i915/gvt/gtt.c 	oos_page->spt = NULL;
spt              1484 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_vgpu_ppgtt_spt *spt)
spt              1486 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_gvt *gvt = spt->vgpu->gvt;
spt              1489 drivers/gpu/drm/i915/gvt/gtt.c 	ret = intel_gvt_hypervisor_read_gpa(spt->vgpu,
spt              1490 drivers/gpu/drm/i915/gvt/gtt.c 			spt->guest_page.gfn << I915_GTT_PAGE_SHIFT,
spt              1495 drivers/gpu/drm/i915/gvt/gtt.c 	oos_page->spt = spt;
spt              1496 drivers/gpu/drm/i915/gvt/gtt.c 	spt->guest_page.oos_page = oos_page;
spt              1500 drivers/gpu/drm/i915/gvt/gtt.c 	trace_oos_change(spt->vgpu->id, "attach", oos_page->id,
spt              1501 drivers/gpu/drm/i915/gvt/gtt.c 			 spt, spt->guest_page.type);
spt              1505 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_set_guest_page_sync(struct intel_vgpu_ppgtt_spt *spt)
spt              1507 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_oos_page *oos_page = spt->guest_page.oos_page;
spt              1510 drivers/gpu/drm/i915/gvt/gtt.c 	ret = intel_vgpu_enable_page_track(spt->vgpu, spt->guest_page.gfn);
spt              1514 drivers/gpu/drm/i915/gvt/gtt.c 	trace_oos_change(spt->vgpu->id, "set page sync", oos_page->id,
spt              1515 drivers/gpu/drm/i915/gvt/gtt.c 			 spt, spt->guest_page.type);
spt              1518 drivers/gpu/drm/i915/gvt/gtt.c 	return sync_oos_page(spt->vgpu, oos_page);
spt              1521 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_allocate_oos_page(struct intel_vgpu_ppgtt_spt *spt)
spt              1523 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_gvt *gvt = spt->vgpu->gvt;
spt              1525 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_oos_page *oos_page = spt->guest_page.oos_page;
spt              1533 drivers/gpu/drm/i915/gvt/gtt.c 		ret = ppgtt_set_guest_page_sync(oos_page->spt);
spt              1536 drivers/gpu/drm/i915/gvt/gtt.c 		ret = detach_oos_page(spt->vgpu, oos_page);
spt              1542 drivers/gpu/drm/i915/gvt/gtt.c 	return attach_oos_page(oos_page, spt);
spt              1545 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_set_guest_page_oos(struct intel_vgpu_ppgtt_spt *spt)
spt              1547 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_oos_page *oos_page = spt->guest_page.oos_page;
spt              1552 drivers/gpu/drm/i915/gvt/gtt.c 	trace_oos_change(spt->vgpu->id, "set page out of sync", oos_page->id,
spt              1553 drivers/gpu/drm/i915/gvt/gtt.c 			 spt, spt->guest_page.type);
spt              1555 drivers/gpu/drm/i915/gvt/gtt.c 	list_add_tail(&oos_page->vm_list, &spt->vgpu->gtt.oos_page_list_head);
spt              1556 drivers/gpu/drm/i915/gvt/gtt.c 	return intel_vgpu_disable_page_track(spt->vgpu, spt->guest_page.gfn);
spt              1581 drivers/gpu/drm/i915/gvt/gtt.c 		ret = ppgtt_set_guest_page_sync(oos_page->spt);
spt              1592 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_vgpu_ppgtt_spt *spt,
spt              1595 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu *vgpu = spt->vgpu;
spt              1596 drivers/gpu/drm/i915/gvt/gtt.c 	int type = spt->shadow_page.type;
spt              1609 drivers/gpu/drm/i915/gvt/gtt.c 	ppgtt_get_shadow_entry(spt, &old_se, index);
spt              1612 drivers/gpu/drm/i915/gvt/gtt.c 		ret = ppgtt_handle_guest_entry_add(spt, we, index);
spt              1617 drivers/gpu/drm/i915/gvt/gtt.c 	ret = ppgtt_handle_guest_entry_removal(spt, &old_se, index);
spt              1630 drivers/gpu/drm/i915/gvt/gtt.c 				ppgtt_set_shadow_entry(spt, &old_se, index + i);
spt              1637 drivers/gpu/drm/i915/gvt/gtt.c 			ppgtt_set_shadow_entry(spt, &old_se, index);
spt              1641 drivers/gpu/drm/i915/gvt/gtt.c 			ppgtt_set_shadow_entry(spt, &old_se, index);
spt              1648 drivers/gpu/drm/i915/gvt/gtt.c 			spt, we->val64, we->type);
spt              1654 drivers/gpu/drm/i915/gvt/gtt.c static inline bool can_do_out_of_sync(struct intel_vgpu_ppgtt_spt *spt)
spt              1657 drivers/gpu/drm/i915/gvt/gtt.c 		&& gtt_type_is_pte_pt(spt->guest_page.type)
spt              1658 drivers/gpu/drm/i915/gvt/gtt.c 		&& spt->guest_page.write_cnt >= 2;
spt              1661 drivers/gpu/drm/i915/gvt/gtt.c static void ppgtt_set_post_shadow(struct intel_vgpu_ppgtt_spt *spt,
spt              1664 drivers/gpu/drm/i915/gvt/gtt.c 	set_bit(index, spt->post_shadow_bitmap);
spt              1665 drivers/gpu/drm/i915/gvt/gtt.c 	if (!list_empty(&spt->post_shadow_list))
spt              1668 drivers/gpu/drm/i915/gvt/gtt.c 	list_add_tail(&spt->post_shadow_list,
spt              1669 drivers/gpu/drm/i915/gvt/gtt.c 			&spt->vgpu->gtt.post_shadow_list_head);
spt              1685 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_ppgtt_spt *spt;
spt              1691 drivers/gpu/drm/i915/gvt/gtt.c 		spt = container_of(pos, struct intel_vgpu_ppgtt_spt,
spt              1694 drivers/gpu/drm/i915/gvt/gtt.c 		for_each_set_bit(index, spt->post_shadow_bitmap,
spt              1696 drivers/gpu/drm/i915/gvt/gtt.c 			ppgtt_get_guest_entry(spt, &ge, index);
spt              1698 drivers/gpu/drm/i915/gvt/gtt.c 			ret = ppgtt_handle_guest_write_page_table(spt,
spt              1702 drivers/gpu/drm/i915/gvt/gtt.c 			clear_bit(index, spt->post_shadow_bitmap);
spt              1704 drivers/gpu/drm/i915/gvt/gtt.c 		list_del_init(&spt->post_shadow_list);
spt              1710 drivers/gpu/drm/i915/gvt/gtt.c 		struct intel_vgpu_ppgtt_spt *spt,
spt              1713 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu *vgpu = spt->vgpu;
spt              1722 drivers/gpu/drm/i915/gvt/gtt.c 	ppgtt_get_guest_entry(spt, &we, index);
spt              1737 drivers/gpu/drm/i915/gvt/gtt.c 		ret = ppgtt_handle_guest_write_page_table(spt, &we, index);
spt              1741 drivers/gpu/drm/i915/gvt/gtt.c 		if (!test_bit(index, spt->post_shadow_bitmap)) {
spt              1742 drivers/gpu/drm/i915/gvt/gtt.c 			int type = spt->shadow_page.type;
spt              1744 drivers/gpu/drm/i915/gvt/gtt.c 			ppgtt_get_shadow_entry(spt, &se, index);
spt              1745 drivers/gpu/drm/i915/gvt/gtt.c 			ret = ppgtt_handle_guest_entry_removal(spt, &se, index);
spt              1749 drivers/gpu/drm/i915/gvt/gtt.c 			ppgtt_set_shadow_entry(spt, &se, index);
spt              1751 drivers/gpu/drm/i915/gvt/gtt.c 		ppgtt_set_post_shadow(spt, index);
spt              1757 drivers/gpu/drm/i915/gvt/gtt.c 	spt->guest_page.write_cnt++;
spt              1759 drivers/gpu/drm/i915/gvt/gtt.c 	if (spt->guest_page.oos_page)
spt              1760 drivers/gpu/drm/i915/gvt/gtt.c 		ops->set_entry(spt->guest_page.oos_page->mem, &we, index,
spt              1763 drivers/gpu/drm/i915/gvt/gtt.c 	if (can_do_out_of_sync(spt)) {
spt              1764 drivers/gpu/drm/i915/gvt/gtt.c 		if (!spt->guest_page.oos_page)
spt              1765 drivers/gpu/drm/i915/gvt/gtt.c 			ppgtt_allocate_oos_page(spt);
spt              1767 drivers/gpu/drm/i915/gvt/gtt.c 		ret = ppgtt_set_guest_page_oos(spt);
spt              1810 drivers/gpu/drm/i915/gvt/gtt.c 	struct intel_vgpu_ppgtt_spt *spt;
spt              1828 drivers/gpu/drm/i915/gvt/gtt.c 		spt = ppgtt_populate_spt_by_guest_entry(vgpu, &ge);
spt              1829 drivers/gpu/drm/i915/gvt/gtt.c 		if (IS_ERR(spt)) {
spt              1831 drivers/gpu/drm/i915/gvt/gtt.c 			ret = PTR_ERR(spt);
spt              1834 drivers/gpu/drm/i915/gvt/gtt.c 		ppgtt_generate_shadow_entry(&se, spt, &ge);
spt               222 drivers/gpu/drm/i915/gvt/gtt.h 	struct intel_vgpu_ppgtt_spt *spt;
spt                43 drivers/gpu/drm/i915/gvt/trace.h 	TP_PROTO(int id, void *spt, int type, unsigned long mfn,
spt                46 drivers/gpu/drm/i915/gvt/trace.h 	TP_ARGS(id, spt, type, mfn, gpt_gfn),
spt                50 drivers/gpu/drm/i915/gvt/trace.h 		__field(void *, spt)
spt                58 drivers/gpu/drm/i915/gvt/trace.h 		__entry->spt = spt;
spt                66 drivers/gpu/drm/i915/gvt/trace.h 		__entry->spt,
spt                73 drivers/gpu/drm/i915/gvt/trace.h 	TP_PROTO(int id, void *spt, int type),
spt                75 drivers/gpu/drm/i915/gvt/trace.h 	TP_ARGS(id, spt, type),
spt                79 drivers/gpu/drm/i915/gvt/trace.h 		__field(void *, spt)
spt                85 drivers/gpu/drm/i915/gvt/trace.h 		__entry->spt = spt;
spt                91 drivers/gpu/drm/i915/gvt/trace.h 		__entry->spt,
spt               135 drivers/gpu/drm/i915/gvt/trace.h 	TP_PROTO(int id, char *action, void *spt, int before, int after),
spt               137 drivers/gpu/drm/i915/gvt/trace.h 	TP_ARGS(id, action, spt, before, after),
spt               146 drivers/gpu/drm/i915/gvt/trace.h 				id, action, spt, before, after);
spt               153 drivers/gpu/drm/i915/gvt/trace.h 	TP_PROTO(int id, char *action, void *spt, unsigned long gfn,
spt               156 drivers/gpu/drm/i915/gvt/trace.h 	TP_ARGS(id, action, spt, gfn, type),
spt               165 drivers/gpu/drm/i915/gvt/trace.h 				id, action, spt, gfn, type);
spt               172 drivers/gpu/drm/i915/gvt/trace.h 	TP_PROTO(int id, const char *tag, void *spt, int type, u64 v,
spt               175 drivers/gpu/drm/i915/gvt/trace.h 	TP_ARGS(id, tag, spt, type, v, index),
spt               184 drivers/gpu/drm/i915/gvt/trace.h 			id, tag, spt, type, v, index);
spt                66 drivers/misc/mic/host/mic_smpt.c static void mic_add_smpt_entry(int spt, s64 *ref, u64 addr,
spt                72 drivers/misc/mic/host/mic_smpt.c 	for (i = spt; i < spt + entries; i++,
spt                79 drivers/misc/mic/host/mic_smpt.c 		smpt_info->entry[i].ref_count += ref[i - spt];
spt                90 drivers/misc/mic/host/mic_smpt.c 	int spt;
spt               124 drivers/misc/mic/host/mic_smpt.c 	spt = i - entries + 1;
spt               125 drivers/misc/mic/host/mic_smpt.c 	mic_addr = mic_smpt_to_pa(mdev, spt);
spt               126 drivers/misc/mic/host/mic_smpt.c 	mic_add_smpt_entry(spt, ref, dma_addr, entries, mdev);
spt               168 drivers/misc/mic/host/mic_smpt.c 	int spt;
spt               176 drivers/misc/mic/host/mic_smpt.c 	spt = mic_sys_addr_to_smpt(mdev, mic_addr);
spt               177 drivers/misc/mic/host/mic_smpt.c 	dma_addr = smpt_info->entry[spt].dma_addr +
spt               247 drivers/misc/mic/host/mic_smpt.c 	int spt;
spt               260 drivers/misc/mic/host/mic_smpt.c 	spt = mic_sys_addr_to_smpt(mdev, mic_addr);
spt               272 drivers/misc/mic/host/mic_smpt.c 	for (i = spt; i < spt + num_smpt; i++) {
spt               273 drivers/misc/mic/host/mic_smpt.c 		smpt_info->entry[i].ref_count -= ref[i - spt];
spt               691 drivers/scsi/esas2r/esas2r_ioctl.c 	struct atto_hba_scsi_pass_thru *spt = &hi->data.scsi_pass_thru;
spt               694 drivers/scsi/esas2r/esas2r_ioctl.c 	spt->scsi_status = rq->func_rsp.scsi_rsp.scsi_stat;
spt               695 drivers/scsi/esas2r/esas2r_ioctl.c 	spt->sense_length = rq->sense_len;
spt               696 drivers/scsi/esas2r/esas2r_ioctl.c 	spt->residual_length =
spt               734 drivers/scsi/esas2r/esas2r_ioctl.c 	spt->req_status = sts;
spt               737 drivers/scsi/esas2r/esas2r_ioctl.c 	spt->target_id =
spt               738 drivers/scsi/esas2r/esas2r_ioctl.c 		esas2r_targ_db_find_next_present(a, (u16)spt->target_id);
spt               982 drivers/scsi/esas2r/esas2r_ioctl.c 		struct atto_hba_scsi_pass_thru *spt = &hi->data.scsi_pass_thru;
spt               985 drivers/scsi/esas2r/esas2r_ioctl.c 		memcpy(&lun, spt->lun, sizeof(struct scsi_lun));
spt              1000 drivers/scsi/esas2r/esas2r_ioctl.c 		if (spt->target_id >= ESAS2R_MAX_TARGETS || !check_lun(lun)) {
spt              1012 drivers/scsi/esas2r/esas2r_ioctl.c 		rq->target_id = (u16)spt->target_id;
spt              1013 drivers/scsi/esas2r/esas2r_ioctl.c 		rq->vrq->scsi.flags |= cpu_to_le32(spt->lun[1]);
spt              1014 drivers/scsi/esas2r/esas2r_ioctl.c 		memcpy(rq->vrq->scsi.cdb, spt->cdb, 16);
spt              1016 drivers/scsi/esas2r/esas2r_ioctl.c 		rq->sense_len = spt->sense_length;
spt              1017 drivers/scsi/esas2r/esas2r_ioctl.c 		rq->sense_buf = (u8 *)spt->sense_data;
spt              1029 drivers/scsi/esas2r/esas2r_ioctl.c 		if (spt->flags & ATTO_SPTF_DATA_IN) {
spt              1031 drivers/scsi/esas2r/esas2r_ioctl.c 		} else if (spt->flags & ATTO_SPTF_DATA_OUT) {
spt              1040 drivers/scsi/esas2r/esas2r_ioctl.c 		if (spt->flags & ATTO_SPTF_ORDERED_Q)
spt              1043 drivers/scsi/esas2r/esas2r_ioctl.c 		else if (spt->flags & ATTO_SPTF_HEAD_OF_Q)
spt              2187 drivers/scsi/qla2xxx/qla_isr.c 			struct t10_pi_tuple *spt;
spt              2209 drivers/scsi/qla2xxx/qla_isr.c 			spt = page_address(sg_page(sg)) + sg->offset;
spt              2210 drivers/scsi/qla2xxx/qla_isr.c 			spt += j;
spt              2212 drivers/scsi/qla2xxx/qla_isr.c 			spt->app_tag = T10_PI_APP_ESCAPE;
spt              2214 drivers/scsi/qla2xxx/qla_isr.c 				spt->ref_tag = T10_PI_REF_ESCAPE;
spt               579 kernel/torture.c 	int spt;
spt               583 kernel/torture.c 	spt = READ_ONCE(stutter_pause_test);
spt               584 kernel/torture.c 	for (; spt; spt = READ_ONCE(stutter_pause_test)) {
spt               586 kernel/torture.c 		if (spt == 1) {
spt               588 kernel/torture.c 		} else if (spt == 2) {