spt 718 arch/powerpc/include/asm/book3s/64/mmu-hash.h struct subpage_prot_table *spt; spt 192 arch/powerpc/include/asm/book3s/64/mmu.h return ctx->hash_context->spt; spt 1166 arch/powerpc/mm/book3s64/hash_utils.c struct subpage_prot_table *spt = mm_ctx_subpage_prot(&mm->context); spt 1170 arch/powerpc/mm/book3s64/hash_utils.c if (!spt) spt 1173 arch/powerpc/mm/book3s64/hash_utils.c if (ea >= spt->maxaddr) spt 1177 arch/powerpc/mm/book3s64/hash_utils.c sbpm = spt->low_prot; spt 1179 arch/powerpc/mm/book3s64/hash_utils.c sbpm = spt->protptrs[ea >> SBP_L3_SHIFT]; spt 122 arch/powerpc/mm/book3s64/mmu_context.c if (current->mm->context.hash_context->spt) { spt 123 arch/powerpc/mm/book3s64/mmu_context.c mm->context.hash_context->spt = kmalloc(sizeof(struct subpage_prot_table), spt 125 arch/powerpc/mm/book3s64/mmu_context.c if (!mm->context.hash_context->spt) { spt 136 arch/powerpc/mm/book3s64/mmu_context.c kfree(mm->context.hash_context->spt); spt 24 arch/powerpc/mm/book3s64/subpage_prot.c struct subpage_prot_table *spt = mm_ctx_subpage_prot(&mm->context); spt 28 arch/powerpc/mm/book3s64/subpage_prot.c if (!spt) spt 32 arch/powerpc/mm/book3s64/subpage_prot.c if (spt->low_prot[i]) { spt 33 arch/powerpc/mm/book3s64/subpage_prot.c free_page((unsigned long)spt->low_prot[i]); spt 34 arch/powerpc/mm/book3s64/subpage_prot.c spt->low_prot[i] = NULL; spt 39 arch/powerpc/mm/book3s64/subpage_prot.c p = spt->protptrs[i]; spt 42 arch/powerpc/mm/book3s64/subpage_prot.c spt->protptrs[i] = NULL; spt 43 arch/powerpc/mm/book3s64/subpage_prot.c for (j = 0; j < SBP_L2_COUNT && addr < spt->maxaddr; spt 49 arch/powerpc/mm/book3s64/subpage_prot.c spt->maxaddr = 0; spt 50 arch/powerpc/mm/book3s64/subpage_prot.c kfree(spt); spt 89 arch/powerpc/mm/book3s64/subpage_prot.c struct subpage_prot_table *spt; spt 97 arch/powerpc/mm/book3s64/subpage_prot.c spt = mm_ctx_subpage_prot(&mm->context); spt 98 arch/powerpc/mm/book3s64/subpage_prot.c if (!spt) spt 102 arch/powerpc/mm/book3s64/subpage_prot.c if (limit > spt->maxaddr) spt 103 arch/powerpc/mm/book3s64/subpage_prot.c limit = spt->maxaddr; spt 107 arch/powerpc/mm/book3s64/subpage_prot.c spm = spt->low_prot; spt 109 arch/powerpc/mm/book3s64/subpage_prot.c spm = spt->protptrs[addr >> SBP_L3_SHIFT]; spt 192 arch/powerpc/mm/book3s64/subpage_prot.c struct subpage_prot_table *spt; spt 222 arch/powerpc/mm/book3s64/subpage_prot.c spt = mm_ctx_subpage_prot(&mm->context); spt 223 arch/powerpc/mm/book3s64/subpage_prot.c if (!spt) { spt 228 arch/powerpc/mm/book3s64/subpage_prot.c spt = kzalloc(sizeof(struct subpage_prot_table), GFP_KERNEL); spt 229 arch/powerpc/mm/book3s64/subpage_prot.c if (!spt) { spt 233 arch/powerpc/mm/book3s64/subpage_prot.c mm->context.hash_context->spt = spt; spt 241 arch/powerpc/mm/book3s64/subpage_prot.c spm = spt->low_prot; spt 243 arch/powerpc/mm/book3s64/subpage_prot.c spm = spt->protptrs[addr >> SBP_L3_SHIFT]; spt 248 arch/powerpc/mm/book3s64/subpage_prot.c spt->protptrs[addr >> SBP_L3_SHIFT] = spm; spt 279 arch/powerpc/mm/book3s64/subpage_prot.c if (limit > spt->maxaddr) spt 280 arch/powerpc/mm/book3s64/subpage_prot.c spt->maxaddr = limit; spt 329 arch/x86/include/asm/kvm_host.h u64 *spt; spt 1503 arch/x86/kvm/mmu.c kvm_mmu_page_set_gfn(sp, spte - sp->spt, gfn); spt 1515 arch/x86/kvm/mmu.c gfn = kvm_mmu_page_get_gfn(sp, spte - sp->spt); spt 2113 arch/x86/kvm/mmu.c static int is_empty_shadow_page(u64 *spt) spt 2118 arch/x86/kvm/mmu.c for (pos = spt, end = pos + PAGE_SIZE / sizeof(u64); pos != end; pos++) spt 2142 arch/x86/kvm/mmu.c MMU_WARN_ON(!is_empty_shadow_page(sp->spt)); spt 2145 arch/x86/kvm/mmu.c free_page((unsigned long)sp->spt); spt 2183 arch/x86/kvm/mmu.c sp->spt = mmu_memory_cache_alloc(&vcpu->arch.mmu_page_cache); spt 2186 arch/x86/kvm/mmu.c set_page_private(virt_to_page(sp->spt), (unsigned long)sp); spt 2216 arch/x86/kvm/mmu.c index = spte - sp->spt; spt 2281 arch/x86/kvm/mmu.c u64 ent = sp->spt[i]; spt 2630 arch/x86/kvm/mmu.c clear_page(sp->spt); spt 2710 arch/x86/kvm/mmu.c spte = __pa(sp->spt) | shadow_present_mask | PT_WRITABLE_MASK | spt 2779 arch/x86/kvm/mmu.c mmu_page_zap_pte(kvm, sp, sp->spt + i); spt 3241 arch/x86/kvm/mmu.c gfn = kvm_mmu_page_get_gfn(sp, start - sp->spt); spt 3267 arch/x86/kvm/mmu.c i = (sptep - sp->spt) & ~(PTE_PREFETCH_NUM - 1); spt 3268 arch/x86/kvm/mmu.c spte = sp->spt + i; spt 3512 arch/x86/kvm/mmu.c gfn = kvm_mmu_page_get_gfn(sp, sptep - sp->spt); spt 3792 arch/x86/kvm/mmu.c vcpu->arch.mmu->root_hpa = __pa(sp->spt); spt 3805 arch/x86/kvm/mmu.c root = __pa(sp->spt); spt 3847 arch/x86/kvm/mmu.c root = __pa(sp->spt); spt 3884 arch/x86/kvm/mmu.c root = __pa(sp->spt); spt 5427 arch/x86/kvm/mmu.c spte = &sp->spt[page_offset / sizeof(*spte)]; spt 40 arch/x86/kvm/mmu_audit.c u64 *ent = sp->spt; spt 113 arch/x86/kvm/mmu_audit.c gfn = kvm_mmu_page_get_gfn(sp, sptep - sp->spt); spt 136 arch/x86/kvm/mmu_audit.c gfn = kvm_mmu_page_get_gfn(rev_sp, sptep - rev_sp->spt); spt 145 arch/x86/kvm/mmu_audit.c (long int)(sptep - rev_sp->spt), rev_sp->gfn); spt 183 arch/x86/kvm/mmu_audit.c if (!is_shadow_present_pte(sp->spt[i])) spt 186 arch/x86/kvm/mmu_audit.c inspect_spte_has_rmap(kvm, sp->spt + i); spt 594 arch/x86/kvm/paging_tmpl.h i = (sptep - sp->spt) & ~(PTE_PREFETCH_NUM - 1); spt 595 arch/x86/kvm/paging_tmpl.h spte = sp->spt + i; spt 926 arch/x86/kvm/paging_tmpl.h pte_gpa += (sptep - sp->spt) * sizeof(pt_element_t); spt 1025 arch/x86/kvm/paging_tmpl.h if (!sp->spt[i]) spt 1034 arch/x86/kvm/paging_tmpl.h if (FNAME(prefetch_invalid_gpte)(vcpu, sp, &sp->spt[i], gpte)) { spt 1050 arch/x86/kvm/paging_tmpl.h if (sync_mmio_spte(vcpu, &sp->spt[i], gfn, pte_access, spt 1055 arch/x86/kvm/paging_tmpl.h drop_spte(vcpu->kvm, &sp->spt[i]); spt 1067 arch/x86/kvm/paging_tmpl.h host_writable = sp->spt[i] & SPTE_HOST_WRITEABLE; spt 1069 arch/x86/kvm/paging_tmpl.h set_spte_ret |= set_spte(vcpu, &sp->spt[i], spt 1071 arch/x86/kvm/paging_tmpl.h gfn, spte_to_pfn(sp->spt[i]), spt 170 drivers/block/ataflop.c unsigned spt; /* sectors per track */ spt 750 drivers/block/ataflop.c if (!UDT || desc->track >= UDT->blocks/UDT->spt/2 || desc->head >= 2) { spt 755 drivers/block/ataflop.c nsect = UDT->spt; spt 1018 drivers/block/ataflop.c dma_wd.fdc_acces_seccount = read_track ? SUDT->spt : 1; spt 1083 drivers/block/ataflop.c if (addr >= PhysTrackBuffer + SUDT->spt*512) { spt 1151 drivers/block/ataflop.c !(read_track && FDC_READ(FDCREG_SECTOR) > SUDT->spt)) { spt 1460 drivers/block/ataflop.c ReqTrack = block / UDT->spt; spt 1461 drivers/block/ataflop.c ReqSector = block - ReqTrack * UDT->spt + 1; spt 1594 drivers/block/ataflop.c getprm.sect = dtp->spt; spt 1596 drivers/block/ataflop.c getprm.track = dtp->blocks/dtp->spt/2; spt 1655 drivers/block/ataflop.c && dtp->spt == setprm.sect spt 1666 drivers/block/ataflop.c default_params[drive].spt = dtp->spt; spt 1688 drivers/block/ataflop.c dtp->spt = setprm.sect; spt 1697 drivers/block/ataflop.c drive, dtp->blocks, dtp->spt, dtp->stretch); spt 1700 drivers/block/ataflop.c if (setprm.track != dtp->blocks/dtp->spt/2 || spt 646 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt, spt 651 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt *gvt = spt->vgpu->gvt; spt 661 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.gfn << I915_GTT_PAGE_SHIFT, spt 662 drivers/gpu/drm/i915/gvt/gtt.c spt->vgpu); spt 667 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.pde_ips : false); spt 675 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt, spt 680 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt *gvt = spt->vgpu->gvt; spt 690 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.gfn << I915_GTT_PAGE_SHIFT, spt 691 drivers/gpu/drm/i915/gvt/gtt.c spt->vgpu); spt 694 drivers/gpu/drm/i915/gvt/gtt.c #define ppgtt_get_guest_entry(spt, e, index) \ spt 695 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_spt_get_entry(spt, NULL, \ spt 696 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.type, e, index, true) spt 698 drivers/gpu/drm/i915/gvt/gtt.c #define ppgtt_set_guest_entry(spt, e, index) \ spt 699 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_spt_set_entry(spt, NULL, \ spt 700 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.type, e, index, true) spt 702 drivers/gpu/drm/i915/gvt/gtt.c #define ppgtt_get_shadow_entry(spt, e, index) \ spt 703 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_spt_get_entry(spt, spt->shadow_page.vaddr, \ spt 704 drivers/gpu/drm/i915/gvt/gtt.c spt->shadow_page.type, e, index, false) spt 706 drivers/gpu/drm/i915/gvt/gtt.c #define ppgtt_set_shadow_entry(spt, e, index) \ spt 707 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_spt_set_entry(spt, spt->shadow_page.vaddr, \ spt 708 drivers/gpu/drm/i915/gvt/gtt.c spt->shadow_page.type, e, index, false) spt 712 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt; spt 714 drivers/gpu/drm/i915/gvt/gtt.c spt = kzalloc(sizeof(*spt), gfp_mask); spt 715 drivers/gpu/drm/i915/gvt/gtt.c if (!spt) spt 718 drivers/gpu/drm/i915/gvt/gtt.c spt->shadow_page.page = alloc_page(gfp_mask); spt 719 drivers/gpu/drm/i915/gvt/gtt.c if (!spt->shadow_page.page) { spt 720 drivers/gpu/drm/i915/gvt/gtt.c kfree(spt); spt 723 drivers/gpu/drm/i915/gvt/gtt.c return spt; spt 726 drivers/gpu/drm/i915/gvt/gtt.c static void free_spt(struct intel_vgpu_ppgtt_spt *spt) spt 728 drivers/gpu/drm/i915/gvt/gtt.c __free_page(spt->shadow_page.page); spt 729 drivers/gpu/drm/i915/gvt/gtt.c kfree(spt); spt 735 drivers/gpu/drm/i915/gvt/gtt.c static void ppgtt_free_spt(struct intel_vgpu_ppgtt_spt *spt) spt 737 drivers/gpu/drm/i915/gvt/gtt.c struct device *kdev = &spt->vgpu->gvt->dev_priv->drm.pdev->dev; spt 739 drivers/gpu/drm/i915/gvt/gtt.c trace_spt_free(spt->vgpu->id, spt, spt->guest_page.type); spt 741 drivers/gpu/drm/i915/gvt/gtt.c dma_unmap_page(kdev, spt->shadow_page.mfn << I915_GTT_PAGE_SHIFT, 4096, spt 744 drivers/gpu/drm/i915/gvt/gtt.c radix_tree_delete(&spt->vgpu->gtt.spt_tree, spt->shadow_page.mfn); spt 746 drivers/gpu/drm/i915/gvt/gtt.c if (spt->guest_page.gfn) { spt 747 drivers/gpu/drm/i915/gvt/gtt.c if (spt->guest_page.oos_page) spt 748 drivers/gpu/drm/i915/gvt/gtt.c detach_oos_page(spt->vgpu, spt->guest_page.oos_page); spt 750 drivers/gpu/drm/i915/gvt/gtt.c intel_vgpu_unregister_page_track(spt->vgpu, spt->guest_page.gfn); spt 753 drivers/gpu/drm/i915/gvt/gtt.c list_del_init(&spt->post_shadow_list); spt 754 drivers/gpu/drm/i915/gvt/gtt.c free_spt(spt); spt 759 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt, *spn; spt 766 drivers/gpu/drm/i915/gvt/gtt.c spt = radix_tree_deref_slot(slot); spt 767 drivers/gpu/drm/i915/gvt/gtt.c list_move(&spt->post_shadow_list, &all_spt); spt 771 drivers/gpu/drm/i915/gvt/gtt.c list_for_each_entry_safe(spt, spn, &all_spt, post_shadow_list) spt 772 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_free_spt(spt); spt 776 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt, spt 783 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt = page_track->priv_data; spt 790 drivers/gpu/drm/i915/gvt/gtt.c ret = ppgtt_handle_guest_write_page_table_bytes(spt, gpa, data, bytes); spt 823 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt = NULL; spt 828 drivers/gpu/drm/i915/gvt/gtt.c spt = alloc_spt(GFP_KERNEL | __GFP_ZERO); spt 829 drivers/gpu/drm/i915/gvt/gtt.c if (!spt) { spt 837 drivers/gpu/drm/i915/gvt/gtt.c spt->vgpu = vgpu; spt 838 drivers/gpu/drm/i915/gvt/gtt.c atomic_set(&spt->refcount, 1); spt 839 drivers/gpu/drm/i915/gvt/gtt.c INIT_LIST_HEAD(&spt->post_shadow_list); spt 844 drivers/gpu/drm/i915/gvt/gtt.c spt->shadow_page.type = type; spt 845 drivers/gpu/drm/i915/gvt/gtt.c daddr = dma_map_page(kdev, spt->shadow_page.page, spt 852 drivers/gpu/drm/i915/gvt/gtt.c spt->shadow_page.vaddr = page_address(spt->shadow_page.page); spt 853 drivers/gpu/drm/i915/gvt/gtt.c spt->shadow_page.mfn = daddr >> I915_GTT_PAGE_SHIFT; spt 855 drivers/gpu/drm/i915/gvt/gtt.c ret = radix_tree_insert(&vgpu->gtt.spt_tree, spt->shadow_page.mfn, spt); spt 859 drivers/gpu/drm/i915/gvt/gtt.c return spt; spt 864 drivers/gpu/drm/i915/gvt/gtt.c free_spt(spt); spt 873 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt; spt 876 drivers/gpu/drm/i915/gvt/gtt.c spt = ppgtt_alloc_spt(vgpu, type); spt 877 drivers/gpu/drm/i915/gvt/gtt.c if (IS_ERR(spt)) spt 878 drivers/gpu/drm/i915/gvt/gtt.c return spt; spt 884 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_write_protection_handler, spt); spt 886 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_free_spt(spt); spt 890 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.type = type; spt 891 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.gfn = gfn; spt 892 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.pde_ips = guest_pde_ips; spt 894 drivers/gpu/drm/i915/gvt/gtt.c trace_spt_alloc(vgpu->id, spt, type, spt->shadow_page.mfn, gfn); spt 896 drivers/gpu/drm/i915/gvt/gtt.c return spt; spt 899 drivers/gpu/drm/i915/gvt/gtt.c #define pt_entry_size_shift(spt) \ spt 900 drivers/gpu/drm/i915/gvt/gtt.c ((spt)->vgpu->gvt->device_info.gtt_entry_size_shift) spt 902 drivers/gpu/drm/i915/gvt/gtt.c #define pt_entries(spt) \ spt 903 drivers/gpu/drm/i915/gvt/gtt.c (I915_GTT_PAGE_SIZE >> pt_entry_size_shift(spt)) spt 905 drivers/gpu/drm/i915/gvt/gtt.c #define for_each_present_guest_entry(spt, e, i) \ spt 906 drivers/gpu/drm/i915/gvt/gtt.c for (i = 0; i < pt_entries(spt); \ spt 907 drivers/gpu/drm/i915/gvt/gtt.c i += spt->guest_page.pde_ips ? GTT_64K_PTE_STRIDE : 1) \ spt 908 drivers/gpu/drm/i915/gvt/gtt.c if (!ppgtt_get_guest_entry(spt, e, i) && \ spt 909 drivers/gpu/drm/i915/gvt/gtt.c spt->vgpu->gvt->gtt.pte_ops->test_present(e)) spt 911 drivers/gpu/drm/i915/gvt/gtt.c #define for_each_present_shadow_entry(spt, e, i) \ spt 912 drivers/gpu/drm/i915/gvt/gtt.c for (i = 0; i < pt_entries(spt); \ spt 913 drivers/gpu/drm/i915/gvt/gtt.c i += spt->shadow_page.pde_ips ? GTT_64K_PTE_STRIDE : 1) \ spt 914 drivers/gpu/drm/i915/gvt/gtt.c if (!ppgtt_get_shadow_entry(spt, e, i) && \ spt 915 drivers/gpu/drm/i915/gvt/gtt.c spt->vgpu->gvt->gtt.pte_ops->test_present(e)) spt 917 drivers/gpu/drm/i915/gvt/gtt.c #define for_each_shadow_entry(spt, e, i) \ spt 918 drivers/gpu/drm/i915/gvt/gtt.c for (i = 0; i < pt_entries(spt); \ spt 919 drivers/gpu/drm/i915/gvt/gtt.c i += (spt->shadow_page.pde_ips ? GTT_64K_PTE_STRIDE : 1)) \ spt 920 drivers/gpu/drm/i915/gvt/gtt.c if (!ppgtt_get_shadow_entry(spt, e, i)) spt 922 drivers/gpu/drm/i915/gvt/gtt.c static inline void ppgtt_get_spt(struct intel_vgpu_ppgtt_spt *spt) spt 924 drivers/gpu/drm/i915/gvt/gtt.c int v = atomic_read(&spt->refcount); spt 926 drivers/gpu/drm/i915/gvt/gtt.c trace_spt_refcount(spt->vgpu->id, "inc", spt, v, (v + 1)); spt 927 drivers/gpu/drm/i915/gvt/gtt.c atomic_inc(&spt->refcount); spt 930 drivers/gpu/drm/i915/gvt/gtt.c static inline int ppgtt_put_spt(struct intel_vgpu_ppgtt_spt *spt) spt 932 drivers/gpu/drm/i915/gvt/gtt.c int v = atomic_read(&spt->refcount); spt 934 drivers/gpu/drm/i915/gvt/gtt.c trace_spt_refcount(spt->vgpu->id, "dec", spt, v, (v - 1)); spt 935 drivers/gpu/drm/i915/gvt/gtt.c return atomic_dec_return(&spt->refcount); spt 938 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_invalidate_spt(struct intel_vgpu_ppgtt_spt *spt); spt 974 drivers/gpu/drm/i915/gvt/gtt.c static inline void ppgtt_invalidate_pte(struct intel_vgpu_ppgtt_spt *spt, spt 977 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu *vgpu = spt->vgpu; spt 983 drivers/gpu/drm/i915/gvt/gtt.c type = spt->shadow_page.type; spt 992 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_invalidate_spt(struct intel_vgpu_ppgtt_spt *spt) spt 994 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu *vgpu = spt->vgpu; spt 999 drivers/gpu/drm/i915/gvt/gtt.c trace_spt_change(spt->vgpu->id, "die", spt, spt 1000 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.gfn, spt->shadow_page.type); spt 1002 drivers/gpu/drm/i915/gvt/gtt.c if (ppgtt_put_spt(spt) > 0) spt 1005 drivers/gpu/drm/i915/gvt/gtt.c for_each_present_shadow_entry(spt, &e, index) { spt 1009 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_invalidate_pte(spt, &e); spt 1026 drivers/gpu/drm/i915/gvt/gtt.c spt->vgpu, &e); spt 1035 drivers/gpu/drm/i915/gvt/gtt.c trace_spt_change(spt->vgpu->id, "release", spt, spt 1036 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.gfn, spt->shadow_page.type); spt 1037 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_free_spt(spt); spt 1041 drivers/gpu/drm/i915/gvt/gtt.c spt, e.val64, e.type); spt 1061 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_populate_spt(struct intel_vgpu_ppgtt_spt *spt); spt 1067 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt = NULL; spt 1076 drivers/gpu/drm/i915/gvt/gtt.c spt = intel_vgpu_find_spt_by_gfn(vgpu, ops->get_pfn(we)); spt 1077 drivers/gpu/drm/i915/gvt/gtt.c if (spt) { spt 1078 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_get_spt(spt); spt 1080 drivers/gpu/drm/i915/gvt/gtt.c if (ips != spt->guest_page.pde_ips) { spt 1081 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.pde_ips = ips; spt 1084 drivers/gpu/drm/i915/gvt/gtt.c clear_page(spt->shadow_page.vaddr); spt 1085 drivers/gpu/drm/i915/gvt/gtt.c ret = ppgtt_populate_spt(spt); spt 1087 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_put_spt(spt); spt 1099 drivers/gpu/drm/i915/gvt/gtt.c spt = ppgtt_alloc_spt_gfn(vgpu, type, ops->get_pfn(we), ips); spt 1100 drivers/gpu/drm/i915/gvt/gtt.c if (IS_ERR(spt)) { spt 1101 drivers/gpu/drm/i915/gvt/gtt.c ret = PTR_ERR(spt); spt 1105 drivers/gpu/drm/i915/gvt/gtt.c ret = intel_vgpu_enable_page_track(vgpu, spt->guest_page.gfn); spt 1109 drivers/gpu/drm/i915/gvt/gtt.c ret = ppgtt_populate_spt(spt); spt 1113 drivers/gpu/drm/i915/gvt/gtt.c trace_spt_change(vgpu->id, "new", spt, spt->guest_page.gfn, spt 1114 drivers/gpu/drm/i915/gvt/gtt.c spt->shadow_page.type); spt 1116 drivers/gpu/drm/i915/gvt/gtt.c return spt; spt 1119 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_free_spt(spt); spt 1120 drivers/gpu/drm/i915/gvt/gtt.c spt = NULL; spt 1123 drivers/gpu/drm/i915/gvt/gtt.c spt, we->val64, we->type); spt 1167 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt, unsigned long index, spt 1190 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_invalidate_spt(spt); spt 1209 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_set_shadow_entry(spt, se, index); spt 1214 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt, unsigned long index, spt 1239 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_set_shadow_entry(spt, &entry, index + i); spt 1245 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt, unsigned long index, spt 1270 drivers/gpu/drm/i915/gvt/gtt.c return split_64KB_gtt_entry(vgpu, spt, index, &se); spt 1275 drivers/gpu/drm/i915/gvt/gtt.c return split_2MB_gtt_entry(vgpu, spt, index, &se); spt 1294 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_set_shadow_entry(spt, &se, index); spt 1298 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_populate_spt(struct intel_vgpu_ppgtt_spt *spt) spt 1300 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu *vgpu = spt->vgpu; spt 1308 drivers/gpu/drm/i915/gvt/gtt.c trace_spt_change(spt->vgpu->id, "born", spt, spt 1309 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.gfn, spt->shadow_page.type); spt 1311 drivers/gpu/drm/i915/gvt/gtt.c for_each_present_guest_entry(spt, &ge, i) { spt 1318 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_get_shadow_entry(spt, &se, i); spt 1320 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_set_shadow_entry(spt, &se, i); spt 1325 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_set_shadow_entry(spt, &se, i); spt 1329 drivers/gpu/drm/i915/gvt/gtt.c ret = ppgtt_populate_shadow_entry(vgpu, spt, i, &ge); spt 1337 drivers/gpu/drm/i915/gvt/gtt.c spt, ge.val64, ge.type); spt 1341 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_handle_guest_entry_removal(struct intel_vgpu_ppgtt_spt *spt, spt 1344 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu *vgpu = spt->vgpu; spt 1348 drivers/gpu/drm/i915/gvt/gtt.c trace_spt_guest_change(spt->vgpu->id, "remove", spt, spt 1349 drivers/gpu/drm/i915/gvt/gtt.c spt->shadow_page.type, se->val64, index); spt 1358 drivers/gpu/drm/i915/gvt/gtt.c vgpu->gtt.scratch_pt[spt->shadow_page.type].page_mfn) spt 1376 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_invalidate_pte(spt, se); spt 1382 drivers/gpu/drm/i915/gvt/gtt.c spt, se->val64, se->type); spt 1386 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_handle_guest_entry_add(struct intel_vgpu_ppgtt_spt *spt, spt 1389 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu *vgpu = spt->vgpu; spt 1394 drivers/gpu/drm/i915/gvt/gtt.c trace_spt_guest_change(spt->vgpu->id, "add", spt, spt->shadow_page.type, spt 1406 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_get_shadow_entry(spt, &m, index); spt 1408 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_set_shadow_entry(spt, &m, index); spt 1410 drivers/gpu/drm/i915/gvt/gtt.c ret = ppgtt_populate_shadow_entry(vgpu, spt, index, we); spt 1417 drivers/gpu/drm/i915/gvt/gtt.c spt, we->val64, we->type); spt 1427 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt = oos_page->spt; spt 1433 drivers/gpu/drm/i915/gvt/gtt.c spt, spt->guest_page.type); spt 1435 drivers/gpu/drm/i915/gvt/gtt.c old.type = new.type = get_entry_type(spt->guest_page.type); spt 1442 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.gfn << PAGE_SHIFT, vgpu); spt 1445 drivers/gpu/drm/i915/gvt/gtt.c && !test_and_clear_bit(index, spt->post_shadow_bitmap)) spt 1449 drivers/gpu/drm/i915/gvt/gtt.c spt, spt->guest_page.type, spt 1452 drivers/gpu/drm/i915/gvt/gtt.c ret = ppgtt_populate_shadow_entry(vgpu, spt, index, &new); spt 1459 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.write_cnt = 0; spt 1460 drivers/gpu/drm/i915/gvt/gtt.c list_del_init(&spt->post_shadow_list); spt 1468 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt = oos_page->spt; spt 1471 drivers/gpu/drm/i915/gvt/gtt.c spt, spt->guest_page.type); spt 1473 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.write_cnt = 0; spt 1474 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.oos_page = NULL; spt 1475 drivers/gpu/drm/i915/gvt/gtt.c oos_page->spt = NULL; spt 1484 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt) spt 1486 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt *gvt = spt->vgpu->gvt; spt 1489 drivers/gpu/drm/i915/gvt/gtt.c ret = intel_gvt_hypervisor_read_gpa(spt->vgpu, spt 1490 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.gfn << I915_GTT_PAGE_SHIFT, spt 1495 drivers/gpu/drm/i915/gvt/gtt.c oos_page->spt = spt; spt 1496 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.oos_page = oos_page; spt 1500 drivers/gpu/drm/i915/gvt/gtt.c trace_oos_change(spt->vgpu->id, "attach", oos_page->id, spt 1501 drivers/gpu/drm/i915/gvt/gtt.c spt, spt->guest_page.type); spt 1505 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_set_guest_page_sync(struct intel_vgpu_ppgtt_spt *spt) spt 1507 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_oos_page *oos_page = spt->guest_page.oos_page; spt 1510 drivers/gpu/drm/i915/gvt/gtt.c ret = intel_vgpu_enable_page_track(spt->vgpu, spt->guest_page.gfn); spt 1514 drivers/gpu/drm/i915/gvt/gtt.c trace_oos_change(spt->vgpu->id, "set page sync", oos_page->id, spt 1515 drivers/gpu/drm/i915/gvt/gtt.c spt, spt->guest_page.type); spt 1518 drivers/gpu/drm/i915/gvt/gtt.c return sync_oos_page(spt->vgpu, oos_page); spt 1521 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_allocate_oos_page(struct intel_vgpu_ppgtt_spt *spt) spt 1523 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt *gvt = spt->vgpu->gvt; spt 1525 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_oos_page *oos_page = spt->guest_page.oos_page; spt 1533 drivers/gpu/drm/i915/gvt/gtt.c ret = ppgtt_set_guest_page_sync(oos_page->spt); spt 1536 drivers/gpu/drm/i915/gvt/gtt.c ret = detach_oos_page(spt->vgpu, oos_page); spt 1542 drivers/gpu/drm/i915/gvt/gtt.c return attach_oos_page(oos_page, spt); spt 1545 drivers/gpu/drm/i915/gvt/gtt.c static int ppgtt_set_guest_page_oos(struct intel_vgpu_ppgtt_spt *spt) spt 1547 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_oos_page *oos_page = spt->guest_page.oos_page; spt 1552 drivers/gpu/drm/i915/gvt/gtt.c trace_oos_change(spt->vgpu->id, "set page out of sync", oos_page->id, spt 1553 drivers/gpu/drm/i915/gvt/gtt.c spt, spt->guest_page.type); spt 1555 drivers/gpu/drm/i915/gvt/gtt.c list_add_tail(&oos_page->vm_list, &spt->vgpu->gtt.oos_page_list_head); spt 1556 drivers/gpu/drm/i915/gvt/gtt.c return intel_vgpu_disable_page_track(spt->vgpu, spt->guest_page.gfn); spt 1581 drivers/gpu/drm/i915/gvt/gtt.c ret = ppgtt_set_guest_page_sync(oos_page->spt); spt 1592 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt, spt 1595 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu *vgpu = spt->vgpu; spt 1596 drivers/gpu/drm/i915/gvt/gtt.c int type = spt->shadow_page.type; spt 1609 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_get_shadow_entry(spt, &old_se, index); spt 1612 drivers/gpu/drm/i915/gvt/gtt.c ret = ppgtt_handle_guest_entry_add(spt, we, index); spt 1617 drivers/gpu/drm/i915/gvt/gtt.c ret = ppgtt_handle_guest_entry_removal(spt, &old_se, index); spt 1630 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_set_shadow_entry(spt, &old_se, index + i); spt 1637 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_set_shadow_entry(spt, &old_se, index); spt 1641 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_set_shadow_entry(spt, &old_se, index); spt 1648 drivers/gpu/drm/i915/gvt/gtt.c spt, we->val64, we->type); spt 1654 drivers/gpu/drm/i915/gvt/gtt.c static inline bool can_do_out_of_sync(struct intel_vgpu_ppgtt_spt *spt) spt 1657 drivers/gpu/drm/i915/gvt/gtt.c && gtt_type_is_pte_pt(spt->guest_page.type) spt 1658 drivers/gpu/drm/i915/gvt/gtt.c && spt->guest_page.write_cnt >= 2; spt 1661 drivers/gpu/drm/i915/gvt/gtt.c static void ppgtt_set_post_shadow(struct intel_vgpu_ppgtt_spt *spt, spt 1664 drivers/gpu/drm/i915/gvt/gtt.c set_bit(index, spt->post_shadow_bitmap); spt 1665 drivers/gpu/drm/i915/gvt/gtt.c if (!list_empty(&spt->post_shadow_list)) spt 1668 drivers/gpu/drm/i915/gvt/gtt.c list_add_tail(&spt->post_shadow_list, spt 1669 drivers/gpu/drm/i915/gvt/gtt.c &spt->vgpu->gtt.post_shadow_list_head); spt 1685 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt; spt 1691 drivers/gpu/drm/i915/gvt/gtt.c spt = container_of(pos, struct intel_vgpu_ppgtt_spt, spt 1694 drivers/gpu/drm/i915/gvt/gtt.c for_each_set_bit(index, spt->post_shadow_bitmap, spt 1696 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_get_guest_entry(spt, &ge, index); spt 1698 drivers/gpu/drm/i915/gvt/gtt.c ret = ppgtt_handle_guest_write_page_table(spt, spt 1702 drivers/gpu/drm/i915/gvt/gtt.c clear_bit(index, spt->post_shadow_bitmap); spt 1704 drivers/gpu/drm/i915/gvt/gtt.c list_del_init(&spt->post_shadow_list); spt 1710 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt, spt 1713 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu *vgpu = spt->vgpu; spt 1722 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_get_guest_entry(spt, &we, index); spt 1737 drivers/gpu/drm/i915/gvt/gtt.c ret = ppgtt_handle_guest_write_page_table(spt, &we, index); spt 1741 drivers/gpu/drm/i915/gvt/gtt.c if (!test_bit(index, spt->post_shadow_bitmap)) { spt 1742 drivers/gpu/drm/i915/gvt/gtt.c int type = spt->shadow_page.type; spt 1744 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_get_shadow_entry(spt, &se, index); spt 1745 drivers/gpu/drm/i915/gvt/gtt.c ret = ppgtt_handle_guest_entry_removal(spt, &se, index); spt 1749 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_set_shadow_entry(spt, &se, index); spt 1751 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_set_post_shadow(spt, index); spt 1757 drivers/gpu/drm/i915/gvt/gtt.c spt->guest_page.write_cnt++; spt 1759 drivers/gpu/drm/i915/gvt/gtt.c if (spt->guest_page.oos_page) spt 1760 drivers/gpu/drm/i915/gvt/gtt.c ops->set_entry(spt->guest_page.oos_page->mem, &we, index, spt 1763 drivers/gpu/drm/i915/gvt/gtt.c if (can_do_out_of_sync(spt)) { spt 1764 drivers/gpu/drm/i915/gvt/gtt.c if (!spt->guest_page.oos_page) spt 1765 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_allocate_oos_page(spt); spt 1767 drivers/gpu/drm/i915/gvt/gtt.c ret = ppgtt_set_guest_page_oos(spt); spt 1810 drivers/gpu/drm/i915/gvt/gtt.c struct intel_vgpu_ppgtt_spt *spt; spt 1828 drivers/gpu/drm/i915/gvt/gtt.c spt = ppgtt_populate_spt_by_guest_entry(vgpu, &ge); spt 1829 drivers/gpu/drm/i915/gvt/gtt.c if (IS_ERR(spt)) { spt 1831 drivers/gpu/drm/i915/gvt/gtt.c ret = PTR_ERR(spt); spt 1834 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_generate_shadow_entry(&se, spt, &ge); spt 222 drivers/gpu/drm/i915/gvt/gtt.h struct intel_vgpu_ppgtt_spt *spt; spt 43 drivers/gpu/drm/i915/gvt/trace.h TP_PROTO(int id, void *spt, int type, unsigned long mfn, spt 46 drivers/gpu/drm/i915/gvt/trace.h TP_ARGS(id, spt, type, mfn, gpt_gfn), spt 50 drivers/gpu/drm/i915/gvt/trace.h __field(void *, spt) spt 58 drivers/gpu/drm/i915/gvt/trace.h __entry->spt = spt; spt 66 drivers/gpu/drm/i915/gvt/trace.h __entry->spt, spt 73 drivers/gpu/drm/i915/gvt/trace.h TP_PROTO(int id, void *spt, int type), spt 75 drivers/gpu/drm/i915/gvt/trace.h TP_ARGS(id, spt, type), spt 79 drivers/gpu/drm/i915/gvt/trace.h __field(void *, spt) spt 85 drivers/gpu/drm/i915/gvt/trace.h __entry->spt = spt; spt 91 drivers/gpu/drm/i915/gvt/trace.h __entry->spt, spt 135 drivers/gpu/drm/i915/gvt/trace.h TP_PROTO(int id, char *action, void *spt, int before, int after), spt 137 drivers/gpu/drm/i915/gvt/trace.h TP_ARGS(id, action, spt, before, after), spt 146 drivers/gpu/drm/i915/gvt/trace.h id, action, spt, before, after); spt 153 drivers/gpu/drm/i915/gvt/trace.h TP_PROTO(int id, char *action, void *spt, unsigned long gfn, spt 156 drivers/gpu/drm/i915/gvt/trace.h TP_ARGS(id, action, spt, gfn, type), spt 165 drivers/gpu/drm/i915/gvt/trace.h id, action, spt, gfn, type); spt 172 drivers/gpu/drm/i915/gvt/trace.h TP_PROTO(int id, const char *tag, void *spt, int type, u64 v, spt 175 drivers/gpu/drm/i915/gvt/trace.h TP_ARGS(id, tag, spt, type, v, index), spt 184 drivers/gpu/drm/i915/gvt/trace.h id, tag, spt, type, v, index); spt 66 drivers/misc/mic/host/mic_smpt.c static void mic_add_smpt_entry(int spt, s64 *ref, u64 addr, spt 72 drivers/misc/mic/host/mic_smpt.c for (i = spt; i < spt + entries; i++, spt 79 drivers/misc/mic/host/mic_smpt.c smpt_info->entry[i].ref_count += ref[i - spt]; spt 90 drivers/misc/mic/host/mic_smpt.c int spt; spt 124 drivers/misc/mic/host/mic_smpt.c spt = i - entries + 1; spt 125 drivers/misc/mic/host/mic_smpt.c mic_addr = mic_smpt_to_pa(mdev, spt); spt 126 drivers/misc/mic/host/mic_smpt.c mic_add_smpt_entry(spt, ref, dma_addr, entries, mdev); spt 168 drivers/misc/mic/host/mic_smpt.c int spt; spt 176 drivers/misc/mic/host/mic_smpt.c spt = mic_sys_addr_to_smpt(mdev, mic_addr); spt 177 drivers/misc/mic/host/mic_smpt.c dma_addr = smpt_info->entry[spt].dma_addr + spt 247 drivers/misc/mic/host/mic_smpt.c int spt; spt 260 drivers/misc/mic/host/mic_smpt.c spt = mic_sys_addr_to_smpt(mdev, mic_addr); spt 272 drivers/misc/mic/host/mic_smpt.c for (i = spt; i < spt + num_smpt; i++) { spt 273 drivers/misc/mic/host/mic_smpt.c smpt_info->entry[i].ref_count -= ref[i - spt]; spt 691 drivers/scsi/esas2r/esas2r_ioctl.c struct atto_hba_scsi_pass_thru *spt = &hi->data.scsi_pass_thru; spt 694 drivers/scsi/esas2r/esas2r_ioctl.c spt->scsi_status = rq->func_rsp.scsi_rsp.scsi_stat; spt 695 drivers/scsi/esas2r/esas2r_ioctl.c spt->sense_length = rq->sense_len; spt 696 drivers/scsi/esas2r/esas2r_ioctl.c spt->residual_length = spt 734 drivers/scsi/esas2r/esas2r_ioctl.c spt->req_status = sts; spt 737 drivers/scsi/esas2r/esas2r_ioctl.c spt->target_id = spt 738 drivers/scsi/esas2r/esas2r_ioctl.c esas2r_targ_db_find_next_present(a, (u16)spt->target_id); spt 982 drivers/scsi/esas2r/esas2r_ioctl.c struct atto_hba_scsi_pass_thru *spt = &hi->data.scsi_pass_thru; spt 985 drivers/scsi/esas2r/esas2r_ioctl.c memcpy(&lun, spt->lun, sizeof(struct scsi_lun)); spt 1000 drivers/scsi/esas2r/esas2r_ioctl.c if (spt->target_id >= ESAS2R_MAX_TARGETS || !check_lun(lun)) { spt 1012 drivers/scsi/esas2r/esas2r_ioctl.c rq->target_id = (u16)spt->target_id; spt 1013 drivers/scsi/esas2r/esas2r_ioctl.c rq->vrq->scsi.flags |= cpu_to_le32(spt->lun[1]); spt 1014 drivers/scsi/esas2r/esas2r_ioctl.c memcpy(rq->vrq->scsi.cdb, spt->cdb, 16); spt 1016 drivers/scsi/esas2r/esas2r_ioctl.c rq->sense_len = spt->sense_length; spt 1017 drivers/scsi/esas2r/esas2r_ioctl.c rq->sense_buf = (u8 *)spt->sense_data; spt 1029 drivers/scsi/esas2r/esas2r_ioctl.c if (spt->flags & ATTO_SPTF_DATA_IN) { spt 1031 drivers/scsi/esas2r/esas2r_ioctl.c } else if (spt->flags & ATTO_SPTF_DATA_OUT) { spt 1040 drivers/scsi/esas2r/esas2r_ioctl.c if (spt->flags & ATTO_SPTF_ORDERED_Q) spt 1043 drivers/scsi/esas2r/esas2r_ioctl.c else if (spt->flags & ATTO_SPTF_HEAD_OF_Q) spt 2187 drivers/scsi/qla2xxx/qla_isr.c struct t10_pi_tuple *spt; spt 2209 drivers/scsi/qla2xxx/qla_isr.c spt = page_address(sg_page(sg)) + sg->offset; spt 2210 drivers/scsi/qla2xxx/qla_isr.c spt += j; spt 2212 drivers/scsi/qla2xxx/qla_isr.c spt->app_tag = T10_PI_APP_ESCAPE; spt 2214 drivers/scsi/qla2xxx/qla_isr.c spt->ref_tag = T10_PI_REF_ESCAPE; spt 579 kernel/torture.c int spt; spt 583 kernel/torture.c spt = READ_ONCE(stutter_pause_test); spt 584 kernel/torture.c for (; spt; spt = READ_ONCE(stutter_pause_test)) { spt 586 kernel/torture.c if (spt == 1) { spt 588 kernel/torture.c } else if (spt == 2) {