orig_pte 123 arch/arm64/mm/hugetlbpage.c pte_t orig_pte = huge_ptep_get(ptep); orig_pte 124 arch/arm64/mm/hugetlbpage.c bool valid = pte_valid(orig_pte); orig_pte 136 arch/arm64/mm/hugetlbpage.c orig_pte = pte_mkdirty(orig_pte); orig_pte 139 arch/arm64/mm/hugetlbpage.c orig_pte = pte_mkyoung(orig_pte); orig_pte 146 arch/arm64/mm/hugetlbpage.c return orig_pte; orig_pte 330 arch/arm64/mm/hugetlbpage.c pte_t orig_pte = huge_ptep_get(ptep); orig_pte 332 arch/arm64/mm/hugetlbpage.c if (!pte_cont(orig_pte)) orig_pte 357 arch/arm64/mm/hugetlbpage.c pte_t orig_pte = huge_ptep_get(ptep + i); orig_pte 359 arch/arm64/mm/hugetlbpage.c if (pte_dirty(pte) != pte_dirty(orig_pte)) orig_pte 362 arch/arm64/mm/hugetlbpage.c if (pte_young(pte) != pte_young(orig_pte)) orig_pte 377 arch/arm64/mm/hugetlbpage.c pte_t orig_pte; orig_pte 388 arch/arm64/mm/hugetlbpage.c orig_pte = get_clear_flush(vma->vm_mm, addr, ptep, pgsize, ncontig); orig_pte 391 arch/arm64/mm/hugetlbpage.c if (pte_dirty(orig_pte)) orig_pte 394 arch/arm64/mm/hugetlbpage.c if (pte_young(orig_pte)) orig_pte 130 arch/powerpc/kvm/book3s_32_mmu_host.c int kvmppc_mmu_map_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *orig_pte, orig_pte 138 arch/powerpc/kvm/book3s_32_mmu_host.c u32 eaddr = orig_pte->eaddr; orig_pte 148 arch/powerpc/kvm/book3s_32_mmu_host.c hpaddr = kvmppc_gpa_to_pfn(vcpu, orig_pte->raddr, iswrite, &writable); orig_pte 151 arch/powerpc/kvm/book3s_32_mmu_host.c orig_pte->raddr); orig_pte 158 arch/powerpc/kvm/book3s_32_mmu_host.c vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid); orig_pte 198 arch/powerpc/kvm/book3s_32_mmu_host.c if (orig_pte->may_write && writable) { orig_pte 200 arch/powerpc/kvm/book3s_32_mmu_host.c mark_page_dirty(vcpu->kvm, orig_pte->raddr >> PAGE_SHIFT); orig_pte 205 arch/powerpc/kvm/book3s_32_mmu_host.c if (orig_pte->may_execute) orig_pte 241 arch/powerpc/kvm/book3s_32_mmu_host.c orig_pte->may_write ? 'w' : '-', orig_pte 242 arch/powerpc/kvm/book3s_32_mmu_host.c orig_pte->may_execute ? 'x' : '-', orig_pte 243 arch/powerpc/kvm/book3s_32_mmu_host.c orig_pte->eaddr, (ulong)pteg, vpn, orig_pte 244 arch/powerpc/kvm/book3s_32_mmu_host.c orig_pte->vpage, hpaddr); orig_pte 248 arch/powerpc/kvm/book3s_32_mmu_host.c pte->pte = *orig_pte; orig_pte 70 arch/powerpc/kvm/book3s_64_mmu_host.c int kvmppc_mmu_map_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *orig_pte, orig_pte 88 arch/powerpc/kvm/book3s_64_mmu_host.c unsigned long gfn = orig_pte->raddr >> PAGE_SHIFT; orig_pte 96 arch/powerpc/kvm/book3s_64_mmu_host.c pfn = kvmppc_gpa_to_pfn(vcpu, orig_pte->raddr, iswrite, &writable); orig_pte 99 arch/powerpc/kvm/book3s_64_mmu_host.c orig_pte->raddr); orig_pte 106 arch/powerpc/kvm/book3s_64_mmu_host.c vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid); orig_pte 109 arch/powerpc/kvm/book3s_64_mmu_host.c ret = kvmppc_mmu_map_segment(vcpu, orig_pte->eaddr); orig_pte 115 arch/powerpc/kvm/book3s_64_mmu_host.c vsid, orig_pte->eaddr); orig_pte 121 arch/powerpc/kvm/book3s_64_mmu_host.c vpn = hpt_vpn(orig_pte->eaddr, map->host_vsid, MMU_SEGSIZE_256M); orig_pte 124 arch/powerpc/kvm/book3s_64_mmu_host.c if (!orig_pte->may_write || !writable) orig_pte 131 arch/powerpc/kvm/book3s_64_mmu_host.c if (!orig_pte->may_execute) orig_pte 136 arch/powerpc/kvm/book3s_64_mmu_host.c rflags = (rflags & ~HPTE_R_WIMG) | orig_pte->wimg; orig_pte 145 arch/powerpc/kvm/book3s_64_mmu_host.c hpaddr |= orig_pte->raddr & (~0xfffULL & ~PAGE_MASK); orig_pte 181 arch/powerpc/kvm/book3s_64_mmu_host.c vpn, hpaddr, orig_pte); orig_pte 194 arch/powerpc/kvm/book3s_64_mmu_host.c cpte->pte = *orig_pte; orig_pte 505 arch/powerpc/kvm/book3s_hv_rm_mmu.c u64 pte, orig_pte, pte_r; orig_pte 514 arch/powerpc/kvm/book3s_hv_rm_mmu.c pte = orig_pte = be64_to_cpu(hpte[0]); orig_pte 523 arch/powerpc/kvm/book3s_hv_rm_mmu.c __unlock_hpte(hpte, orig_pte); orig_pte 33 arch/powerpc/kvm/trace_pr.h struct kvmppc_pte *orig_pte), orig_pte 34 arch/powerpc/kvm/trace_pr.h TP_ARGS(rflags, hpteg, va, hpaddr, orig_pte), orig_pte 49 arch/powerpc/kvm/trace_pr.h __entry->eaddr = orig_pte->eaddr; orig_pte 52 arch/powerpc/kvm/trace_pr.h __entry->vpage = orig_pte->vpage; orig_pte 207 arch/sparc/mm/tlb.c pte_t orig_pte = __pte(pmd_val(orig)); orig_pte 208 arch/sparc/mm/tlb.c bool exec = pte_exec(orig_pte); orig_pte 133 arch/x86/kvm/paging_tmpl.h pt_element_t orig_pte, pt_element_t new_pte) orig_pte 143 arch/x86/kvm/paging_tmpl.h ret = CMPXCHG(&table[index], orig_pte, new_pte); orig_pte 166 arch/x86/kvm/paging_tmpl.h ret = CMPXCHG(&table[index], orig_pte, new_pte); orig_pte 171 arch/x86/kvm/paging_tmpl.h return (ret != orig_pte); orig_pte 226 arch/x86/kvm/paging_tmpl.h pt_element_t pte, orig_pte; orig_pte 236 arch/x86/kvm/paging_tmpl.h pte = orig_pte = walker->ptes[level - 1]; orig_pte 253 arch/x86/kvm/paging_tmpl.h if (pte == orig_pte) orig_pte 272 arch/x86/kvm/paging_tmpl.h ret = FNAME(cmpxchg_gpte)(vcpu, mmu, ptep_user, index, orig_pte, pte); orig_pte 1369 fs/proc/task_mmu.c pte_t *pte, *orig_pte; orig_pte 1439 fs/proc/task_mmu.c orig_pte = pte = pte_offset_map_lock(walk->mm, pmdp, addr, &ptl); orig_pte 1448 fs/proc/task_mmu.c pte_unmap_unlock(orig_pte, ptl); orig_pte 1749 fs/proc/task_mmu.c pte_t *orig_pte; orig_pte 1768 fs/proc/task_mmu.c orig_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); orig_pte 1776 fs/proc/task_mmu.c pte_unmap_unlock(orig_pte, ptl); orig_pte 471 include/asm-generic/pgtable.h pte_t orig_pte) orig_pte 428 include/linux/mm.h pte_t orig_pte; /* Value of PTE at the time of fault */ orig_pte 913 mm/khugepaged.c vmf.orig_pte = *vmf.pte; orig_pte 914 mm/khugepaged.c if (!is_swap_pte(vmf.orig_pte)) orig_pte 1033 mm/ksm.c pte_t *orig_pte) orig_pte 1099 mm/ksm.c *orig_pte = *pvmw.pte; orig_pte 1120 mm/ksm.c struct page *kpage, pte_t orig_pte) orig_pte 1144 mm/ksm.c if (!pte_same(*ptep, orig_pte)) { orig_pte 1204 mm/ksm.c pte_t orig_pte = __pte(0); orig_pte 1234 mm/ksm.c if (write_protect_page(vma, page, &orig_pte) == 0) { orig_pte 1251 mm/ksm.c err = replace_page(vma, page, kpage, orig_pte); orig_pte 186 mm/madvise.c pte_t *orig_pte; orig_pte 199 mm/madvise.c orig_pte = pte_offset_map_lock(vma->vm_mm, pmd, start, &ptl); orig_pte 200 mm/madvise.c pte = *(orig_pte + ((index - start) / PAGE_SIZE)); orig_pte 201 mm/madvise.c pte_unmap_unlock(orig_pte, ptl); orig_pte 309 mm/madvise.c pte_t *orig_pte, *pte, ptent; orig_pte 388 mm/madvise.c orig_pte = pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); orig_pte 416 mm/madvise.c pte_unmap_unlock(orig_pte, ptl); orig_pte 465 mm/madvise.c pte_unmap_unlock(orig_pte, ptl); orig_pte 570 mm/madvise.c pte_t *orig_pte, *pte, ptent; orig_pte 584 mm/madvise.c orig_pte = pte = pte_offset_map_lock(mm, pmd, addr, &ptl); orig_pte 626 mm/madvise.c pte_unmap_unlock(orig_pte, ptl); orig_pte 689 mm/madvise.c pte_unmap_unlock(orig_pte, ptl); orig_pte 2133 mm/memory.c pte_t *page_table, pte_t orig_pte) orig_pte 2140 mm/memory.c same = pte_same(*page_table, orig_pte); orig_pte 2297 mm/memory.c flush_cache_page(vma, vmf->address, pte_pfn(vmf->orig_pte)); orig_pte 2298 mm/memory.c entry = pte_mkyoung(vmf->orig_pte); orig_pte 2335 mm/memory.c if (is_zero_pfn(pte_pfn(vmf->orig_pte))) { orig_pte 2362 mm/memory.c if (likely(pte_same(*vmf->pte, vmf->orig_pte))) { orig_pte 2372 mm/memory.c flush_cache_page(vma, vmf->address, pte_pfn(vmf->orig_pte)); orig_pte 2481 mm/memory.c if (!pte_same(*vmf->pte, vmf->orig_pte)) { orig_pte 2568 mm/memory.c vmf->page = vm_normal_page(vma, vmf->address, vmf->orig_pte); orig_pte 2600 mm/memory.c if (!pte_same(*vmf->pte, vmf->orig_pte)) { orig_pte 2762 mm/memory.c if (!pte_unmap_same(vma->vm_mm, vmf->pmd, vmf->pte, vmf->orig_pte)) orig_pte 2765 mm/memory.c entry = pte_to_swp_entry(vmf->orig_pte); orig_pte 2776 mm/memory.c print_bad_pte(vma, vmf->address, vmf->orig_pte, NULL); orig_pte 2815 mm/memory.c if (likely(pte_same(*vmf->pte, vmf->orig_pte))) orig_pte 2871 mm/memory.c if (unlikely(!pte_same(*vmf->pte, vmf->orig_pte))) orig_pte 2899 mm/memory.c if (pte_swp_soft_dirty(vmf->orig_pte)) orig_pte 2902 mm/memory.c arch_do_swap_page(vma->vm_mm, vma, vmf->address, pte, vmf->orig_pte); orig_pte 2903 mm/memory.c vmf->orig_pte = pte; orig_pte 3672 mm/memory.c bool was_writable = pte_savedwrite(vmf->orig_pte); orig_pte 3682 mm/memory.c if (unlikely(!pte_same(*vmf->pte, vmf->orig_pte))) { orig_pte 3844 mm/memory.c vmf->orig_pte = *vmf->pte; orig_pte 3855 mm/memory.c if (pte_none(vmf->orig_pte)) { orig_pte 3868 mm/memory.c if (!pte_present(vmf->orig_pte)) orig_pte 3871 mm/memory.c if (pte_protnone(vmf->orig_pte) && vma_is_accessible(vmf->vma)) orig_pte 3876 mm/memory.c entry = vmf->orig_pte; orig_pte 648 mm/swap_state.c pte_t *pte, *orig_pte; orig_pte 662 mm/swap_state.c orig_pte = pte = pte_offset_map(vmf->pmd, faddr); orig_pte 665 mm/swap_state.c pte_unmap(orig_pte); orig_pte 680 mm/swap_state.c pte_unmap(orig_pte); orig_pte 705 mm/swap_state.c pte_unmap(orig_pte);