write_fault 859 arch/mips/include/asm/kvm_host.h struct kvm_vcpu *vcpu, bool write_fault); write_fault 863 arch/mips/include/asm/kvm_host.h bool write_fault); write_fault 871 arch/mips/include/asm/kvm_host.h bool write_fault); write_fault 877 arch/mips/include/asm/kvm_host.h bool write_fault); write_fault 2761 arch/mips/kvm/emulate.c bool write_fault) write_fault 2818 arch/mips/kvm/emulate.c write_fault)) { write_fault 611 arch/mips/kvm/mmu.c bool write_fault, write_fault 637 arch/mips/kvm/mmu.c if (write_fault && !pte_dirty(*ptep)) { write_fault 687 arch/mips/kvm/mmu.c bool write_fault, write_fault 702 arch/mips/kvm/mmu.c err = _kvm_mips_map_page_fast(vcpu, gpa, write_fault, out_entry, write_fault 733 arch/mips/kvm/mmu.c pfn = gfn_to_pfn_prot(kvm, gfn, write_fault, &writeable); write_fault 759 arch/mips/kvm/mmu.c if (write_fault) { write_fault 989 arch/mips/kvm/mmu.c bool write_fault) write_fault 993 arch/mips/kvm/mmu.c ret = kvm_mips_map_page(vcpu, badvaddr, write_fault, NULL, NULL); write_fault 1005 arch/mips/kvm/mmu.c bool write_fault) write_fault 1020 arch/mips/kvm/mmu.c if (kvm_mips_map_page(vcpu, gpa, write_fault, &pte_gpa[idx], write_fault 1043 arch/mips/kvm/mmu.c bool write_fault) write_fault 1063 arch/mips/kvm/mmu.c write_fault, &pte_gpa[idx], NULL) < 0) write_fault 3154 arch/x86/kvm/mmu.c int write_fault, int level, gfn_t gfn, kvm_pfn_t pfn, write_fault 3164 arch/x86/kvm/mmu.c *sptep, write_fault, gfn); write_fault 3191 arch/x86/kvm/mmu.c if (write_fault) write_fault 223 arch/x86/kvm/paging_tmpl.h int write_fault) write_fault 244 arch/x86/kvm/paging_tmpl.h if (level == walker->level && write_fault && write_fault 311 arch/x86/kvm/paging_tmpl.h const int write_fault = access & PFERR_WRITE_MASK; write_fault 433 arch/x86/kvm/paging_tmpl.h if (!write_fault) write_fault 445 arch/x86/kvm/paging_tmpl.h ret = FNAME(update_accessed_dirty_bits)(vcpu, mmu, walker, write_fault); write_fault 457 arch/x86/kvm/paging_tmpl.h errcode |= write_fault | user_fault; write_fault 482 arch/x86/kvm/paging_tmpl.h if (write_fault) write_fault 616 arch/x86/kvm/paging_tmpl.h int write_fault, int hlevel, write_fault 704 arch/x86/kvm/paging_tmpl.h ret = mmu_set_spte(vcpu, it.sptep, gw->pte_access, write_fault, write_fault 771 arch/x86/kvm/paging_tmpl.h int write_fault = error_code & PFERR_WRITE_MASK; write_fault 833 arch/x86/kvm/paging_tmpl.h if (try_async_pf(vcpu, prefault, walker.gfn, addr, &pfn, write_fault, write_fault 844 arch/x86/kvm/paging_tmpl.h if (write_fault && !(walker.pte_access & ACC_WRITE_MASK) && write_fault 870 arch/x86/kvm/paging_tmpl.h r = FNAME(fetch)(vcpu, addr, &walker, write_fault, write_fault 1142 fs/xfs/xfs_file.c bool write_fault) write_fault 1148 fs/xfs/xfs_file.c trace_xfs_filemap_fault(ip, pe_size, write_fault); write_fault 1150 fs/xfs/xfs_file.c if (write_fault) { write_fault 1163 fs/xfs/xfs_file.c if (write_fault) write_fault 1170 fs/xfs/xfs_file.c if (write_fault) write_fault 660 fs/xfs/xfs_trace.h bool write_fault), write_fault 661 fs/xfs/xfs_trace.h TP_ARGS(ip, pe_size, write_fault), write_fault 666 fs/xfs/xfs_trace.h __field(bool, write_fault) write_fault 672 fs/xfs/xfs_trace.h __entry->write_fault = write_fault; write_fault 681 fs/xfs/xfs_trace.h __entry->write_fault) write_fault 717 include/linux/kvm_host.h kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, write_fault 722 include/linux/kvm_host.h bool atomic, bool *async, bool write_fault, write_fault 224 mm/hmm.c bool write_fault, uint64_t *pfn) write_fault 237 mm/hmm.c if (write_fault) write_fault 284 mm/hmm.c bool fault, bool write_fault, write_fault 295 mm/hmm.c if (write_fault && walk->vma && !(walk->vma->vm_flags & VM_WRITE)) write_fault 300 mm/hmm.c if (fault || write_fault) { write_fault 303 mm/hmm.c ret = hmm_vma_do_fault(walk, addr, write_fault, write_fault 310 mm/hmm.c return (fault || write_fault) ? -EBUSY : 0; write_fault 315 mm/hmm.c bool *fault, bool *write_fault) write_fault 341 mm/hmm.c *write_fault = pfns & range->flags[HMM_PFN_WRITE]; write_fault 352 mm/hmm.c *write_fault = true; write_fault 360 mm/hmm.c bool *write_fault) write_fault 365 mm/hmm.c *fault = *write_fault = false; write_fault 369 mm/hmm.c *fault = *write_fault = false; write_fault 372 mm/hmm.c fault, write_fault); write_fault 373 mm/hmm.c if ((*write_fault)) write_fault 383 mm/hmm.c bool fault, write_fault; write_fault 391 mm/hmm.c 0, &fault, &write_fault); write_fault 392 mm/hmm.c return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk); write_fault 411 mm/hmm.c bool fault, write_fault; write_fault 417 mm/hmm.c &fault, &write_fault); write_fault 419 mm/hmm.c if (pmd_protnone(pmd) || fault || write_fault) write_fault 420 mm/hmm.c return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk); write_fault 460 mm/hmm.c bool fault, write_fault; write_fault 466 mm/hmm.c fault = write_fault = false; write_fault 470 mm/hmm.c &fault, &write_fault); write_fault 471 mm/hmm.c if (fault || write_fault) write_fault 482 mm/hmm.c &fault, &write_fault); write_fault 483 mm/hmm.c if (fault || write_fault) write_fault 498 mm/hmm.c &fault, &write_fault); write_fault 499 mm/hmm.c if (fault || write_fault) write_fault 508 mm/hmm.c if (fault || write_fault) { write_fault 523 mm/hmm.c &fault, &write_fault); write_fault 526 mm/hmm.c if (fault || write_fault) write_fault 549 mm/hmm.c return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk); write_fault 570 mm/hmm.c bool fault, write_fault; write_fault 579 mm/hmm.c 0, &fault, &write_fault); write_fault 580 mm/hmm.c if (fault || write_fault) { write_fault 674 mm/hmm.c bool fault, write_fault; write_fault 685 mm/hmm.c cpu_flags, &fault, &write_fault); write_fault 686 mm/hmm.c if (fault || write_fault) write_fault 688 mm/hmm.c write_fault, walk); write_fault 735 mm/hmm.c bool fault, write_fault; write_fault 747 mm/hmm.c fault = write_fault = false; write_fault 749 mm/hmm.c &fault, &write_fault); write_fault 750 mm/hmm.c if (fault || write_fault) { write_fault 765 mm/hmm.c return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk); write_fault 1674 virt/kvm/arm/mmu.c bool write_fault, writable, force_pte = false; write_fault 1686 virt/kvm/arm/mmu.c write_fault = kvm_is_write_fault(vcpu); write_fault 1688 virt/kvm/arm/mmu.c VM_BUG_ON(write_fault && exec_fault); write_fault 1690 virt/kvm/arm/mmu.c if (fault_status == FSC_PERM && !write_fault && !exec_fault) { write_fault 1742 virt/kvm/arm/mmu.c pfn = gfn_to_pfn_prot(kvm, gfn, write_fault, &writable); write_fault 1765 virt/kvm/arm/mmu.c if (!write_fault) write_fault 1914 virt/kvm/arm/mmu.c bool is_iabt, write_fault, writable; write_fault 1956 virt/kvm/arm/mmu.c write_fault = kvm_is_write_fault(vcpu); write_fault 1957 virt/kvm/arm/mmu.c if (kvm_is_error_hva(hva) || (write_fault && !writable)) { write_fault 1518 virt/kvm/kvm_main.c static bool hva_to_pfn_fast(unsigned long addr, bool write_fault, write_fault 1529 virt/kvm/kvm_main.c if (!(write_fault || writable)) write_fault 1548 virt/kvm/kvm_main.c static int hva_to_pfn_slow(unsigned long addr, bool *async, bool write_fault, write_fault 1558 virt/kvm/kvm_main.c *writable = write_fault; write_fault 1560 virt/kvm/kvm_main.c if (write_fault) write_fault 1570 virt/kvm/kvm_main.c if (unlikely(!write_fault) && writable) { write_fault 1583 virt/kvm/kvm_main.c static bool vma_is_valid(struct vm_area_struct *vma, bool write_fault) write_fault 1588 virt/kvm/kvm_main.c if (write_fault && (unlikely(!(vma->vm_flags & VM_WRITE)))) write_fault 1596 virt/kvm/kvm_main.c bool write_fault, bool *writable, write_fault 1610 virt/kvm/kvm_main.c (write_fault ? FAULT_FLAG_WRITE : 0), write_fault 1658 virt/kvm/kvm_main.c bool write_fault, bool *writable) write_fault 1667 virt/kvm/kvm_main.c if (hva_to_pfn_fast(addr, write_fault, writable, &pfn)) write_fault 1673 virt/kvm/kvm_main.c npages = hva_to_pfn_slow(addr, async, write_fault, writable, &pfn); write_fault 1690 virt/kvm/kvm_main.c r = hva_to_pfn_remapped(vma, addr, async, write_fault, writable, &pfn); write_fault 1696 virt/kvm/kvm_main.c if (async && vma_is_valid(vma, write_fault)) write_fault 1706 virt/kvm/kvm_main.c bool atomic, bool *async, bool write_fault, write_fault 1709 virt/kvm/kvm_main.c unsigned long addr = __gfn_to_hva_many(slot, gfn, NULL, write_fault); write_fault 1729 virt/kvm/kvm_main.c return hva_to_pfn(addr, atomic, async, write_fault, write_fault 1734 virt/kvm/kvm_main.c kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, write_fault 1738 virt/kvm/kvm_main.c write_fault, writable);