write_fault       859 arch/mips/include/asm/kvm_host.h 				      struct kvm_vcpu *vcpu, bool write_fault);
write_fault       863 arch/mips/include/asm/kvm_host.h 					   bool write_fault);
write_fault       871 arch/mips/include/asm/kvm_host.h 						bool write_fault);
write_fault       877 arch/mips/include/asm/kvm_host.h 						     bool write_fault);
write_fault      2761 arch/mips/kvm/emulate.c 					      bool write_fault)
write_fault      2818 arch/mips/kvm/emulate.c 								 write_fault)) {
write_fault       611 arch/mips/kvm/mmu.c 				   bool write_fault,
write_fault       637 arch/mips/kvm/mmu.c 	if (write_fault && !pte_dirty(*ptep)) {
write_fault       687 arch/mips/kvm/mmu.c 			     bool write_fault,
write_fault       702 arch/mips/kvm/mmu.c 	err = _kvm_mips_map_page_fast(vcpu, gpa, write_fault, out_entry,
write_fault       733 arch/mips/kvm/mmu.c 	pfn = gfn_to_pfn_prot(kvm, gfn, write_fault, &writeable);
write_fault       759 arch/mips/kvm/mmu.c 		if (write_fault) {
write_fault       989 arch/mips/kvm/mmu.c 				      bool write_fault)
write_fault       993 arch/mips/kvm/mmu.c 	ret = kvm_mips_map_page(vcpu, badvaddr, write_fault, NULL, NULL);
write_fault      1005 arch/mips/kvm/mmu.c 				    bool write_fault)
write_fault      1020 arch/mips/kvm/mmu.c 	if (kvm_mips_map_page(vcpu, gpa, write_fault, &pte_gpa[idx],
write_fault      1043 arch/mips/kvm/mmu.c 					 bool write_fault)
write_fault      1063 arch/mips/kvm/mmu.c 			      write_fault, &pte_gpa[idx], NULL) < 0)
write_fault      3154 arch/x86/kvm/mmu.c 			int write_fault, int level, gfn_t gfn, kvm_pfn_t pfn,
write_fault      3164 arch/x86/kvm/mmu.c 		 *sptep, write_fault, gfn);
write_fault      3191 arch/x86/kvm/mmu.c 		if (write_fault)
write_fault       223 arch/x86/kvm/paging_tmpl.h 					     int write_fault)
write_fault       244 arch/x86/kvm/paging_tmpl.h 		if (level == walker->level && write_fault &&
write_fault       311 arch/x86/kvm/paging_tmpl.h 	const int write_fault = access & PFERR_WRITE_MASK;
write_fault       433 arch/x86/kvm/paging_tmpl.h 	if (!write_fault)
write_fault       445 arch/x86/kvm/paging_tmpl.h 		ret = FNAME(update_accessed_dirty_bits)(vcpu, mmu, walker, write_fault);
write_fault       457 arch/x86/kvm/paging_tmpl.h 	errcode |= write_fault | user_fault;
write_fault       482 arch/x86/kvm/paging_tmpl.h 		if (write_fault)
write_fault       616 arch/x86/kvm/paging_tmpl.h 			 int write_fault, int hlevel,
write_fault       704 arch/x86/kvm/paging_tmpl.h 	ret = mmu_set_spte(vcpu, it.sptep, gw->pte_access, write_fault,
write_fault       771 arch/x86/kvm/paging_tmpl.h 	int write_fault = error_code & PFERR_WRITE_MASK;
write_fault       833 arch/x86/kvm/paging_tmpl.h 	if (try_async_pf(vcpu, prefault, walker.gfn, addr, &pfn, write_fault,
write_fault       844 arch/x86/kvm/paging_tmpl.h 	if (write_fault && !(walker.pte_access & ACC_WRITE_MASK) &&
write_fault       870 arch/x86/kvm/paging_tmpl.h 	r = FNAME(fetch)(vcpu, addr, &walker, write_fault,
write_fault      1142 fs/xfs/xfs_file.c 	bool			write_fault)
write_fault      1148 fs/xfs/xfs_file.c 	trace_xfs_filemap_fault(ip, pe_size, write_fault);
write_fault      1150 fs/xfs/xfs_file.c 	if (write_fault) {
write_fault      1163 fs/xfs/xfs_file.c 		if (write_fault)
write_fault      1170 fs/xfs/xfs_file.c 	if (write_fault)
write_fault       660 fs/xfs/xfs_trace.h 		 bool write_fault),
write_fault       661 fs/xfs/xfs_trace.h 	TP_ARGS(ip, pe_size, write_fault),
write_fault       666 fs/xfs/xfs_trace.h 		__field(bool, write_fault)
write_fault       672 fs/xfs/xfs_trace.h 		__entry->write_fault = write_fault;
write_fault       681 fs/xfs/xfs_trace.h 		  __entry->write_fault)
write_fault       717 include/linux/kvm_host.h kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault,
write_fault       722 include/linux/kvm_host.h 			       bool atomic, bool *async, bool write_fault,
write_fault       224 mm/hmm.c       			    bool write_fault, uint64_t *pfn)
write_fault       237 mm/hmm.c       	if (write_fault)
write_fault       284 mm/hmm.c       			      bool fault, bool write_fault,
write_fault       295 mm/hmm.c       	if (write_fault && walk->vma && !(walk->vma->vm_flags & VM_WRITE))
write_fault       300 mm/hmm.c       		if (fault || write_fault) {
write_fault       303 mm/hmm.c       			ret = hmm_vma_do_fault(walk, addr, write_fault,
write_fault       310 mm/hmm.c       	return (fault || write_fault) ? -EBUSY : 0;
write_fault       315 mm/hmm.c       				      bool *fault, bool *write_fault)
write_fault       341 mm/hmm.c       			*write_fault = pfns & range->flags[HMM_PFN_WRITE];
write_fault       352 mm/hmm.c       		*write_fault = true;
write_fault       360 mm/hmm.c       				 bool *write_fault)
write_fault       365 mm/hmm.c       		*fault = *write_fault = false;
write_fault       369 mm/hmm.c       	*fault = *write_fault = false;
write_fault       372 mm/hmm.c       				   fault, write_fault);
write_fault       373 mm/hmm.c       		if ((*write_fault))
write_fault       383 mm/hmm.c       	bool fault, write_fault;
write_fault       391 mm/hmm.c       			     0, &fault, &write_fault);
write_fault       392 mm/hmm.c       	return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk);
write_fault       411 mm/hmm.c       	bool fault, write_fault;
write_fault       417 mm/hmm.c       			     &fault, &write_fault);
write_fault       419 mm/hmm.c       	if (pmd_protnone(pmd) || fault || write_fault)
write_fault       420 mm/hmm.c       		return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk);
write_fault       460 mm/hmm.c       	bool fault, write_fault;
write_fault       466 mm/hmm.c       	fault = write_fault = false;
write_fault       470 mm/hmm.c       				   &fault, &write_fault);
write_fault       471 mm/hmm.c       		if (fault || write_fault)
write_fault       482 mm/hmm.c       					   &fault, &write_fault);
write_fault       483 mm/hmm.c       			if (fault || write_fault)
write_fault       498 mm/hmm.c       					   &fault, &write_fault);
write_fault       499 mm/hmm.c       			if (fault || write_fault)
write_fault       508 mm/hmm.c       			if (fault || write_fault) {
write_fault       523 mm/hmm.c       				   &fault, &write_fault);
write_fault       526 mm/hmm.c       	if (fault || write_fault)
write_fault       549 mm/hmm.c       	return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk);
write_fault       570 mm/hmm.c       		bool fault, write_fault;
write_fault       579 mm/hmm.c       				     0, &fault, &write_fault);
write_fault       580 mm/hmm.c       		if (fault || write_fault) {
write_fault       674 mm/hmm.c       		bool fault, write_fault;
write_fault       685 mm/hmm.c       				     cpu_flags, &fault, &write_fault);
write_fault       686 mm/hmm.c       		if (fault || write_fault)
write_fault       688 mm/hmm.c       						write_fault, walk);
write_fault       735 mm/hmm.c       	bool fault, write_fault;
write_fault       747 mm/hmm.c       	fault = write_fault = false;
write_fault       749 mm/hmm.c       			   &fault, &write_fault);
write_fault       750 mm/hmm.c       	if (fault || write_fault) {
write_fault       765 mm/hmm.c       		return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk);
write_fault      1674 virt/kvm/arm/mmu.c 	bool write_fault, writable, force_pte = false;
write_fault      1686 virt/kvm/arm/mmu.c 	write_fault = kvm_is_write_fault(vcpu);
write_fault      1688 virt/kvm/arm/mmu.c 	VM_BUG_ON(write_fault && exec_fault);
write_fault      1690 virt/kvm/arm/mmu.c 	if (fault_status == FSC_PERM && !write_fault && !exec_fault) {
write_fault      1742 virt/kvm/arm/mmu.c 	pfn = gfn_to_pfn_prot(kvm, gfn, write_fault, &writable);
write_fault      1765 virt/kvm/arm/mmu.c 		if (!write_fault)
write_fault      1914 virt/kvm/arm/mmu.c 	bool is_iabt, write_fault, writable;
write_fault      1956 virt/kvm/arm/mmu.c 	write_fault = kvm_is_write_fault(vcpu);
write_fault      1957 virt/kvm/arm/mmu.c 	if (kvm_is_error_hva(hva) || (write_fault && !writable)) {
write_fault      1518 virt/kvm/kvm_main.c static bool hva_to_pfn_fast(unsigned long addr, bool write_fault,
write_fault      1529 virt/kvm/kvm_main.c 	if (!(write_fault || writable))
write_fault      1548 virt/kvm/kvm_main.c static int hva_to_pfn_slow(unsigned long addr, bool *async, bool write_fault,
write_fault      1558 virt/kvm/kvm_main.c 		*writable = write_fault;
write_fault      1560 virt/kvm/kvm_main.c 	if (write_fault)
write_fault      1570 virt/kvm/kvm_main.c 	if (unlikely(!write_fault) && writable) {
write_fault      1583 virt/kvm/kvm_main.c static bool vma_is_valid(struct vm_area_struct *vma, bool write_fault)
write_fault      1588 virt/kvm/kvm_main.c 	if (write_fault && (unlikely(!(vma->vm_flags & VM_WRITE))))
write_fault      1596 virt/kvm/kvm_main.c 			       bool write_fault, bool *writable,
write_fault      1610 virt/kvm/kvm_main.c 				     (write_fault ? FAULT_FLAG_WRITE : 0),
write_fault      1658 virt/kvm/kvm_main.c 			bool write_fault, bool *writable)
write_fault      1667 virt/kvm/kvm_main.c 	if (hva_to_pfn_fast(addr, write_fault, writable, &pfn))
write_fault      1673 virt/kvm/kvm_main.c 	npages = hva_to_pfn_slow(addr, async, write_fault, writable, &pfn);
write_fault      1690 virt/kvm/kvm_main.c 		r = hva_to_pfn_remapped(vma, addr, async, write_fault, writable, &pfn);
write_fault      1696 virt/kvm/kvm_main.c 		if (async && vma_is_valid(vma, write_fault))
write_fault      1706 virt/kvm/kvm_main.c 			       bool atomic, bool *async, bool write_fault,
write_fault      1709 virt/kvm/kvm_main.c 	unsigned long addr = __gfn_to_hva_many(slot, gfn, NULL, write_fault);
write_fault      1729 virt/kvm/kvm_main.c 	return hva_to_pfn(addr, atomic, async, write_fault,
write_fault      1734 virt/kvm/kvm_main.c kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault,
write_fault      1738 virt/kvm/kvm_main.c 				    write_fault, writable);