Lines Matching refs:pte

86 static void kvm_flush_dcache_pte(pte_t pte)  in kvm_flush_dcache_pte()  argument
88 __kvm_flush_dcache_pte(pte); in kvm_flush_dcache_pte()
210 pte_t *pte, *start_pte; in unmap_ptes() local
212 start_pte = pte = pte_offset_kernel(pmd, addr); in unmap_ptes()
214 if (!pte_none(*pte)) { in unmap_ptes()
215 pte_t old_pte = *pte; in unmap_ptes()
217 kvm_set_pte(pte, __pte(0)); in unmap_ptes()
224 put_page(virt_to_page(pte)); in unmap_ptes()
226 } while (pte++, addr += PAGE_SIZE, addr != end); in unmap_ptes()
309 pte_t *pte; in stage2_flush_ptes() local
311 pte = pte_offset_kernel(pmd, addr); in stage2_flush_ptes()
313 if (!pte_none(*pte) && !kvm_is_device_pfn(pte_pfn(*pte))) in stage2_flush_ptes()
314 kvm_flush_dcache_pte(*pte); in stage2_flush_ptes()
315 } while (pte++, addr += PAGE_SIZE, addr != end); in stage2_flush_ptes()
455 pte_t *pte; in create_hyp_pte_mappings() local
460 pte = pte_offset_kernel(pmd, addr); in create_hyp_pte_mappings()
461 kvm_set_pte(pte, pfn_pte(pfn, prot)); in create_hyp_pte_mappings()
462 get_page(virt_to_page(pte)); in create_hyp_pte_mappings()
463 kvm_flush_dcache_to_poc(pte, sizeof(*pte)); in create_hyp_pte_mappings()
473 pte_t *pte; in create_hyp_pmd_mappings() local
483 pte = pte_alloc_one_kernel(NULL, addr); in create_hyp_pmd_mappings()
484 if (!pte) { in create_hyp_pmd_mappings()
488 pmd_populate_kernel(NULL, pmd, pte); in create_hyp_pmd_mappings()
905 pte_t *pte, old_pte; in stage2_set_pte() local
932 pte = mmu_memory_cache_alloc(cache); in stage2_set_pte()
933 kvm_clean_pte(pte); in stage2_set_pte()
934 pmd_populate_kernel(NULL, pmd, pte); in stage2_set_pte()
938 pte = pte_offset_kernel(pmd, addr); in stage2_set_pte()
940 if (iomap && pte_present(*pte)) in stage2_set_pte()
944 old_pte = *pte; in stage2_set_pte()
946 kvm_set_pte(pte, __pte(0)); in stage2_set_pte()
949 get_page(virt_to_page(pte)); in stage2_set_pte()
952 kvm_set_pte(pte, *new_pte); in stage2_set_pte()
976 pte_t pte = pfn_pte(pfn, PAGE_S2_DEVICE); in kvm_phys_addr_ioremap() local
979 kvm_set_s2pte_writable(&pte); in kvm_phys_addr_ioremap()
986 ret = stage2_set_pte(kvm, &cache, addr, &pte, in kvm_phys_addr_ioremap()
1057 pte_t *pte; in stage2_wp_ptes() local
1059 pte = pte_offset_kernel(pmd, addr); in stage2_wp_ptes()
1061 if (!pte_none(*pte)) { in stage2_wp_ptes()
1062 if (!kvm_s2pte_readonly(pte)) in stage2_wp_ptes()
1063 kvm_set_s2pte_readonly(pte); in stage2_wp_ptes()
1065 } while (pte++, addr += PAGE_SIZE, addr != end); in stage2_wp_ptes()
1352 pte_t *pte; in handle_access_fault() local
1371 pte = pte_offset_kernel(pmd, fault_ipa); in handle_access_fault()
1372 if (pte_none(*pte)) /* Nothing there either */ in handle_access_fault()
1375 *pte = pte_mkyoung(*pte); /* Just a page... */ in handle_access_fault()
1376 pfn = pte_pfn(*pte); in handle_access_fault()
1536 pte_t *pte = (pte_t *)data; in kvm_set_spte_handler() local
1545 stage2_set_pte(kvm, NULL, gpa, pte, 0); in kvm_set_spte_handler()
1550 void kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte) in kvm_set_spte_hva() argument
1559 stage2_pte = pfn_pte(pte_pfn(pte), PAGE_S2); in kvm_set_spte_hva()
1566 pte_t *pte; in kvm_age_hva_handler() local
1581 pte = pte_offset_kernel(pmd, gpa); in kvm_age_hva_handler()
1582 if (pte_none(*pte)) in kvm_age_hva_handler()
1585 if (pte_young(*pte)) { in kvm_age_hva_handler()
1586 *pte = pte_mkold(*pte); /* Just a page... */ in kvm_age_hva_handler()
1596 pte_t *pte; in kvm_test_age_hva_handler() local
1605 pte = pte_offset_kernel(pmd, gpa); in kvm_test_age_hva_handler()
1606 if (!pte_none(*pte)) /* Just a page... */ in kvm_test_age_hva_handler()
1607 return pte_young(*pte); in kvm_test_age_hva_handler()