Lines Matching refs:gpa
778 gpa_t gpa = addr + (vm_start - memslot->userspace_addr); in stage2_unmap_memslot() local
779 unmap_stage2_range(kvm, gpa, vm_end - vm_start); in stage2_unmap_memslot()
1469 gpa_t gpa, void *data), in handle_hva_to_gpa() argument
1497 gpa_t gpa = gfn << PAGE_SHIFT; in handle_hva_to_gpa() local
1498 ret |= handler(kvm, gpa, data); in handle_hva_to_gpa()
1505 static int kvm_unmap_hva_handler(struct kvm *kvm, gpa_t gpa, void *data) in kvm_unmap_hva_handler() argument
1507 unmap_stage2_range(kvm, gpa, PAGE_SIZE); in kvm_unmap_hva_handler()
1534 static int kvm_set_spte_handler(struct kvm *kvm, gpa_t gpa, void *data) in kvm_set_spte_handler() argument
1545 stage2_set_pte(kvm, NULL, gpa, pte, 0); in kvm_set_spte_handler()
1563 static int kvm_age_hva_handler(struct kvm *kvm, gpa_t gpa, void *data) in kvm_age_hva_handler() argument
1568 pmd = stage2_get_pmd(kvm, NULL, gpa); in kvm_age_hva_handler()
1581 pte = pte_offset_kernel(pmd, gpa); in kvm_age_hva_handler()
1593 static int kvm_test_age_hva_handler(struct kvm *kvm, gpa_t gpa, void *data) in kvm_test_age_hva_handler() argument
1598 pmd = stage2_get_pmd(kvm, NULL, gpa); in kvm_test_age_hva_handler()
1605 pte = pte_offset_kernel(pmd, gpa); in kvm_test_age_hva_handler()
1795 gpa_t gpa = mem->guest_phys_addr + in kvm_arch_prepare_memory_region() local
1806 ret = kvm_phys_addr_ioremap(kvm, gpa, pa, in kvm_arch_prepare_memory_region()
1858 gpa_t gpa = slot->base_gfn << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot() local
1862 unmap_stage2_range(kvm, gpa, size); in kvm_arch_flush_shadow_memslot()