/linux-4.1.27/arch/s390/kvm/ |
D | gaccess.h | 127 unsigned long gpa = gra + kvm_s390_get_prefix(vcpu); in write_guest_lc() local 129 return kvm_write_guest(vcpu->kvm, gpa, data, len); in write_guest_lc() 153 unsigned long gpa = gra + kvm_s390_get_prefix(vcpu); in read_guest_lc() local 155 return kvm_read_guest(vcpu->kvm, gpa, data, len); in read_guest_lc() 159 ar_t ar, unsigned long *gpa, int write); 258 int write_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, in write_guest_abs() argument 261 return kvm_write_guest(vcpu->kvm, gpa, data, len); in write_guest_abs() 281 int read_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, in read_guest_abs() argument 284 return kvm_read_guest(vcpu->kvm, gpa, data, len); in read_guest_abs()
|
D | gaccess.c | 514 static int deref_table(struct kvm *kvm, unsigned long gpa, unsigned long *val) in deref_table() argument 516 return kvm_read_guest(kvm, gpa, val, sizeof(*val)); in deref_table() 540 unsigned long *gpa, const union asce asce, in guest_translate() argument 690 *gpa = raddr.addr; in guest_translate() 758 unsigned long _len, nr_pages, gpa, idx; in access_guest() local 781 gpa = *(pages + idx) + (ga & ~PAGE_MASK); in access_guest() 782 _len = min(PAGE_SIZE - (gpa & ~PAGE_MASK), len); in access_guest() 784 rc = kvm_write_guest(vcpu->kvm, gpa, data, _len); in access_guest() 786 rc = kvm_read_guest(vcpu->kvm, gpa, data, _len); in access_guest() 801 unsigned long _len, gpa; in access_guest_real() local [all …]
|
D | kvm-s390.c | 1787 long kvm_arch_fault_in_page(struct kvm_vcpu *vcpu, gpa_t gpa, int writable) in kvm_arch_fault_in_page() argument 1789 return gmap_fault(vcpu->arch.gmap, gpa, in kvm_arch_fault_in_page() 1961 gpa_t gpa = current->thread.gmap_addr; in vcpu_post_run() local 1962 rc = kvm_arch_fault_in_page(vcpu, gpa, 1); in vcpu_post_run() 2127 int kvm_s390_store_status_unloaded(struct kvm_vcpu *vcpu, unsigned long gpa) in kvm_s390_store_status_unloaded() argument 2134 if (gpa == KVM_S390_STORE_STATUS_NOADDR) { in kvm_s390_store_status_unloaded() 2137 gpa = SAVE_AREA_BASE; in kvm_s390_store_status_unloaded() 2138 } else if (gpa == KVM_S390_STORE_STATUS_PREFIXED) { in kvm_s390_store_status_unloaded() 2141 gpa = kvm_s390_real_to_abs(vcpu, SAVE_AREA_BASE); in kvm_s390_store_status_unloaded() 2143 rc = write_guest_abs(vcpu, gpa + offsetof(struct save_area, fp_regs), in kvm_s390_store_status_unloaded() [all …]
|
D | priv.c | 972 unsigned long hva, gpa; in handle_tprot() local 991 ret = guest_translate_address(vcpu, address1, ar, &gpa, 1); in handle_tprot() 995 ret = guest_translate_address(vcpu, address1, ar, &gpa, 0); in handle_tprot() 1008 hva = gfn_to_hva_prot(vcpu->kvm, gpa_to_gfn(gpa), &writable); in handle_tprot()
|
D | kvm-s390.h | 206 long kvm_arch_fault_in_page(struct kvm_vcpu *vcpu, gpa_t gpa, int writable);
|
/linux-4.1.27/arch/x86/include/asm/uv/ |
D | uv_hub.h | 330 uv_gpa_in_mmr_space(unsigned long gpa) in uv_gpa_in_mmr_space() argument 332 return (gpa >> 62) == 0x3UL; in uv_gpa_in_mmr_space() 336 static inline unsigned long uv_gpa_to_soc_phys_ram(unsigned long gpa) in uv_gpa_to_soc_phys_ram() argument 342 gpa = ((gpa << uv_hub_info->m_shift) >> uv_hub_info->m_shift) | in uv_gpa_to_soc_phys_ram() 343 ((gpa >> uv_hub_info->n_lshift) << uv_hub_info->m_val); in uv_gpa_to_soc_phys_ram() 344 paddr = gpa & uv_hub_info->gpa_mask; in uv_gpa_to_soc_phys_ram() 352 static inline unsigned long uv_gpa_to_gnode(unsigned long gpa) in uv_gpa_to_gnode() argument 354 return gpa >> uv_hub_info->n_lshift; in uv_gpa_to_gnode() 358 static inline int uv_gpa_to_pnode(unsigned long gpa) in uv_gpa_to_pnode() argument 362 return uv_gpa_to_gnode(gpa) & n_mask; in uv_gpa_to_pnode() [all …]
|
/linux-4.1.27/include/trace/events/ |
D | kvm.h | 207 TP_PROTO(int type, int len, u64 gpa, u64 val), 208 TP_ARGS(type, len, gpa, val), 213 __field( u64, gpa ) 220 __entry->gpa = gpa; 226 __entry->len, __entry->gpa, __entry->val)
|
/linux-4.1.27/drivers/misc/sgi-gru/ |
D | grufault.c | 262 int write, int atomic, unsigned long *gpa, int *pageshift) in gru_vtop() argument 288 *gpa = uv_soc_phys_ram_to_gpa(paddr); in gru_vtop() 325 unsigned long vaddr = 0, gpa; in gru_preload_tlb() local 341 ret = gru_vtop(gts, vaddr, write, atomic, &gpa, &pageshift); in gru_preload_tlb() 342 if (ret || tfh_write_only(tfh, gpa, GAA_RAM, vaddr, asid, write, in gru_preload_tlb() 348 vaddr, asid, write, pageshift, gpa); in gru_preload_tlb() 372 unsigned long gpa = 0, vaddr = 0; in gru_try_dropin() local 422 ret = gru_vtop(gts, vaddr, write, atomic, &gpa, &pageshift); in gru_try_dropin() 443 tfh_write_restart(tfh, gpa, GAA_RAM, vaddr, asid, write, in gru_try_dropin() 449 indexway, write, pageshift, gpa); in gru_try_dropin()
|
D | grukservices.h | 144 int gru_read_gpa(unsigned long *value, unsigned long gpa);
|
D | gru_instructions.h | 359 static inline void gru_vload_phys(void *cb, unsigned long gpa, in gru_vload_phys() argument 364 ins->baddr0 = (long)gpa | ((unsigned long)iaa << 62); in gru_vload_phys() 371 static inline void gru_vstore_phys(void *cb, unsigned long gpa, in gru_vstore_phys() argument 376 ins->baddr0 = (long)gpa | ((unsigned long)iaa << 62); in gru_vstore_phys()
|
D | grukservices.c | 716 unsigned long m, *val = mesg, gpa, save; in send_message_put_nacked() local 739 gpa = uv_global_gru_mmr_address(mqd->interrupt_pnode, UVH_IPI_INT); in send_message_put_nacked() 743 gru_vstore_phys(cb, gpa, gru_get_tri(mesg), IAA_REGISTER, IMA); in send_message_put_nacked() 906 int gru_read_gpa(unsigned long *value, unsigned long gpa) in gru_read_gpa() argument 915 iaa = gpa >> 62; in gru_read_gpa() 916 gru_vload_phys(cb, gpa, gru_get_tri(dsr), iaa, IMA); in gru_read_gpa()
|
/linux-4.1.27/arch/arm/kvm/ |
D | mmu.c | 778 gpa_t gpa = addr + (vm_start - memslot->userspace_addr); in stage2_unmap_memslot() local 779 unmap_stage2_range(kvm, gpa, vm_end - vm_start); in stage2_unmap_memslot() 1469 gpa_t gpa, void *data), in handle_hva_to_gpa() argument 1497 gpa_t gpa = gfn << PAGE_SHIFT; in handle_hva_to_gpa() local 1498 ret |= handler(kvm, gpa, data); in handle_hva_to_gpa() 1505 static int kvm_unmap_hva_handler(struct kvm *kvm, gpa_t gpa, void *data) in kvm_unmap_hva_handler() argument 1507 unmap_stage2_range(kvm, gpa, PAGE_SIZE); in kvm_unmap_hva_handler() 1534 static int kvm_set_spte_handler(struct kvm *kvm, gpa_t gpa, void *data) in kvm_set_spte_handler() argument 1545 stage2_set_pte(kvm, NULL, gpa, pte, 0); in kvm_set_spte_handler() 1563 static int kvm_age_hva_handler(struct kvm *kvm, gpa_t gpa, void *data) in kvm_age_hva_handler() argument [all …]
|
/linux-4.1.27/arch/arm/boot/dts/ |
D | s3c64xx-pinctrl.dtsi | 24 gpa: gpa { label 140 samsung,pins = "gpa-0", "gpa-1"; 146 samsung,pins = "gpa-2", "gpa-3"; 152 samsung,pins = "gpa-4", "gpa-5"; 158 samsung,pins = "gpa-6", "gpa-7";
|
D | s3c2416-pinctrl.dtsi | 16 gpa: gpa { label
|
/linux-4.1.27/arch/x86/kvm/ |
D | x86.c | 2077 gpa_t gpa = data & ~0x3f; in kvm_pv_enable_async_pf() local 2091 if (kvm_gfn_to_hva_cache_init(vcpu->kvm, &vcpu->arch.apf.data, gpa, in kvm_pv_enable_async_pf() 4198 gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, in translate_nested_gpa() argument 4207 t_gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, gpa, access, exception); in translate_nested_gpa() 4250 gpa_t gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, addr, access, in kvm_read_guest_virt_helper() local 4256 if (gpa == UNMAPPED_GVA) in kvm_read_guest_virt_helper() 4258 ret = kvm_read_guest_page(vcpu->kvm, gpa >> PAGE_SHIFT, data, in kvm_read_guest_virt_helper() 4284 gpa_t gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, addr, access|PFERR_FETCH_MASK, in kvm_fetch_guest_virt() local 4286 if (unlikely(gpa == UNMAPPED_GVA)) in kvm_fetch_guest_virt() 4292 ret = kvm_read_guest_page(vcpu->kvm, gpa >> PAGE_SHIFT, val, in kvm_fetch_guest_virt() [all …]
|
D | mmutrace.h | 103 __field(__u64, gpa) 107 __entry->gpa = ((u64)table_gfn << PAGE_SHIFT) 111 TP_printk("gpa %llx", __entry->gpa)
|
D | paging_tmpl.h | 889 gpa_t gpa = UNMAPPED_GVA; in FNAME() local 895 gpa = gfn_to_gpa(walker.gfn); in FNAME() 896 gpa |= vaddr & ~PAGE_MASK; in FNAME() 900 return gpa; in FNAME() 909 gpa_t gpa = UNMAPPED_GVA; in FNAME() local 915 gpa = gfn_to_gpa(walker.gfn); in FNAME() 916 gpa |= vaddr & ~PAGE_MASK; in FNAME() 920 return gpa; in FNAME()
|
D | x86.h | 122 static inline bool vcpu_match_mmio_gpa(struct kvm_vcpu *vcpu, gpa_t gpa) in vcpu_match_mmio_gpa() argument 125 vcpu->arch.mmio_gfn == gpa >> PAGE_SHIFT) in vcpu_match_mmio_gpa()
|
D | trace.h | 747 TP_PROTO(gva_t gva, gpa_t gpa, bool write, bool gpa_match), 748 TP_ARGS(gva, gpa, write, gpa_match), 752 __field(gpa_t, gpa) 759 __entry->gpa = gpa; 764 TP_printk("gva %#lx gpa %#llx %s %s", __entry->gva, __entry->gpa,
|
D | mmu.c | 3455 static int tdp_page_fault(struct kvm_vcpu *vcpu, gva_t gpa, u32 error_code, in tdp_page_fault() argument 3462 gfn_t gfn = gpa >> PAGE_SHIFT; in tdp_page_fault() 3470 r = handle_mmio_page_fault(vcpu, gpa, error_code, true); in tdp_page_fault() 3487 if (fast_page_fault(vcpu, gpa, level, error_code)) in tdp_page_fault() 3493 if (try_async_pf(vcpu, prefault, gfn, gpa, &pfn, write, &map_writable)) in tdp_page_fault() 3505 r = __direct_map(vcpu, gpa, write, map_writable, in tdp_page_fault() 4053 static u64 mmu_pte_write_fetch_gpte(struct kvm_vcpu *vcpu, gpa_t *gpa, in mmu_pte_write_fetch_gpte() argument 4066 *gpa &= ~(gpa_t)7; in mmu_pte_write_fetch_gpte() 4068 r = kvm_read_guest(vcpu->kvm, *gpa, &gentry, 8); in mmu_pte_write_fetch_gpte() 4109 static bool detect_write_misaligned(struct kvm_mmu_page *sp, gpa_t gpa, in detect_write_misaligned() argument [all …]
|
D | vmx.c | 1212 static inline void __invept(int ext, u64 eptp, gpa_t gpa) in __invept() argument 1215 u64 eptp, gpa; in __invept() member 1216 } operand = {eptp, gpa}; in __invept() 5748 gpa_t gpa; in handle_ept_violation() local 5778 gpa = vmcs_read64(GUEST_PHYSICAL_ADDRESS); in handle_ept_violation() 5779 trace_kvm_page_fault(gpa, exit_qualification); in handle_ept_violation() 5790 return kvm_mmu_page_fault(vcpu, gpa, error_code, NULL, 0); in handle_ept_violation() 5860 gpa_t gpa; in handle_ept_misconfig() local 5862 gpa = vmcs_read64(GUEST_PHYSICAL_ADDRESS); in handle_ept_misconfig() 5863 if (!kvm_io_bus_write(vcpu, KVM_FAST_MMIO_BUS, gpa, 0, NULL)) { in handle_ept_misconfig() [all …]
|
D | svm.c | 2111 static void *nested_svm_map(struct vcpu_svm *svm, u64 gpa, struct page **_page) in nested_svm_map() argument 2117 page = gfn_to_page(svm->vcpu.kvm, gpa >> PAGE_SHIFT); in nested_svm_map() 2142 u64 gpa; in nested_svm_intercept_ioio() local 2150 gpa = svm->nested.vmcb_iopm + (port / 8); in nested_svm_intercept_ioio() 2156 if (kvm_read_guest(svm->vcpu.kvm, gpa, &val, iopm_len)) in nested_svm_intercept_ioio()
|
/linux-4.1.27/drivers/misc/sgi-xp/ |
D | xp_sn2.c | 90 xp_socket_pa_sn2(unsigned long gpa) in xp_socket_pa_sn2() argument 92 return gpa; in xp_socket_pa_sn2()
|
D | xp_uv.c | 39 xp_socket_pa_uv(unsigned long gpa) in xp_socket_pa_uv() argument 41 return uv_gpa_to_soc_phys_ram(gpa); in xp_socket_pa_uv()
|
D | xp_main.c | 47 unsigned long (*xp_socket_pa) (unsigned long gpa);
|
/linux-4.1.27/include/linux/ |
D | kvm_host.h | 211 gpa_t gpa; member 567 int kvm_read_guest_atomic(struct kvm *kvm, gpa_t gpa, void *data, 569 int kvm_read_guest(struct kvm *kvm, gpa_t gpa, void *data, unsigned long len); 574 int kvm_write_guest(struct kvm *kvm, gpa_t gpa, const void *data, 579 gpa_t gpa, unsigned long len); 581 int kvm_clear_guest(struct kvm *kvm, gpa_t gpa, unsigned long len); 871 static inline gfn_t gpa_to_gfn(gpa_t gpa) in gpa_to_gfn() argument 873 return (gfn_t)(gpa >> PAGE_SHIFT); in gpa_to_gfn() 881 static inline bool kvm_is_error_gpa(struct kvm *kvm, gpa_t gpa) in kvm_is_error_gpa() argument 883 unsigned long hva = gfn_to_hva(kvm, gpa_to_gfn(gpa)); in kvm_is_error_gpa()
|
D | kvm_types.h | 59 gpa_t gpa; member
|
/linux-4.1.27/arch/ia64/include/asm/uv/ |
D | uv_hub.h | 166 static inline void *uv_va(unsigned long gpa) in uv_va() argument 168 return __va(gpa & uv_hub_info->gpa_mask); in uv_va()
|
/linux-4.1.27/arch/powerpc/kvm/ |
D | e500.h | 258 gpa_t gpa; in tlbe_is_host_safe() local 270 gpa = get_tlb_raddr(tlbe); in tlbe_is_host_safe() 271 if (!gfn_to_memslot(vcpu->kvm, gpa >> PAGE_SHIFT)) in tlbe_is_host_safe()
|
D | book3s_64_mmu_hv.c | 391 unsigned long gpa, gva_t ea, int is_store) in kvmppc_hv_emulate_mmio() argument 430 vcpu->arch.paddr_accessed = gpa; in kvmppc_hv_emulate_mmio() 443 unsigned long gpa, gfn, hva, pfn; in kvmppc_book3s_hv_page_fault() local 482 gpa = gpa_base | (ea & (psize - 1)); in kvmppc_book3s_hv_page_fault() 483 gfn = gpa >> PAGE_SHIFT; in kvmppc_book3s_hv_page_fault() 490 return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea, in kvmppc_book3s_hv_page_fault() 1011 gfn = vpa->gpa >> PAGE_SHIFT; in harvest_vpa_dirty() 1055 void *kvmppc_pin_guest_page(struct kvm *kvm, unsigned long gpa, in kvmppc_pin_guest_page() argument 1059 unsigned long gfn = gpa >> PAGE_SHIFT; in kvmppc_pin_guest_page() 1076 offset = gpa & (PAGE_SIZE - 1); in kvmppc_pin_guest_page() [all …]
|
D | book3s_hv_rm_mmu.c | 141 unsigned long i, pa, gpa, gfn, psize; in kvmppc_do_h_enter() local 168 gpa = (ptel & HPTE_R_RPN) & ~(psize - 1); in kvmppc_do_h_enter() 169 gfn = gpa >> PAGE_SHIFT; in kvmppc_do_h_enter() 224 pa |= gpa & ~PAGE_MASK; in kvmppc_do_h_enter()
|
D | book3s.c | 367 pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa, bool writing, in kvmppc_gpa_to_pfn() argument 371 gfn_t gfn = gpa >> PAGE_SHIFT; in kvmppc_gpa_to_pfn() 377 gpa &= ~0xFFFULL; in kvmppc_gpa_to_pfn() 378 if (unlikely(mp_pa) && unlikely((gpa & KVM_PAM) == mp_pa)) { in kvmppc_gpa_to_pfn()
|
D | book3s_hv.c | 470 unsigned long gpa; in kvmppc_update_vpa() local 481 gpa = vpap->next_gpa; in kvmppc_update_vpa() 485 if (gpa) in kvmppc_update_vpa() 486 va = kvmppc_pin_guest_page(kvm, gpa, &nb); in kvmppc_update_vpa() 488 if (gpa == vpap->next_gpa) in kvmppc_update_vpa() 492 kvmppc_unpin_guest_page(kvm, va, gpa, false); in kvmppc_update_vpa() 502 kvmppc_unpin_guest_page(kvm, va, gpa, false); in kvmppc_update_vpa() 506 kvmppc_unpin_guest_page(kvm, vpap->pinned_addr, vpap->gpa, in kvmppc_update_vpa() 508 vpap->gpa = gpa; in kvmppc_update_vpa() 1681 kvmppc_unpin_guest_page(kvm, vpa->pinned_addr, vpa->gpa, in unpin_vpa()
|
D | book3s_pr.c | 515 static int kvmppc_visible_gpa(struct kvm_vcpu *vcpu, gpa_t gpa) in kvmppc_visible_gpa() argument 522 gpa &= ~0xFFFULL; in kvmppc_visible_gpa() 523 if (unlikely(mp_pa) && unlikely((mp_pa & KVM_PAM) == (gpa & KVM_PAM))) { in kvmppc_visible_gpa() 527 return kvm_is_visible_gfn(vcpu->kvm, gpa >> PAGE_SHIFT); in kvmppc_visible_gpa()
|
/linux-4.1.27/Documentation/virtual/kvm/ |
D | mmu.txt | 35 gpa guest physical address 63 host physical addresses (gpa->hpa) 65 guest physical addresses, to host physical addresses (gva->gpa->hpa) 68 addresses, to host physical addresses (ngva->ngpa->gpa->hpa) 78 Guest memory (gpa) is part of the user address space of the process that is 80 addresses (gpa->hva); note that two gpas may alias to the same hva, but not 98 - changes in the gpa->hpa translation (either through gpa->hva changes or 121 nonpaging: gpa->hpa 122 paging: gva->gpa->hpa 123 paging, tdp: (gva->)gpa->hpa [all …]
|
/linux-4.1.27/virt/kvm/ |
D | kvm_main.c | 1550 int kvm_read_guest(struct kvm *kvm, gpa_t gpa, void *data, unsigned long len) in kvm_read_guest() argument 1552 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_read_guest() 1554 int offset = offset_in_page(gpa); in kvm_read_guest() 1570 int kvm_read_guest_atomic(struct kvm *kvm, gpa_t gpa, void *data, in kvm_read_guest_atomic() argument 1575 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_read_guest_atomic() 1576 int offset = offset_in_page(gpa); in kvm_read_guest_atomic() 1607 int kvm_write_guest(struct kvm *kvm, gpa_t gpa, const void *data, in kvm_write_guest() argument 1610 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_write_guest() 1612 int offset = offset_in_page(gpa); in kvm_write_guest() 1629 gpa_t gpa, unsigned long len) in kvm_gfn_to_hva_cache_init() argument [all …]
|
/linux-4.1.27/include/xen/interface/hvm/ |
D | hvm_op.h | 41 aligned_u64 gpa; member
|
/linux-4.1.27/arch/mips/kvm/ |
D | trap_emul.c | 24 gpa_t gpa; in kvm_trap_emul_gva_to_gpa_cb() local 28 gpa = CPHYSADDR(gva); in kvm_trap_emul_gva_to_gpa_cb() 32 gpa = KVM_INVALID_ADDR; in kvm_trap_emul_gva_to_gpa_cb() 35 kvm_debug("%s: gva %#lx, gpa: %#llx\n", __func__, gva, gpa); in kvm_trap_emul_gva_to_gpa_cb() 37 return gpa; in kvm_trap_emul_gva_to_gpa_cb()
|
/linux-4.1.27/arch/x86/include/asm/ |
D | kvm_host.h | 272 gpa_t (*translate_gpa)(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, 891 int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa, 1007 void kvm_mmu_pte_write(struct kvm_vcpu *vcpu, gpa_t gpa, 1015 gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, 1036 static inline gpa_t translate_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, in translate_gpa() argument 1039 return gpa; in translate_gpa()
|
/linux-4.1.27/arch/powerpc/include/asm/ |
D | kvm_book3s.h | 157 extern pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa, bool writing, 168 unsigned long gpa, bool dirty);
|
D | kvm_host.h | 316 unsigned long gpa; /* Current guest phys addr */ member
|
/linux-4.1.27/kernel/rcu/ |
D | tree.c | 1132 unsigned long gpa; in rcu_check_gp_kthread_starvation() local 1136 gpa = ACCESS_ONCE(rsp->gp_activity); in rcu_check_gp_kthread_starvation() 1137 if (j - gpa > 2 * HZ) in rcu_check_gp_kthread_starvation() 1139 rsp->name, j - gpa); in rcu_check_gp_kthread_starvation() 1167 unsigned long gpa; in print_other_cpu_stall() local 1220 gpa = ACCESS_ONCE(rsp->gp_activity); in print_other_cpu_stall() 1222 rsp->name, j - gpa, j, gpa, in print_other_cpu_stall()
|
/linux-4.1.27/arch/x86/platform/uv/ |
D | tlb_uv.c | 1693 unsigned long gpa; in activation_descriptor_init() local 1711 gpa = uv_gpa(bau_desc); in activation_descriptor_init() 1712 n = uv_gpa_to_gnode(gpa); in activation_descriptor_init() 1713 m = uv_gpa_to_offset(gpa); in activation_descriptor_init()
|
/linux-4.1.27/arch/x86/xen/ |
D | mmu.c | 2409 a.gpa = __pa(mm->pgd); in xen_hvm_exit_mmap() 2420 a.gpa = 0x00; in is_pagetable_dying_supported()
|