Lines Matching refs:slot
487 struct kvm_memory_slot *slot; in id_to_memslot() local
489 slot = &slots->memslots[index]; in id_to_memslot()
491 WARN_ON(slot->id != id); in id_to_memslot()
492 return slot; in id_to_memslot()
519 int kvm_arch_create_memslot(struct kvm *kvm, struct kvm_memory_slot *slot,
536 struct kvm_memory_slot *slot);
544 unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
545 unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn,
557 pfn_t gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
558 pfn_t gfn_to_pfn_memslot_atomic(struct kvm_memory_slot *slot, gfn_t gfn);
615 struct kvm_memory_slot *slot,
761 int kvm_iommu_map_pages(struct kvm *kvm, struct kvm_memory_slot *slot);
762 void kvm_iommu_unmap_pages(struct kvm *kvm, struct kvm_memory_slot *slot);
765 struct kvm_memory_slot *slot) in kvm_iommu_map_pages() argument
771 struct kvm_memory_slot *slot) in kvm_iommu_unmap_pages() argument
816 int slot = atomic_read(&slots->lru_slot); in search_memslots() local
819 if (gfn >= memslots[slot].base_gfn && in search_memslots()
820 gfn < memslots[slot].base_gfn + memslots[slot].npages) in search_memslots()
821 return &memslots[slot]; in search_memslots()
824 slot = start + (end - start) / 2; in search_memslots()
826 if (gfn >= memslots[slot].base_gfn) in search_memslots()
827 end = slot; in search_memslots()
829 start = slot + 1; in search_memslots()
848 __gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn) in __gfn_to_hva_memslot() argument
850 return slot->userspace_addr + (gfn - slot->base_gfn) * PAGE_SIZE; in __gfn_to_hva_memslot()
859 hva_to_gfn_memslot(unsigned long hva, struct kvm_memory_slot *slot) in hva_to_gfn_memslot() argument
861 gfn_t gfn_offset = (hva - slot->userspace_addr) >> PAGE_SHIFT; in hva_to_gfn_memslot()
863 return slot->base_gfn + gfn_offset; in hva_to_gfn_memslot()