/linux-4.1.27/arch/x86/xen/ |
D | p2m.c | 110 static inline unsigned p2m_top_index(unsigned long pfn) in p2m_top_index() argument 112 BUG_ON(pfn >= MAX_P2M_PFN); in p2m_top_index() 113 return pfn / (P2M_MID_PER_PAGE * P2M_PER_PAGE); in p2m_top_index() 116 static inline unsigned p2m_mid_index(unsigned long pfn) in p2m_mid_index() argument 118 return (pfn / P2M_PER_PAGE) % P2M_MID_PER_PAGE; in p2m_mid_index() 121 static inline unsigned p2m_index(unsigned long pfn) in p2m_index() argument 123 return pfn % P2M_PER_PAGE; in p2m_index() 158 static void p2m_init_identity(unsigned long *p2m, unsigned long pfn) in p2m_init_identity() argument 163 p2m[i] = IDENTITY_FRAME(pfn + i); in p2m_init_identity() 196 unsigned long pfn, mfn; in xen_build_mfn_list_list() local [all …]
|
D | setup.c | 130 unsigned long __ref xen_chk_extra_mem(unsigned long pfn) in xen_chk_extra_mem() argument 133 phys_addr_t addr = PFN_PHYS(pfn); in xen_chk_extra_mem() 141 return IDENTITY_FRAME(pfn); in xen_chk_extra_mem() 149 unsigned long pfn, pfn_s, pfn_e; in xen_inv_extra_mem() local 157 for (pfn = pfn_s; pfn < pfn_e; pfn++) in xen_inv_extra_mem() 158 set_phys_to_machine(pfn, INVALID_P2M_ENTRY); in xen_inv_extra_mem() 226 unsigned long pfn, end; in xen_set_identity_and_release_chunk() local 233 for (pfn = start_pfn; pfn < end; pfn++) { in xen_set_identity_and_release_chunk() 234 unsigned long mfn = pfn_to_mfn(pfn); in xen_set_identity_and_release_chunk() 237 if (mfn == INVALID_P2M_ENTRY || mfn_to_pfn(mfn) != pfn) in xen_set_identity_and_release_chunk() [all …]
|
D | mmu.c | 370 unsigned long pfn = mfn_to_pfn(mfn); in pte_mfn_to_pfn() local 373 if (unlikely(pfn == ~0)) in pte_mfn_to_pfn() 376 val = ((pteval_t)pfn << PAGE_SHIFT) | flags; in pte_mfn_to_pfn() 385 unsigned long pfn = (val & PTE_PFN_MASK) >> PAGE_SHIFT; in pte_pfn_to_mfn() local 390 mfn = __pfn_to_mfn(pfn); in pte_pfn_to_mfn() 392 mfn = pfn; in pte_pfn_to_mfn() 728 static void xen_do_pin(unsigned level, unsigned long pfn) in xen_do_pin() argument 733 op.arg1.mfn = pfn_to_mfn(pfn); in xen_do_pin() 752 unsigned long pfn = page_to_pfn(page); in xen_pin_page() local 783 pfn_pte(pfn, PAGE_KERNEL_RO), in xen_pin_page() [all …]
|
D | mmu.h | 14 bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn); 16 void set_pte_mfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags);
|
D | enlighten.c | 488 unsigned long pfn; in set_aliased_prot() local 495 pfn = pte_pfn(*ptep); in set_aliased_prot() 496 page = pfn_to_page(pfn); in set_aliased_prot() 498 pte = pfn_pte(pfn, prot); in set_aliased_prot() 530 void *av = __va(PFN_PHYS(pfn)); in set_aliased_prot() 606 unsigned long pfn, mfn; in xen_load_gdt() local 619 pfn = pte_pfn(*ptep); in xen_load_gdt() 620 mfn = pfn_to_mfn(pfn); in xen_load_gdt() 621 virt = __va(PFN_PHYS(pfn)); in xen_load_gdt() 654 unsigned long pfn, mfn; in xen_load_gdt_boot() local [all …]
|
/linux-4.1.27/mm/ |
D | page_isolation.c | 15 unsigned long flags, pfn; in set_migratetype_isolate() local 24 pfn = page_to_pfn(page); in set_migratetype_isolate() 25 arg.start_pfn = pfn; in set_migratetype_isolate() 132 __first_valid_page(unsigned long pfn, unsigned long nr_pages) in __first_valid_page() argument 136 if (pfn_valid_within(pfn + i)) in __first_valid_page() 140 return pfn_to_page(pfn + i); in __first_valid_page() 160 unsigned long pfn; in start_isolate_page_range() local 167 for (pfn = start_pfn; in start_isolate_page_range() 168 pfn < end_pfn; in start_isolate_page_range() 169 pfn += pageblock_nr_pages) { in start_isolate_page_range() [all …]
|
D | page_ext.c | 100 unsigned long pfn = page_to_pfn(page); in lookup_page_ext() local 115 offset = pfn - round_down(node_start_pfn(page_to_nid(page)), in lookup_page_ext() 177 unsigned long pfn = page_to_pfn(page); in lookup_page_ext() local 178 struct mem_section *section = __pfn_to_section(pfn); in lookup_page_ext() 189 return section->page_ext + pfn; in lookup_page_ext() 211 static int __meminit init_section_page_ext(unsigned long pfn, int nid) in init_section_page_ext() argument 217 section = __pfn_to_section(pfn); in init_section_page_ext() 241 pfn &= PAGE_SECTION_MASK; in init_section_page_ext() 242 section->page_ext = base - pfn; in init_section_page_ext() 262 static void __free_page_ext(unsigned long pfn) in __free_page_ext() argument [all …]
|
D | page_owner.c | 80 print_page_owner(char __user *buf, size_t count, unsigned long pfn, in print_page_owner() argument 103 pageblock_mt = get_pfnblock_migratetype(page, pfn); in print_page_owner() 107 pfn, in print_page_owner() 108 pfn >> pageblock_order, in print_page_owner() 149 unsigned long pfn; in read_page_owner() local 157 pfn = min_low_pfn + *ppos; in read_page_owner() 160 while (!pfn_valid(pfn) && (pfn & (MAX_ORDER_NR_PAGES - 1)) != 0) in read_page_owner() 161 pfn++; in read_page_owner() 166 for (; pfn < max_pfn; pfn++) { in read_page_owner() 171 if ((pfn & (MAX_ORDER_NR_PAGES - 1)) == 0 && !pfn_valid(pfn)) { in read_page_owner() [all …]
|
D | memory-failure.c | 191 unsigned long pfn, struct page *page, int flags) in kill_proc() argument 198 pfn, t->comm, t->pid); in kill_proc() 335 int fail, struct page *page, unsigned long pfn, in kill_procs() argument 350 pfn, tk->tsk->comm, tk->tsk->pid); in kill_procs() 361 pfn, page, flags) < 0) in kill_procs() 364 pfn, tk->tsk->comm, tk->tsk->pid); in kill_procs() 599 static int me_kernel(struct page *p, unsigned long pfn) in me_kernel() argument 607 static int me_unknown(struct page *p, unsigned long pfn) in me_unknown() argument 609 printk(KERN_ERR "MCE %#lx: Unknown page state\n", pfn); in me_unknown() 616 static int me_pagecache_clean(struct page *p, unsigned long pfn) in me_pagecache_clean() argument [all …]
|
D | memory_hotplug.c | 247 unsigned long i, pfn, end_pfn, nr_pages; in register_page_bootmem_info_node() local 271 pfn = pgdat->node_start_pfn; in register_page_bootmem_info_node() 275 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node() 282 if (pfn_valid(pfn) && (pfn_to_nid(pfn) == node)) in register_page_bootmem_info_node() 283 register_page_bootmem_info_section(pfn); in register_page_bootmem_info_node() 330 unsigned long pfn; in fix_zone_id() local 332 for (pfn = start_pfn; pfn < end_pfn; pfn++) in fix_zone_id() 333 set_page_links(pfn_to_page(pfn), zid, nid, pfn); in fix_zone_id() 553 unsigned long pfn; in find_biggest_section_pfn() local 556 pfn = end_pfn - 1; in find_biggest_section_pfn() [all …]
|
D | cma.c | 84 static void cma_clear_bitmap(struct cma *cma, unsigned long pfn, in cma_clear_bitmap() argument 89 bitmap_no = (pfn - cma->base_pfn) >> cma->order_per_bit; in cma_clear_bitmap() 100 unsigned long base_pfn = cma->base_pfn, pfn = base_pfn; in cma_activate_area() local 109 WARN_ON_ONCE(!pfn_valid(pfn)); in cma_activate_area() 110 zone = page_zone(pfn_to_page(pfn)); in cma_activate_area() 115 base_pfn = pfn; in cma_activate_area() 116 for (j = pageblock_nr_pages; j; --j, pfn++) { in cma_activate_area() 117 WARN_ON_ONCE(!pfn_valid(pfn)); in cma_activate_area() 124 if (page_zone(pfn_to_page(pfn)) != zone) in cma_activate_area() 364 unsigned long mask, offset, pfn, start = 0; in cma_alloc() local [all …]
|
D | compaction.c | 59 unsigned long pfn = page_to_pfn(page); in release_freepages() local 62 if (pfn > high_pfn) in release_freepages() 63 high_pfn = pfn; in release_freepages() 219 unsigned long pfn; in __reset_isolation_suitable() local 227 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in __reset_isolation_suitable() 232 if (!pfn_valid(pfn)) in __reset_isolation_suitable() 235 page = pfn_to_page(pfn); in __reset_isolation_suitable() 267 unsigned long pfn; in update_pageblock_skip() local 280 pfn = page_to_pfn(page); in update_pageblock_skip() 284 if (pfn > zone->compact_cached_migrate_pfn[0]) in update_pageblock_skip() [all …]
|
D | hwpoison-inject.c | 15 unsigned long pfn = val; in hwpoison_inject() local 23 if (!pfn_valid(pfn)) in hwpoison_inject() 26 p = pfn_to_page(pfn); in hwpoison_inject() 58 pr_info("Injecting memory failure at pfn %#lx\n", pfn); in hwpoison_inject() 59 return memory_failure(pfn, 18, MF_COUNT_INCREASED); in hwpoison_inject()
|
D | page_alloc.c | 253 unsigned long pfn = page_to_pfn(page); in page_outside_zone_boundaries() local 260 if (!zone_spans_pfn(zone, pfn)) in page_outside_zone_boundaries() 266 pfn, zone_to_nid(zone), zone->name, in page_outside_zone_boundaries() 574 unsigned long pfn, in __free_one_page() argument 593 page_idx = pfn & ((1 << MAX_ORDER) - 1); in __free_one_page() 768 struct page *page, unsigned long pfn, in free_one_page() argument 780 migratetype = get_pfnblock_migratetype(page, pfn); in free_one_page() 782 __free_one_page(page, pfn, zone, order, migratetype); in free_one_page() 842 unsigned long pfn = page_to_pfn(page); in __free_pages_ok() local 847 migratetype = get_pfnblock_migratetype(page, pfn); in __free_pages_ok() [all …]
|
D | vmstat.c | 980 unsigned long pfn; in pagetypeinfo_showblockcount_print() local 985 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in pagetypeinfo_showblockcount_print() 988 if (!pfn_valid(pfn)) in pagetypeinfo_showblockcount_print() 991 page = pfn_to_page(pfn); in pagetypeinfo_showblockcount_print() 994 if (!memmap_valid_within(pfn, page, zone)) in pagetypeinfo_showblockcount_print() 1032 unsigned long pfn = zone->zone_start_pfn, block_end_pfn; in pagetypeinfo_showmixedcount_print() local 1033 unsigned long end_pfn = pfn + zone->spanned_pages; in pagetypeinfo_showmixedcount_print() 1039 pfn = zone->zone_start_pfn; in pagetypeinfo_showmixedcount_print() 1046 for (; pfn < end_pfn; ) { in pagetypeinfo_showmixedcount_print() 1047 if (!pfn_valid(pfn)) { in pagetypeinfo_showmixedcount_print() [all …]
|
D | internal.h | 158 extern void __free_pages_bootmem(struct page *page, unsigned long pfn, 323 unsigned long pfn = page_to_pfn(base) + offset; in mem_map_next() local 324 if (!pfn_valid(pfn)) in mem_map_next() 326 return pfn_to_page(pfn); in mem_map_next() 366 enum zone_type zone, unsigned long nid, unsigned long pfn); 381 enum zone_type zone, unsigned long nid, unsigned long pfn) in mminit_verify_page_links() argument
|
D | sparse.c | 172 unsigned long pfn; in memory_present() local 176 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { in memory_present() 177 unsigned long section = pfn_to_section_nr(pfn); in memory_present() 197 unsigned long pfn; in node_memmap_size_bytes() local 201 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in node_memmap_size_bytes() 202 if (nid != early_pfn_to_nid(pfn)) in node_memmap_size_bytes() 205 if (pfn_present(pfn)) in node_memmap_size_bytes()
|
/linux-4.1.27/arch/x86/include/asm/xen/ |
D | page.h | 48 extern unsigned long get_phys_to_machine(unsigned long pfn); 49 extern bool set_phys_to_machine(unsigned long pfn, unsigned long mfn); 50 extern bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn); 85 static inline unsigned long __pfn_to_mfn(unsigned long pfn) in __pfn_to_mfn() argument 89 if (pfn < xen_p2m_size) in __pfn_to_mfn() 90 mfn = xen_p2m_addr[pfn]; in __pfn_to_mfn() 91 else if (unlikely(pfn < xen_max_p2m_pfn)) in __pfn_to_mfn() 92 return get_phys_to_machine(pfn); in __pfn_to_mfn() 94 return IDENTITY_FRAME(pfn); in __pfn_to_mfn() 97 return get_phys_to_machine(pfn); in __pfn_to_mfn() [all …]
|
D | interface_32.h | 99 #define xen_pfn_to_cr3(pfn) (((unsigned)(pfn) << 12) | ((unsigned)(pfn) >> 20)) argument
|
/linux-4.1.27/arch/arm/xen/ |
D | p2m.c | 21 unsigned long pfn; member 42 if (new->pfn == entry->pfn) in xen_add_phys_to_mach_entry() 45 if (new->pfn < entry->pfn) in xen_add_phys_to_mach_entry() 57 __func__, &new->pfn, &new->mfn, &entry->pfn, &entry->mfn); in xen_add_phys_to_mach_entry() 62 unsigned long __pfn_to_mfn(unsigned long pfn) in __pfn_to_mfn() argument 71 if (entry->pfn <= pfn && in __pfn_to_mfn() 72 entry->pfn + entry->nr_pages > pfn) { in __pfn_to_mfn() 74 return entry->mfn + (pfn - entry->pfn); in __pfn_to_mfn() 76 if (pfn < entry->pfn) in __pfn_to_mfn() 119 bool __set_phys_to_machine_multi(unsigned long pfn, in __set_phys_to_machine_multi() argument [all …]
|
D | mm.c | 51 unsigned long pfn; in dma_cache_maint() local 54 pfn = (handle >> PAGE_SHIFT) + offset / PAGE_SIZE; in dma_cache_maint() 66 cflush.a.dev_bus_addr = pfn << PAGE_SHIFT; in dma_cache_maint() 82 pfn++; in dma_cache_maint() 141 unsigned long pfn, in xen_arch_need_swiotlb() argument 144 return (!hypercall_cflush && (pfn != mfn) && !is_device_dma_coherent(dev)); in xen_arch_need_swiotlb()
|
/linux-4.1.27/include/trace/events/ |
D | cma.h | 12 TP_PROTO(unsigned long pfn, const struct page *page, 15 TP_ARGS(pfn, page, count, align), 18 __field(unsigned long, pfn) 25 __entry->pfn = pfn; 32 __entry->pfn, 40 TP_PROTO(unsigned long pfn, const struct page *page, 43 TP_ARGS(pfn, page, count), 46 __field(unsigned long, pfn) 52 __entry->pfn = pfn; 58 __entry->pfn,
|
D | kmem.h | 180 __field( unsigned long, pfn ) 185 __entry->pfn = page_to_pfn(page); 190 pfn_to_page(__entry->pfn), 191 __entry->pfn, 202 __field( unsigned long, pfn ) 207 __entry->pfn = page_to_pfn(page); 212 pfn_to_page(__entry->pfn), 213 __entry->pfn, 225 __field( unsigned long, pfn ) 232 __entry->pfn = page ? page_to_pfn(page) : -1UL; [all …]
|
D | pagemap.h | 38 __field(unsigned long, pfn ) 45 __entry->pfn = page_to_pfn(page); 53 __entry->pfn, 71 __field(unsigned long, pfn ) 76 __entry->pfn = page_to_pfn(page); 80 TP_printk("page=%p pfn=%lu", __entry->page, __entry->pfn)
|
D | filemap.h | 21 __field(unsigned long, pfn) 28 __entry->pfn = page_to_pfn(page); 40 pfn_to_page(__entry->pfn), 41 __entry->pfn,
|
D | xen.h | 328 TP_PROTO(struct mm_struct *mm, unsigned long pfn, unsigned level, bool pinned), 329 TP_ARGS(mm, pfn, level, pinned), 332 __field(unsigned long, pfn) 337 __entry->pfn = pfn; 341 __entry->mm, __entry->pfn, __entry->level, 346 TP_PROTO(unsigned long pfn, unsigned level, bool pinned), 347 TP_ARGS(pfn, level, pinned), 349 __field(unsigned long, pfn) 353 TP_fast_assign(__entry->pfn = pfn; 357 __entry->pfn, __entry->level,
|
/linux-4.1.27/include/asm-generic/ |
D | memory_model.h | 15 #define arch_pfn_to_nid(pfn) pfn_to_nid(pfn) argument 19 #define arch_local_page_offset(pfn, nid) \ argument 20 ((pfn) - NODE_DATA(nid)->node_start_pfn) 30 #define __pfn_to_page(pfn) (mem_map + ((pfn) - ARCH_PFN_OFFSET)) argument 35 #define __pfn_to_page(pfn) \ argument 36 ({ unsigned long __pfn = (pfn); \ 51 #define __pfn_to_page(pfn) (vmemmap + (pfn)) argument 65 #define __pfn_to_page(pfn) \ argument 66 ({ unsigned long __pfn = (pfn); \
|
D | page.h | 84 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument 93 #define pfn_valid(pfn) ((pfn) >= ARCH_PFN_OFFSET && ((pfn) - ARCH_PFN_OFFSET) < max_mapnr) argument
|
D | pgtable.h | 493 unsigned long pfn, unsigned long addr, in track_pfn_remap() argument 504 unsigned long pfn) in track_pfn_insert() argument 524 unsigned long pfn, unsigned long size) in untrack_pfn() argument 529 unsigned long pfn, unsigned long addr, 532 unsigned long pfn); 534 extern void untrack_pfn(struct vm_area_struct *vma, unsigned long pfn, 539 static inline int is_zero_pfn(unsigned long pfn) in is_zero_pfn() argument 542 unsigned long offset_from_zero_pfn = pfn - zero_pfn; in is_zero_pfn() 549 static inline int is_zero_pfn(unsigned long pfn) in is_zero_pfn() argument 552 return pfn == zero_pfn; in is_zero_pfn()
|
/linux-4.1.27/arch/m32r/include/asm/ |
D | mmzone.h | 16 #define node_localnr(pfn, nid) ((pfn) - NODE_DATA(nid)->node_start_pfn) argument 27 #define pfn_valid(pfn) (1) argument 29 #define pfn_valid(pfn) ((pfn) < num_physpages) argument 36 static __inline__ int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument 41 if (pfn >= node_start_pfn(node) && pfn < node_end_pfn(node)) in pfn_to_nid() 47 static __inline__ struct pglist_data *pfn_to_pgdat(unsigned long pfn) in pfn_to_pgdat() argument 49 return(NODE_DATA(pfn_to_nid(pfn))); in pfn_to_pgdat()
|
D | pgtable-2level.h | 71 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument 72 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument
|
D | cacheflush.h | 14 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) argument 36 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) argument 50 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) argument
|
/linux-4.1.27/arch/parisc/include/asm/ |
D | mmzone.h | 33 #define pfn_is_io(pfn) ((pfn & (0xf0000000UL >> PAGE_SHIFT)) == (0xf0000000UL >> PAGE_SHIFT)) argument 36 #define pfn_is_io(pfn) ((pfn & (0xf000000000000000UL >> PAGE_SHIFT)) == (0xf000000000000000UL >> PA… argument 39 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument 43 if (unlikely(pfn_is_io(pfn))) in pfn_to_nid() 46 i = pfn >> PFNNID_SHIFT; in pfn_to_nid() 52 static inline int pfn_valid(int pfn) in pfn_valid() argument 54 int nid = pfn_to_nid(pfn); in pfn_valid() 57 return (pfn < node_end_pfn(nid)); in pfn_valid()
|
/linux-4.1.27/arch/arm/mach-gemini/ |
D | mm.c | 23 .pfn =__phys_to_pfn(GEMINI_GLOBAL_BASE), 28 .pfn = __phys_to_pfn(GEMINI_UART_BASE), 33 .pfn = __phys_to_pfn(GEMINI_TIMER_BASE), 38 .pfn = __phys_to_pfn(GEMINI_INTERRUPT_BASE), 43 .pfn = __phys_to_pfn(GEMINI_POWER_CTRL_BASE), 48 .pfn = __phys_to_pfn(GEMINI_GPIO_BASE(0)), 53 .pfn = __phys_to_pfn(GEMINI_GPIO_BASE(1)), 58 .pfn = __phys_to_pfn(GEMINI_GPIO_BASE(2)), 63 .pfn = __phys_to_pfn(GEMINI_FLASH_CTRL_BASE), 68 .pfn = __phys_to_pfn(GEMINI_DRAM_CTRL_BASE), [all …]
|
/linux-4.1.27/arch/s390/kernel/ |
D | suspend.c | 99 void page_key_read(unsigned long *pfn) in page_key_read() argument 103 addr = (unsigned long) page_address(pfn_to_page(*pfn)); in page_key_read() 104 *(unsigned char *) pfn = (unsigned char) page_get_storage_key(addr); in page_key_read() 111 void page_key_memorize(unsigned long *pfn) in page_key_memorize() argument 113 page_key_wp->data[page_key_wx] = *(unsigned char *) pfn; in page_key_memorize() 114 *(unsigned char *) pfn = 0; in page_key_memorize() 137 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 145 if (pfn <= LC_PAGES) in pfn_is_nosave() 147 if (pfn >= nosave_begin_pfn && pfn < nosave_end_pfn) in pfn_is_nosave() 150 if (pfn >= stext_pfn && pfn <= eshared_pfn) in pfn_is_nosave() [all …]
|
D | crash_dump.c | 166 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, size_t csize, in copy_oldmem_page() argument 173 src = (pfn << PAGE_SHIFT) + offset; in copy_oldmem_page() 187 unsigned long from, unsigned long pfn, in remap_oldmem_pfn_range_kdump() argument 193 if (pfn < OLDMEM_SIZE >> PAGE_SHIFT) { in remap_oldmem_pfn_range_kdump() 194 size_old = min(size, OLDMEM_SIZE - (pfn << PAGE_SHIFT)); in remap_oldmem_pfn_range_kdump() 196 pfn + (OLDMEM_BASE >> PAGE_SHIFT), in remap_oldmem_pfn_range_kdump() 202 pfn += size_old >> PAGE_SHIFT; in remap_oldmem_pfn_range_kdump() 204 return remap_pfn_range(vma, from, pfn, size, prot); in remap_oldmem_pfn_range_kdump() 215 unsigned long pfn, in remap_oldmem_pfn_range_zfcpdump() argument 221 if (pfn < hsa_end >> PAGE_SHIFT) { in remap_oldmem_pfn_range_zfcpdump() [all …]
|
/linux-4.1.27/arch/x86/power/ |
D | hibernate_32.c | 83 unsigned long pfn; in resume_physical_mapping_init() local 91 pfn = 0; in resume_physical_mapping_init() 98 if (pfn >= max_low_pfn) in resume_physical_mapping_init() 102 if (pfn >= max_low_pfn) in resume_physical_mapping_init() 110 set_pmd(pmd, pfn_pmd(pfn, PAGE_KERNEL_LARGE_EXEC)); in resume_physical_mapping_init() 111 pfn += PTRS_PER_PTE; in resume_physical_mapping_init() 120 for (; pte < max_pte; pte++, pfn++) { in resume_physical_mapping_init() 121 if (pfn >= max_low_pfn) in resume_physical_mapping_init() 124 set_pte(pte, pfn_pte(pfn, PAGE_KERNEL_EXEC)); in resume_physical_mapping_init() 167 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument [all …]
|
D | hibernate_64.c | 103 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 107 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
|
/linux-4.1.27/arch/alpha/include/asm/ |
D | mmzone.h | 34 #define node_localnr(pfn, nid) ((pfn) - NODE_DATA(nid)->node_start_pfn) argument 81 unsigned long pfn; \ 83 pfn = page_to_pfn(page) << 32; \ 84 pte_val(pte) = pfn | pgprot_val(pgprot); \ 103 #define pfn_to_nid(pfn) pa_to_nid(((u64)(pfn) << PAGE_SHIFT)) argument 104 #define pfn_valid(pfn) \ argument 105 (((pfn) - node_start_pfn(pfn_to_nid(pfn))) < \ 106 node_spanned_pages(pfn_to_nid(pfn))) \
|
D | pgtable.h | 194 #define PHYS_TWIDDLE(pfn) \ argument 195 ((((pfn) & KSEG_PFN) == (0x40000000000UL >> PAGE_SHIFT)) \ 196 ? ((pfn) ^= KSEG_PFN) : (pfn)) 198 #define PHYS_TWIDDLE(pfn) (pfn) argument
|
/linux-4.1.27/arch/unicore32/mm/ |
D | ioremap.c | 102 remap_area_sections(unsigned long virt, unsigned long pfn, in remap_area_sections() argument 118 set_pmd(pmd, __pmd(__pfn_to_phys(pfn) | type->prot_sect)); in remap_area_sections() 119 pfn += SZ_4M >> PAGE_SHIFT; in remap_area_sections() 129 void __iomem *__uc32_ioremap_pfn_caller(unsigned long pfn, in __uc32_ioremap_pfn_caller() argument 140 if (pfn >= 0x100000 && (__pfn_to_phys(pfn) & ~SECTION_MASK)) in __uc32_ioremap_pfn_caller() 146 if (pfn_valid(pfn)) { in __uc32_ioremap_pfn_caller() 168 if (!((__pfn_to_phys(pfn) | size | addr) & ~PMD_MASK)) { in __uc32_ioremap_pfn_caller() 170 err = remap_area_sections(addr, pfn, size, type); in __uc32_ioremap_pfn_caller() 172 err = ioremap_page_range(addr, addr + size, __pfn_to_phys(pfn), in __uc32_ioremap_pfn_caller() 189 unsigned long pfn = __phys_to_pfn(phys_addr); in __uc32_ioremap_caller() local [all …]
|
D | mmu.c | 81 unsigned long pfn; member 164 unsigned long end, unsigned long pfn, in alloc_init_pte() argument 169 set_pte(pte, pfn_pte(pfn, __pgprot(type->prot_pte))); in alloc_init_pte() 170 pfn++; in alloc_init_pte() 217 __pfn_to_phys((u64)md->pfn), md->virtual); in create_mapping() 225 __pfn_to_phys((u64)md->pfn), md->virtual); in create_mapping() 231 phys = (unsigned long)__pfn_to_phys(md->pfn); in create_mapping() 237 __pfn_to_phys(md->pfn), addr); in create_mapping() 368 map.pfn = __phys_to_pfn(virt_to_phys(vectors)); in devicemaps_init() 378 map.pfn = __phys_to_pfn(virt_to_phys(vectors)); in devicemaps_init() [all …]
|
/linux-4.1.27/arch/unicore32/kernel/ |
D | hibernate.c | 75 unsigned long pfn; in resume_physical_mapping_init() local 83 pfn = 0; in resume_physical_mapping_init() 90 if (pfn >= max_low_pfn) in resume_physical_mapping_init() 96 if (pfn >= max_low_pfn) in resume_physical_mapping_init() 107 for (; pte < max_pte; pte++, pfn++) { in resume_physical_mapping_init() 108 if (pfn >= max_low_pfn) in resume_physical_mapping_init() 111 set_pte(pte, pfn_pte(pfn, PAGE_KERNEL_EXEC)); in resume_physical_mapping_init() 145 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 150 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
|
/linux-4.1.27/arch/x86/include/asm/ |
D | mmzone_32.h | 33 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument 36 return((int) physnode_map[(pfn) / PAGES_PER_SECTION]); in pfn_to_nid() 42 static inline int pfn_valid(int pfn) in pfn_valid() argument 44 int nid = pfn_to_nid(pfn); in pfn_valid() 47 return (pfn < node_end_pfn(nid)); in pfn_valid() 51 #define early_pfn_valid(pfn) pfn_valid((pfn)) argument
|
D | pgalloc.h | 15 static inline void paravirt_alloc_pte(struct mm_struct *mm, unsigned long pfn) {} in paravirt_alloc_pte() argument 16 static inline void paravirt_alloc_pmd(struct mm_struct *mm, unsigned long pfn) {} in paravirt_alloc_pmd() argument 17 static inline void paravirt_alloc_pmd_clone(unsigned long pfn, unsigned long clonepfn, in paravirt_alloc_pmd_clone() argument 19 static inline void paravirt_alloc_pud(struct mm_struct *mm, unsigned long pfn) {} in paravirt_alloc_pud() argument 20 static inline void paravirt_release_pte(unsigned long pfn) {} in paravirt_release_pte() argument 21 static inline void paravirt_release_pmd(unsigned long pfn) {} in paravirt_release_pmd() argument 22 static inline void paravirt_release_pud(unsigned long pfn) {} in paravirt_release_pud() argument 72 unsigned long pfn = page_to_pfn(pte); in pmd_populate() local 74 paravirt_alloc_pte(mm, pfn); in pmd_populate() 75 set_pmd(pmd, __pmd(((pteval_t)pfn << PAGE_SHIFT) | _PAGE_TABLE)); in pmd_populate()
|
D | page_64.h | 34 #define pfn_valid(pfn) ((pfn) < max_pfn) argument
|
D | page_32.h | 18 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
|
D | highmem.h | 69 void *kmap_atomic_pfn(unsigned long pfn); 70 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot);
|
/linux-4.1.27/include/linux/ |
D | mmzone.h | 85 static inline int get_pfnblock_migratetype(struct page *page, unsigned long pfn) in get_pfnblock_migratetype() argument 88 return get_pfnblock_flags_mask(page, pfn, PB_migrate_end, in get_pfnblock_migratetype() 553 static inline bool zone_spans_pfn(const struct zone *zone, unsigned long pfn) in zone_spans_pfn() argument 555 return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone); in zone_spans_pfn() 1041 static inline unsigned long early_pfn_to_nid(unsigned long pfn) in early_pfn_to_nid() argument 1048 #define pfn_to_nid(pfn) (0) argument 1074 #define pfn_to_section_nr(pfn) ((pfn) >> PFN_SECTION_SHIFT) argument 1077 #define SECTION_ALIGN_UP(pfn) (((pfn) + PAGES_PER_SECTION - 1) & PAGE_SECTION_MASK) argument 1078 #define SECTION_ALIGN_DOWN(pfn) ((pfn) & PAGE_SECTION_MASK) argument 1176 static inline struct mem_section *__pfn_to_section(unsigned long pfn) in __pfn_to_section() argument [all …]
|
D | io-mapping.h | 88 unsigned long pfn; in io_mapping_map_atomic_wc() local 92 pfn = (unsigned long) (phys_addr >> PAGE_SHIFT); in io_mapping_map_atomic_wc() 93 return iomap_atomic_prot_pfn(pfn, mapping->prot); in io_mapping_map_atomic_wc()
|
D | suspend.h | 477 void page_key_read(unsigned long *pfn); 478 void page_key_memorize(unsigned long *pfn); 494 static inline void page_key_read(unsigned long *pfn) {} in page_key_read() argument 495 static inline void page_key_memorize(unsigned long *pfn) {} in page_key_memorize() argument
|
D | crash_dump.h | 22 unsigned long from, unsigned long pfn, 83 extern int register_oldmem_pfn_is_ram(int (*fn)(unsigned long pfn));
|
/linux-4.1.27/arch/arm/include/asm/xen/ |
D | page.h | 14 #define phys_to_machine_mapping_valid(pfn) (1) argument 35 unsigned long __pfn_to_mfn(unsigned long pfn); 38 static inline unsigned long pfn_to_mfn(unsigned long pfn) in pfn_to_mfn() argument 43 mfn = __pfn_to_mfn(pfn); in pfn_to_mfn() 48 return pfn; in pfn_to_mfn() 98 bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn); 99 bool __set_phys_to_machine_multi(unsigned long pfn, unsigned long mfn, 102 static inline bool set_phys_to_machine(unsigned long pfn, unsigned long mfn) in set_phys_to_machine() argument 104 return __set_phys_to_machine(pfn, mfn); in set_phys_to_machine() 111 unsigned long pfn,
|
D | page-coherent.h | 53 unsigned long pfn = PFN_DOWN(handle); in xen_dma_unmap_page() local 58 if (pfn_valid(pfn)) { in xen_dma_unmap_page() 68 unsigned long pfn = PFN_DOWN(handle); in xen_dma_sync_single_for_cpu() local 69 if (pfn_valid(pfn)) { in xen_dma_sync_single_for_cpu() 79 unsigned long pfn = PFN_DOWN(handle); in xen_dma_sync_single_for_device() local 80 if (pfn_valid(pfn)) { in xen_dma_sync_single_for_device()
|
/linux-4.1.27/kernel/power/ |
D | snapshot.c | 634 static int memory_bm_find_bit(struct memory_bitmap *bm, unsigned long pfn, in memory_bm_find_bit() argument 643 if (pfn >= zone->start_pfn && pfn < zone->end_pfn) in memory_bm_find_bit() 650 if (pfn >= curr->start_pfn && pfn < curr->end_pfn) { in memory_bm_find_bit() 666 if (((pfn - zone->start_pfn) & ~BM_BLOCK_MASK) == bm->cur.node_pfn) in memory_bm_find_bit() 670 block_nr = (pfn - zone->start_pfn) >> BM_BLOCK_SHIFT; in memory_bm_find_bit() 685 bm->cur.node_pfn = (pfn - zone->start_pfn) & ~BM_BLOCK_MASK; in memory_bm_find_bit() 689 *bit_nr = (pfn - zone->start_pfn) & BM_BLOCK_MASK; in memory_bm_find_bit() 694 static void memory_bm_set_bit(struct memory_bitmap *bm, unsigned long pfn) in memory_bm_set_bit() argument 700 error = memory_bm_find_bit(bm, pfn, &addr, &bit); in memory_bm_set_bit() 705 static int mem_bm_set_bit_check(struct memory_bitmap *bm, unsigned long pfn) in mem_bm_set_bit_check() argument [all …]
|
/linux-4.1.27/arch/arm/mach-omap2/ |
D | io.c | 75 .pfn = __phys_to_pfn(L3_24XX_PHYS), 81 .pfn = __phys_to_pfn(L4_24XX_PHYS), 91 .pfn = __phys_to_pfn(DSP_MEM_2420_PHYS), 97 .pfn = __phys_to_pfn(DSP_IPI_2420_PHYS), 103 .pfn = __phys_to_pfn(DSP_MMU_2420_PHYS), 115 .pfn = __phys_to_pfn(L4_WK_243X_PHYS), 121 .pfn = __phys_to_pfn(OMAP243X_GPMC_PHYS), 127 .pfn = __phys_to_pfn(OMAP243X_SDRC_PHYS), 133 .pfn = __phys_to_pfn(OMAP243X_SMS_PHYS), 145 .pfn = __phys_to_pfn(L3_34XX_PHYS), [all …]
|
/linux-4.1.27/arch/arm/mm/ |
D | fault-armv.c | 41 unsigned long pfn, pte_t *ptep) in do_adjust_pte() argument 56 flush_cache_page(vma, address, pfn); in do_adjust_pte() 57 outer_flush_range((pfn << PAGE_SHIFT), in do_adjust_pte() 58 (pfn << PAGE_SHIFT) + PAGE_SIZE); in do_adjust_pte() 93 unsigned long pfn) in adjust_pte() argument 123 ret = do_adjust_pte(vma, address, pfn, pte); in adjust_pte() 133 unsigned long addr, pte_t *ptep, unsigned long pfn) in make_coherent() argument 160 aliases += adjust_pte(mpnt, mpnt->vm_start + offset, pfn); in make_coherent() 164 do_adjust_pte(vma, addr, pfn, ptep); in make_coherent() 183 unsigned long pfn = pte_pfn(*ptep); in update_mmu_cache() local [all …]
|
D | flush.c | 26 static void flush_pfn_alias(unsigned long pfn, unsigned long vaddr) in flush_pfn_alias() argument 31 set_top_pte(to, pfn_pte(pfn, PAGE_KERNEL)); in flush_pfn_alias() 40 static void flush_icache_alias(unsigned long pfn, unsigned long vaddr, unsigned long len) in flush_icache_alias() argument 46 set_top_pte(va, pfn_pte(pfn, PAGE_KERNEL)); in flush_icache_alias() 86 void flush_cache_page(struct vm_area_struct *vma, unsigned long user_addr, unsigned long pfn) in flush_cache_page() argument 89 vivt_flush_cache_page(vma, user_addr, pfn); in flush_cache_page() 94 flush_pfn_alias(pfn, user_addr); in flush_cache_page() 103 #define flush_pfn_alias(pfn,vaddr) do { } while (0) argument 104 #define flush_icache_alias(pfn,vaddr,len) do { } while (0) argument 260 unsigned long pfn; in __sync_icache_dcache() local [all …]
|
D | ioremap.c | 186 remap_area_sections(unsigned long virt, unsigned long pfn, in remap_area_sections() argument 204 pmd[0] = __pmd(__pfn_to_phys(pfn) | type->prot_sect); in remap_area_sections() 205 pfn += SZ_1M >> PAGE_SHIFT; in remap_area_sections() 206 pmd[1] = __pmd(__pfn_to_phys(pfn) | type->prot_sect); in remap_area_sections() 207 pfn += SZ_1M >> PAGE_SHIFT; in remap_area_sections() 218 remap_area_supersections(unsigned long virt, unsigned long pfn, in remap_area_supersections() argument 238 super_pmd_val = __pfn_to_phys(pfn) | type->prot_sect | in remap_area_supersections() 240 super_pmd_val |= ((pfn >> (32 - PAGE_SHIFT)) & 0xf) << 20; in remap_area_supersections() 251 pfn += SUPERSECTION_SIZE >> PAGE_SHIFT; in remap_area_supersections() 258 void __iomem * __arm_ioremap_pfn_caller(unsigned long pfn, in __arm_ioremap_pfn_caller() argument [all …]
|
D | mmu.c | 659 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument 662 if (!pfn_valid(pfn)) in phys_mem_access_prot() 696 unsigned long end, unsigned long pfn, in alloc_init_pte() argument 701 set_pte_ext(pte, pfn_pte(pfn, __pgprot(type->prot_pte)), 0); in alloc_init_pte() 702 pfn++; in alloc_init_pte() 787 phys = __pfn_to_phys(md->pfn); in create_36bit_mapping() 792 (long long)__pfn_to_phys((u64)md->pfn), addr); in create_36bit_mapping() 804 (long long)__pfn_to_phys((u64)md->pfn), addr); in create_36bit_mapping() 808 if ((addr | length | __pfn_to_phys(md->pfn)) & ~SUPERSECTION_MASK) { in create_36bit_mapping() 810 (long long)__pfn_to_phys((u64)md->pfn), addr); in create_36bit_mapping() [all …]
|
D | mmap.c | 215 int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) in valid_mmap_phys_addr_range() argument 217 return (pfn + (size >> PAGE_SHIFT)) <= (1 + (PHYS_MASK >> PAGE_SHIFT)); in valid_mmap_phys_addr_range() 231 int devmem_is_allowed(unsigned long pfn) in devmem_is_allowed() argument 233 if (iomem_is_exclusive(pfn << PAGE_SHIFT)) in devmem_is_allowed() 235 if (!page_is_ram(pfn)) in devmem_is_allowed()
|
D | nommu.c | 354 void __iomem *__arm_ioremap_pfn(unsigned long pfn, unsigned long offset, in __arm_ioremap_pfn() argument 357 if (pfn >= (0x100000000ULL >> PAGE_SHIFT)) in __arm_ioremap_pfn() 359 return (void __iomem *) (offset + (pfn << PAGE_SHIFT)); in __arm_ioremap_pfn() 363 void __iomem *__arm_ioremap_pfn_caller(unsigned long pfn, unsigned long offset, in __arm_ioremap_pfn_caller() argument 366 return __arm_ioremap_pfn(pfn, offset, size, mtype); in __arm_ioremap_pfn_caller()
|
D | highmem.c | 127 void *kmap_atomic_pfn(unsigned long pfn) in kmap_atomic_pfn() argument 131 struct page *page = pfn_to_page(pfn); in kmap_atomic_pfn() 143 set_fixmap_pte(idx, pfn_pte(pfn, kmap_prot)); in kmap_atomic_pfn()
|
/linux-4.1.27/arch/ia64/include/asm/ |
D | page.h | 98 extern int ia64_pfn_valid (unsigned long pfn); 100 # define ia64_pfn_valid(pfn) 1 argument 107 # define pfn_to_page(pfn) (vmem_map + (pfn)) argument 116 # define pfn_valid(pfn) (((pfn) < max_mapnr) && ia64_pfn_valid(pfn)) argument 120 # define pfn_valid(pfn) (((pfn) >= min_low_pfn) && ((pfn) < max_low_pfn) && ia64_pfn_valid(pfn)) argument 125 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
|
D | mmzone.h | 20 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument 23 int nid = paddr_to_nid(pfn << PAGE_SHIFT); in pfn_to_nid()
|
/linux-4.1.27/arch/x86/kvm/ |
D | iommu.c | 50 pfn_t pfn; in kvm_pin_pages() local 52 pfn = gfn_to_pfn_memslot(slot, gfn); in kvm_pin_pages() 56 if (is_error_noslot_pfn(pfn)) in kvm_pin_pages() 57 return pfn; in kvm_pin_pages() 62 return pfn; in kvm_pin_pages() 65 static void kvm_unpin_pages(struct kvm *kvm, pfn_t pfn, unsigned long npages) in kvm_unpin_pages() argument 70 kvm_release_pfn_clean(pfn + i); in kvm_unpin_pages() 76 pfn_t pfn; in kvm_iommu_map_pages() local 123 pfn = kvm_pin_pages(slot, gfn, page_size >> PAGE_SHIFT); in kvm_iommu_map_pages() 124 if (is_error_noslot_pfn(pfn)) { in kvm_iommu_map_pages() [all …]
|
D | mmu_audit.c | 100 pfn_t pfn; in audit_mappings() local 117 pfn = gfn_to_pfn_atomic(vcpu->kvm, gfn); in audit_mappings() 119 if (is_error_pfn(pfn)) in audit_mappings() 122 hpa = pfn << PAGE_SHIFT; in audit_mappings() 125 "ent %llxn", vcpu->arch.mmu.root_level, pfn, in audit_mappings()
|
D | mmu.c | 262 pfn_t pfn, unsigned access) in set_mmio_spte() argument 264 if (unlikely(is_noslot_pfn(pfn))) { in set_mmio_spte() 590 pfn_t pfn; in mmu_spte_clear_track_bits() local 601 pfn = spte_to_pfn(old_spte); in mmu_spte_clear_track_bits() 608 WARN_ON(!kvm_is_reserved_pfn(pfn) && !page_count(pfn_to_page(pfn))); in mmu_spte_clear_track_bits() 611 kvm_set_pfn_accessed(pfn); in mmu_spte_clear_track_bits() 613 kvm_set_pfn_dirty(pfn); in mmu_spte_clear_track_bits() 2515 gfn_t gfn, pfn_t pfn, bool speculative, in set_spte() argument 2521 if (set_mmio_spte(vcpu->kvm, sptep, gfn, pfn, pte_access)) in set_spte() 2540 kvm_is_reserved_pfn(pfn)); in set_spte() [all …]
|
/linux-4.1.27/arch/tile/mm/ |
D | init.c | 93 pte_t *get_prealloc_pte(unsigned long pfn) in get_prealloc_pte() argument 95 int node = pfn_to_nid(pfn); in get_prealloc_pte() 96 pfn &= ~(-1UL << (NR_PA_HIGHBIT_SHIFT - PAGE_SHIFT)); in get_prealloc_pte() 98 BUG_ON(pfn >= num_l2_ptes[node]); in get_prealloc_pte() 99 return &l2_ptes[node][pfn]; in get_prealloc_pte() 408 unsigned long address, pfn; in kernel_physical_mapping_init() local 453 pfn = node_start_pfn[i]; in kernel_physical_mapping_init() 456 init_prealloc_ptes(i, end_pfn - pfn); in kernel_physical_mapping_init() 458 address = (unsigned long) pfn_to_kaddr(pfn); in kernel_physical_mapping_init() 459 while (pfn < end_pfn) { in kernel_physical_mapping_init() [all …]
|
/linux-4.1.27/arch/x86/mm/ |
D | init_32.c | 261 unsigned long pfn; in kernel_physical_mapping_init() local 292 pfn = start_pfn; in kernel_physical_mapping_init() 293 pgd_idx = pgd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET); in kernel_physical_mapping_init() 298 if (pfn >= end_pfn) in kernel_physical_mapping_init() 301 pmd_idx = pmd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET); in kernel_physical_mapping_init() 306 for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn; in kernel_physical_mapping_init() 308 unsigned int addr = pfn * PAGE_SIZE + PAGE_OFFSET; in kernel_physical_mapping_init() 325 pfn &= PMD_MASK >> PAGE_SHIFT; in kernel_physical_mapping_init() 326 addr2 = (pfn + PTRS_PER_PTE-1) * PAGE_SIZE + in kernel_physical_mapping_init() 335 set_pmd(pmd, pfn_pmd(pfn, init_prot)); in kernel_physical_mapping_init() [all …]
|
D | pat.c | 336 u64 pfn; in reserve_ram_pages_type() local 344 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in reserve_ram_pages_type() 347 page = pfn_to_page(pfn); in reserve_ram_pages_type() 362 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in reserve_ram_pages_type() 363 page = pfn_to_page(pfn); in reserve_ram_pages_type() 372 u64 pfn; in free_ram_pages_type() local 374 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in free_ram_pages_type() 375 page = pfn_to_page(pfn); in free_ram_pages_type() 606 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument 614 static inline int range_is_allowed(unsigned long pfn, unsigned long size) in range_is_allowed() argument [all …]
|
D | numa_32.c | 49 unsigned long pfn; in memory_present() local 57 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { in memory_present() 58 physnode_map[pfn / PAGES_PER_SECTION] = nid; in memory_present() 59 printk(KERN_CONT "%lx ", pfn); in memory_present()
|
D | init.c | 82 unsigned long pfn; in alloc_low_pages() local 103 pfn = ret >> PAGE_SHIFT; in alloc_low_pages() 105 pfn = pgt_buf_end; in alloc_low_pages() 108 pfn << PAGE_SHIFT, (pgt_buf_end << PAGE_SHIFT) - 1); in alloc_low_pages() 114 adr = __va((pfn + i) << PAGE_SHIFT); in alloc_low_pages() 118 return __va(pfn << PAGE_SHIFT); in alloc_low_pages() 266 unsigned long pfn; in split_mem_range() local 272 pfn = start_pfn = PFN_DOWN(start); in split_mem_range() 280 if (pfn == 0) in split_mem_range() 283 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() [all …]
|
D | pageattr.c | 38 unsigned long pfn; member 253 unsigned long pfn) in static_protections() argument 262 if (pcibios_enabled && within(pfn, BIOS_BEGIN >> PAGE_SHIFT, BIOS_END >> PAGE_SHIFT)) in static_protections() 278 if (within(pfn, __pa_symbol(__start_rodata) >> PAGE_SHIFT, in static_protections() 464 unsigned long nextpage_addr, numpages, pmask, psize, addr, pfn; in try_preserve_large_page() local 539 pfn = pte_pfn(old_pte) + ((address & (psize - 1)) >> PAGE_SHIFT); in try_preserve_large_page() 540 cpa->pfn = pfn; in try_preserve_large_page() 542 new_prot = static_protections(req_prot, address, pfn); in try_preserve_large_page() 550 pfn = pte_pfn(old_pte); in try_preserve_large_page() 551 for (i = 0; i < (psize >> PAGE_SHIFT); i++, addr += PAGE_SIZE, pfn++) { in try_preserve_large_page() [all …]
|
D | iomap_32.c | 57 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) in kmap_atomic_prot_pfn() argument 67 set_pte(kmap_pte - idx, pfn_pte(pfn, prot)); in kmap_atomic_prot_pfn() 77 iomap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) in iomap_atomic_prot_pfn() argument 90 return (void __force __iomem *) kmap_atomic_prot_pfn(pfn, prot); in iomap_atomic_prot_pfn()
|
D | pageattr-test.c | 134 unsigned long pfn = prandom_u32() % max_pfn_mapped; in pageattr_test() local 136 addr[i] = (unsigned long)__va(pfn << PAGE_SHIFT); in pageattr_test() 138 len[i] = min_t(unsigned long, len[i], max_pfn_mapped - pfn - 1); in pageattr_test() 162 if (test_bit(pfn + k, bm)) { in pageattr_test() 166 __set_bit(pfn + k, bm); in pageattr_test()
|
/linux-4.1.27/arch/frv/include/asm/ |
D | page.h | 43 #define devmem_is_allowed(pfn) 1 argument 48 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument 55 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument 58 #define pfn_valid(pfn) ((pfn) >= min_low_pfn && (pfn) < max_low_pfn) argument
|
/linux-4.1.27/arch/x86/kernel/ |
D | crash_dump_32.c | 17 static inline bool is_crashed_pfn_valid(unsigned long pfn) in is_crashed_pfn_valid() argument 27 return pte_pfn(pfn_pte(pfn, __pgprot(0))) == pfn; in is_crashed_pfn_valid() 50 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument 58 if (!is_crashed_pfn_valid(pfn)) in copy_oldmem_page() 61 vaddr = kmap_atomic_pfn(pfn); in copy_oldmem_page()
|
D | crash_dump_64.c | 26 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument 34 vaddr = ioremap_cache(pfn << PAGE_SHIFT, PAGE_SIZE); in copy_oldmem_page()
|
/linux-4.1.27/arch/powerpc/kvm/ |
D | e500_mmu_host.c | 166 pfn_t pfn; in kvmppc_map_magic() local 168 pfn = (pfn_t)virt_to_phys((void *)shared_page) >> PAGE_SHIFT; in kvmppc_map_magic() 169 get_page(pfn_to_page(pfn)); in kvmppc_map_magic() 177 magic.mas7_3 = ((u64)pfn << PAGE_SHIFT) | in kvmppc_map_magic() 249 pfn_t pfn, unsigned int wimg) in kvmppc_e500_ref_setup() argument 251 ref->pfn = pfn; in kvmppc_e500_ref_setup() 258 kvm_set_pfn_accessed(pfn); in kvmppc_e500_ref_setup() 261 kvm_set_pfn_dirty(pfn); in kvmppc_e500_ref_setup() 268 trace_kvm_booke206_ref_release(ref->pfn, ref->flags); in kvmppc_e500_ref_release() 312 pfn_t pfn = ref->pfn; in kvmppc_e500_setup_stlbe() local [all …]
|
D | trace_booke.h | 137 TP_PROTO(__u64 pfn, __u32 flags), 138 TP_ARGS(pfn, flags), 141 __field( __u64, pfn ) 146 __entry->pfn = pfn; 151 __entry->pfn, __entry->flags)
|
D | book3s_64_mmu_host.c | 100 unsigned long pfn; in kvmppc_mmu_map_page() local 107 pfn = kvmppc_gpa_to_pfn(vcpu, orig_pte->raddr, iswrite, &writable); in kvmppc_mmu_map_page() 108 if (is_error_noslot_pfn(pfn)) { in kvmppc_mmu_map_page() 114 hpaddr = pfn << PAGE_SHIFT; in kvmppc_mmu_map_page() 134 kvm_set_pfn_accessed(pfn); in kvmppc_mmu_map_page() 139 kvm_set_pfn_dirty(pfn); in kvmppc_mmu_map_page() 145 kvmppc_mmu_flush_icache(pfn); in kvmppc_mmu_map_page() 199 cpte->pfn = pfn; in kvmppc_mmu_map_page() 208 kvm_release_pfn_clean(pfn); in kvmppc_mmu_map_page()
|
D | trace_pr.h | 70 __field( u64, pfn ) 79 __entry->pfn = pte->pfn; 89 __entry->host_vpn, __entry->pfn, __entry->eaddr, 99 __field( u64, pfn ) 108 __entry->pfn = pte->pfn; 118 __entry->host_vpn, __entry->pfn, __entry->eaddr,
|
/linux-4.1.27/arch/s390/pci/ |
D | pci_mmio.c | 15 unsigned long *pfn) in get_pfn() argument 28 ret = follow_pfn(vma, user_addr, pfn); in get_pfn() 40 unsigned long pfn; in SYSCALL_DEFINE3() local 55 ret = get_pfn(mmio_addr, VM_WRITE, &pfn); in SYSCALL_DEFINE3() 58 io_addr = (void __iomem *)((pfn << PAGE_SHIFT) | (mmio_addr & ~PAGE_MASK)); in SYSCALL_DEFINE3() 80 unsigned long pfn; in SYSCALL_DEFINE3() local 95 ret = get_pfn(mmio_addr, VM_READ, &pfn); in SYSCALL_DEFINE3() 98 io_addr = (void __iomem *)((pfn << PAGE_SHIFT) | (mmio_addr & ~PAGE_MASK)); in SYSCALL_DEFINE3()
|
/linux-4.1.27/arch/powerpc/mm/ |
D | mem.c | 79 int page_is_ram(unsigned long pfn) in page_is_ram() argument 82 return pfn < max_pfn; in page_is_ram() 84 unsigned long paddr = (pfn << PAGE_SHIFT); in page_is_ram() 94 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument 98 return ppc_md.phys_mem_access_prot(file, pfn, size, vma_prot); in phys_mem_access_prot() 100 if (!page_is_ram(pfn)) in phys_mem_access_prot() 341 unsigned long pfn, highmem_mapnr; in mem_init() local 344 for (pfn = highmem_mapnr; pfn < max_mapnr; ++pfn) { in mem_init() 345 phys_addr_t paddr = (phys_addr_t)pfn << PAGE_SHIFT; in mem_init() 346 struct page *page = pfn_to_page(pfn); in mem_init() [all …]
|
/linux-4.1.27/arch/tile/include/asm/ |
D | page.h | 153 #define __pfn_to_highbits(pfn) ((pfn) >> (NR_PA_HIGHBIT_SHIFT - PAGE_SHIFT)) argument 253 static inline void *pfn_to_kaddr(unsigned long pfn) in pfn_to_kaddr() argument 255 return vbase_map[__pfn_to_highbits(pfn)] + (pfn << PAGE_SHIFT); in pfn_to_kaddr() 260 unsigned long pfn = kaddr_to_pfn(kaddr); in virt_to_phys() local 261 return ((phys_addr_t)pfn << PAGE_SHIFT) + in virt_to_phys() 284 static inline void *pfn_to_kaddr(unsigned long pfn) in pfn_to_kaddr() argument 286 return (void *)((pfn << PAGE_SHIFT) + PAGE_OFFSET); in pfn_to_kaddr() 311 static inline int pfn_valid(unsigned long pfn) in pfn_valid() argument 313 return pfn < max_mapnr; in pfn_valid()
|
D | mmzone.h | 38 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument 40 return highbits_to_node[__pfn_to_highbits(pfn)]; in pfn_to_nid() 45 static inline int pfn_valid(unsigned long pfn) in pfn_valid() argument 47 int nid = pfn_to_nid(pfn); in pfn_valid() 50 return (pfn < node_end_pfn(nid)); in pfn_valid()
|
D | highmem.h | 64 void *kmap_atomic_pfn(unsigned long pfn); 65 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot);
|
/linux-4.1.27/arch/microblaze/include/asm/ |
D | page.h | 142 extern int page_is_ram(unsigned long pfn); 145 # define pfn_to_phys(pfn) (PFN_PHYS(pfn)) argument 148 # define pfn_to_virt(pfn) __va(pfn_to_phys((pfn))) argument 165 # define pfn_valid(pfn) (((pfn) >= min_low_pfn) && \ argument 166 ((pfn) <= (min_low_pfn + max_mapnr))) 170 # define pfn_valid(pfn) ((pfn) < (max_mapnr + ARCH_PFN_OFFSET)) argument
|
D | cacheflush.h | 92 #define flush_cache_page(vma, vmaddr, pfn) \ argument 93 flush_dcache_range(pfn << PAGE_SHIFT, (pfn << PAGE_SHIFT) + PAGE_SIZE);
|
/linux-4.1.27/drivers/xen/ |
D | balloon.c | 243 unsigned long pfn, i; in reserve_additional_memory() local 245 pfn = PFN_DOWN(hotplug_start_paddr); in reserve_additional_memory() 247 if (!set_phys_to_machine(pfn + i, INVALID_P2M_ENTRY)) { in reserve_additional_memory() 329 unsigned long pfn, i; in increase_reservation() local 369 pfn = page_to_pfn(page); in increase_reservation() 373 set_phys_to_machine(pfn, frame_list[i]); in increase_reservation() 379 (unsigned long)__va(pfn << PAGE_SHIFT), in increase_reservation() 399 unsigned long pfn, i; in decrease_reservation() local 443 pfn = frame_list[i]; in decrease_reservation() 444 frame_list[i] = pfn_to_mfn(pfn); in decrease_reservation() [all …]
|
D | tmem.c | 132 u32 index, unsigned long pfn) in xen_tmem_put_page() argument 134 unsigned long gmfn = xen_pv_domain() ? pfn_to_mfn(pfn) : pfn; in xen_tmem_put_page() 141 u32 index, unsigned long pfn) in xen_tmem_get_page() argument 143 unsigned long gmfn = xen_pv_domain() ? pfn_to_mfn(pfn) : pfn; in xen_tmem_get_page() 176 unsigned long pfn = page_to_pfn(page); in tmem_cleancache_put_page() local 183 (void)xen_tmem_put_page((u32)pool, oid, ind, pfn); in tmem_cleancache_put_page() 191 unsigned long pfn = page_to_pfn(page); in tmem_cleancache_get_page() local 199 ret = xen_tmem_get_page((u32)pool, oid, ind, pfn); in tmem_cleancache_get_page() 290 unsigned long pfn = page_to_pfn(page); in tmem_frontswap_store() local 299 ret = xen_tmem_put_page(pool, oswiz(type, ind), iswiz(ind), pfn); in tmem_frontswap_store() [all …]
|
D | xlate_mmu.c | 81 unsigned long pfn = page_to_pfn(page); in remap_pte_fn() local 82 pte_t pte = pte_mkspecial(pfn_pte(pfn, info->prot)); in remap_pte_fn() 85 rc = map_foreign_page(pfn, *info->fgmfn, info->domid); in remap_pte_fn() 133 unsigned long pfn; in xen_xlate_unmap_gfn_range() local 135 pfn = page_to_pfn(pages[i]); in xen_xlate_unmap_gfn_range() 138 xrp.gpfn = pfn; in xen_xlate_unmap_gfn_range()
|
D | swiotlb-xen.c | 95 unsigned long pfn = mfn_to_pfn(PFN_DOWN(baddr)); in xen_bus_to_phys() local 96 dma_addr_t dma = (dma_addr_t)pfn << PAGE_SHIFT; in xen_bus_to_phys() 109 static int check_pages_physically_contiguous(unsigned long pfn, in check_pages_physically_contiguous() argument 117 next_mfn = pfn_to_mfn(pfn); in check_pages_physically_contiguous() 121 if (pfn_to_mfn(++pfn) != ++next_mfn) in check_pages_physically_contiguous() 129 unsigned long pfn = PFN_DOWN(p); in range_straddles_page_boundary() local 134 if (check_pages_physically_contiguous(pfn, offset, size)) in range_straddles_page_boundary() 142 unsigned long pfn = mfn_to_local_pfn(mfn); in is_xen_swiotlb_buffer() local 149 if (pfn_valid(pfn)) { in is_xen_swiotlb_buffer() 150 paddr = PFN_PHYS(pfn); in is_xen_swiotlb_buffer()
|
D | grant-table.c | 395 int gnttab_grant_foreign_transfer(domid_t domid, unsigned long pfn) in gnttab_grant_foreign_transfer() argument 402 gnttab_grant_foreign_transfer_ref(ref, domid, pfn); in gnttab_grant_foreign_transfer() 409 unsigned long pfn) in gnttab_grant_foreign_transfer_ref() argument 411 gnttab_interface->update_entry(ref, domid, pfn, GTF_accept_transfer); in gnttab_grant_foreign_transfer_ref() 638 xen_pfn_t *pfn; in gnttab_setup_auto_xlat_frames() local 652 pfn = kcalloc(max_nr_gframes, sizeof(pfn[0]), GFP_KERNEL); in gnttab_setup_auto_xlat_frames() 653 if (!pfn) { in gnttab_setup_auto_xlat_frames() 658 pfn[i] = PFN_DOWN(addr) + i; in gnttab_setup_auto_xlat_frames() 661 xen_auto_xlat_grant_frames.pfn = pfn; in gnttab_setup_auto_xlat_frames() 672 kfree(xen_auto_xlat_grant_frames.pfn); in gnttab_free_auto_xlat_frames() [all …]
|
/linux-4.1.27/arch/metag/include/asm/ |
D | mmzone.h | 10 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument 15 if (pfn >= node_start_pfn(nid) && pfn <= node_end_pfn(nid)) in pfn_to_nid() 21 static inline struct pglist_data *pfn_to_pgdat(unsigned long pfn) in pfn_to_pgdat() argument 23 return NODE_DATA(pfn_to_nid(pfn)); in pfn_to_pgdat()
|
D | page.h | 115 #define pfn_valid(pfn) ((pfn) >= min_low_pfn && (pfn) < max_pfn) argument 118 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
|
/linux-4.1.27/arch/m68k/include/asm/ |
D | page_no.h | 26 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument 31 #define pfn_to_page(pfn) virt_to_page(pfn_to_virt(pfn)) argument 33 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
|
D | page_mm.h | 127 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument 158 #define pfn_to_page(pfn) ({ \ argument 159 unsigned long __pfn = (pfn); \ 172 #define pfn_valid(pfn) virt_addr_valid(pfn_to_virt(pfn)) argument
|
/linux-4.1.27/arch/sh/include/asm/ |
D | mmzone.h | 12 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument 17 if (pfn >= node_start_pfn(nid) && pfn <= node_end_pfn(nid)) in pfn_to_nid() 23 static inline struct pglist_data *pfn_to_pgdat(unsigned long pfn) in pfn_to_pgdat() argument 25 return NODE_DATA(pfn_to_nid(pfn)); in pfn_to_pgdat()
|
D | page.h | 166 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument 179 #define pfn_valid(pfn) ((pfn) >= min_low_pfn && (pfn) < max_low_pfn) argument
|
/linux-4.1.27/arch/arm/mach-omap1/ |
D | io.c | 34 .pfn = __phys_to_pfn(OMAP1_IO_PHYS), 44 .pfn = __phys_to_pfn(OMAP7XX_DSP_START), 49 .pfn = __phys_to_pfn(OMAP7XX_DSPREG_START), 60 .pfn = __phys_to_pfn(OMAP1510_DSP_START), 65 .pfn = __phys_to_pfn(OMAP1510_DSPREG_START), 76 .pfn = __phys_to_pfn(OMAP16XX_DSP_START), 81 .pfn = __phys_to_pfn(OMAP16XX_DSPREG_START),
|
/linux-4.1.27/arch/score/include/asm/ |
D | page.h | 65 #define pfn_to_phys(pfn) (PFN_PHYS(pfn)) argument 68 #define pfn_to_virt(pfn) __va(pfn_to_phys((pfn))) argument 77 #define pfn_valid(pfn) (((pfn) >= min_low_pfn) && ((pfn) < max_low_pfn)) argument
|
/linux-4.1.27/drivers/base/ |
D | node.c | 362 static int get_nid_for_pfn(unsigned long pfn) in get_nid_for_pfn() argument 366 if (!pfn_valid_within(pfn)) in get_nid_for_pfn() 368 page = pfn_to_page(pfn); in get_nid_for_pfn() 371 return pfn_to_nid(pfn); in get_nid_for_pfn() 378 unsigned long pfn, sect_start_pfn, sect_end_pfn; in register_mem_sect_under_node() local 388 for (pfn = sect_start_pfn; pfn <= sect_end_pfn; pfn++) { in register_mem_sect_under_node() 395 if (!pfn_present(pfn)) { in register_mem_sect_under_node() 396 pfn = round_down(pfn + PAGES_PER_SECTION, in register_mem_sect_under_node() 401 page_nid = get_nid_for_pfn(pfn); in register_mem_sect_under_node() 425 unsigned long pfn, sect_start_pfn, sect_end_pfn; in unregister_mem_sect_under_nodes() local [all …]
|
D | memory.c | 127 unsigned long i, pfn; in show_mem_removable() local 134 pfn = section_nr_to_pfn(mem->start_section_nr + i); in show_mem_removable() 135 ret &= is_mem_section_removable(pfn, PAGES_PER_SECTION); in show_mem_removable() 192 unsigned long pfn = start_pfn; in pages_correctly_reserved() local 199 for (i = 0; i < sections_per_block; i++, pfn += PAGES_PER_SECTION) { in pages_correctly_reserved() 200 if (WARN_ON_ONCE(!pfn_valid(pfn))) in pages_correctly_reserved() 202 page = pfn_to_page(pfn); in pages_correctly_reserved() 210 pfn_to_section_nr(pfn), j); in pages_correctly_reserved() 490 u64 pfn; in store_soft_offline_page() local 493 if (kstrtoull(buf, 0, &pfn) < 0) in store_soft_offline_page() [all …]
|
D | dma-mapping.c | 253 unsigned long pfn = page_to_pfn(virt_to_page(cpu_addr)); in dma_common_mmap() local 263 pfn + off, in dma_common_mmap() 310 unsigned long pfn; in dma_common_contiguous_remap() local 316 for (i = 0, pfn = page_to_pfn(page); i < (size >> PAGE_SHIFT); i++) in dma_common_contiguous_remap() 317 pages[i] = pfn_to_page(pfn + i); in dma_common_contiguous_remap()
|
/linux-4.1.27/fs/proc/ |
D | page.c | 30 unsigned long pfn; in kpagecount_read() local 34 pfn = src / KPMSIZE; in kpagecount_read() 40 if (pfn_valid(pfn)) in kpagecount_read() 41 ppage = pfn_to_page(pfn); in kpagecount_read() 54 pfn++; in kpagecount_read() 193 unsigned long pfn; in kpageflags_read() local 196 pfn = src / KPMSIZE; in kpageflags_read() 202 if (pfn_valid(pfn)) in kpageflags_read() 203 ppage = pfn_to_page(pfn); in kpageflags_read() 212 pfn++; in kpageflags_read()
|
D | vmcore.c | 51 static int (*oldmem_pfn_is_ram)(unsigned long pfn); 53 int register_oldmem_pfn_is_ram(int (*fn)(unsigned long pfn)) in register_oldmem_pfn_is_ram() argument 69 static int pfn_is_ram(unsigned long pfn) in pfn_is_ram() argument 71 int (*fn)(unsigned long pfn); in pfn_is_ram() 82 ret = fn(pfn); in pfn_is_ram() 91 unsigned long pfn, offset; in read_from_oldmem() local 99 pfn = (unsigned long)(*ppos / PAGE_SIZE); in read_from_oldmem() 108 if (pfn_is_ram(pfn) == 0) in read_from_oldmem() 111 tmp = copy_oldmem_page(pfn, buf, nr_bytes, in read_from_oldmem() 120 ++pfn; in read_from_oldmem() [all …]
|
/linux-4.1.27/arch/parisc/kernel/ |
D | cache.c | 75 #define pfn_va(pfn) __va(PFN_PHYS(pfn)) argument 80 unsigned long pfn = pte_pfn(*ptep); in update_mmu_cache() local 86 if (!pfn_valid(pfn)) in update_mmu_cache() 89 page = pfn_to_page(pfn); in update_mmu_cache() 91 flush_kernel_dcache_page_addr(pfn_va(pfn)); in update_mmu_cache() 94 flush_kernel_dcache_page_addr(pfn_va(pfn)); in update_mmu_cache() 542 unsigned long pfn; in flush_cache_mm() local 546 pfn = pte_pfn(*ptep); in flush_cache_mm() 547 if (!pfn_valid(pfn)) in flush_cache_mm() 549 __flush_cache_page(vma, addr, PFN_PHYS(pfn)); in flush_cache_mm() [all …]
|
/linux-4.1.27/arch/powerpc/kernel/ |
D | dma.c | 31 u64 pfn = (dev->coherent_dma_mask >> PAGE_SHIFT) + 1; in get_pfn_limit() local 36 pfn = min_t(u64, pfn, sd->max_direct_dma_addr >> PAGE_SHIFT); in get_pfn_limit() 39 return pfn; in get_pfn_limit() 57 u64 pfn = get_pfn_limit(dev); in dma_direct_alloc_coherent() local 66 zone = dma_pfn_limit_to_zone(pfn); in dma_direct_alloc_coherent() 69 __func__, pfn); in dma_direct_alloc_coherent() 114 unsigned long pfn; in dma_direct_mmap_coherent() local 118 pfn = __dma_get_coherent_pfn((unsigned long)cpu_addr); in dma_direct_mmap_coherent() 120 pfn = page_to_pfn(virt_to_page(cpu_addr)); in dma_direct_mmap_coherent() 123 pfn + vma->vm_pgoff, in dma_direct_mmap_coherent()
|
D | suspend.c | 18 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 22 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
|
/linux-4.1.27/arch/arm64/mm/ |
D | mmap.c | 114 int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) in valid_mmap_phys_addr_range() argument 116 return !(((pfn << PAGE_SHIFT) + size) & ~PHYS_MASK); in valid_mmap_phys_addr_range() 129 int devmem_is_allowed(unsigned long pfn) in devmem_is_allowed() argument 131 if (iomem_is_exclusive(pfn << PAGE_SHIFT)) in devmem_is_allowed() 133 if (!page_is_ram(pfn)) in devmem_is_allowed()
|
/linux-4.1.27/arch/arm/mach-exynos/ |
D | exynos.c | 41 .pfn = __phys_to_pfn(EXYNOS4_PA_SROMC), 46 .pfn = __phys_to_pfn(EXYNOS4_PA_CMU), 51 .pfn = __phys_to_pfn(EXYNOS4_PA_COREPERI), 56 .pfn = __phys_to_pfn(EXYNOS4_PA_DMC0), 61 .pfn = __phys_to_pfn(EXYNOS4_PA_DMC1), 70 .pfn = __phys_to_pfn(EXYNOS5_PA_SROMC), 75 .pfn = __phys_to_pfn(EXYNOS5_PA_CMU), 135 iodesc.pfn = __phys_to_pfn(be32_to_cpu(reg[0])); in exynos_fdt_map_chipid()
|
/linux-4.1.27/drivers/vfio/ |
D | vfio_iommu_type1.c | 191 static bool is_invalid_reserved_pfn(unsigned long pfn) in is_invalid_reserved_pfn() argument 193 if (pfn_valid(pfn)) { in is_invalid_reserved_pfn() 195 struct page *tail = pfn_to_page(pfn); in is_invalid_reserved_pfn() 219 static int put_pfn(unsigned long pfn, int prot) in put_pfn() argument 221 if (!is_invalid_reserved_pfn(pfn)) { in put_pfn() 222 struct page *page = pfn_to_page(pfn); in put_pfn() 231 static int vaddr_get_pfn(unsigned long vaddr, int prot, unsigned long *pfn) in vaddr_get_pfn() argument 238 *pfn = page_to_pfn(page[0]); in vaddr_get_pfn() 247 *pfn = ((vaddr - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; in vaddr_get_pfn() 248 if (is_invalid_reserved_pfn(*pfn)) in vaddr_get_pfn() [all …]
|
/linux-4.1.27/arch/mn10300/include/asm/ |
D | page.h | 107 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument 108 #define pfn_to_page(pfn) (mem_map + ((pfn) - __pfn_disp)) argument 111 #define pfn_valid(pfn) \ argument 113 unsigned long __pfn = (pfn) - __pfn_disp; \
|
D | pgtable.h | 317 #define __pfn_addr(pfn) ((pfn) << PAGE_SHIFT) argument 318 #define pfn_pte(pfn, prot) __pte(__pfn_addr(pfn) | pgprot_val(prot)) argument 319 #define pfn_pmd(pfn, prot) __pmd(__pfn_addr(pfn) | pgprot_val(prot)) argument 480 #define MK_IOSPACE_PFN(space, pfn) (pfn) argument 481 #define GET_IOSPACE(pfn) 0 argument 482 #define GET_PFN(pfn) (pfn) argument
|
/linux-4.1.27/arch/arm/mach-footbridge/ |
D | common.c | 143 .pfn = __phys_to_pfn(DC21285_ARMCSR_BASE), 157 .pfn = __phys_to_pfn(DC21285_PCI_MEM), 162 .pfn = __phys_to_pfn(DC21285_PCI_TYPE_0_CONFIG), 167 .pfn = __phys_to_pfn(DC21285_PCI_TYPE_1_CONFIG), 172 .pfn = __phys_to_pfn(DC21285_PCI_IACK), 256 unsigned long __pfn_to_bus(unsigned long pfn) in __pfn_to_bus() argument 258 return __pfn_to_phys(pfn) + (fb_bus_sdram_offset() - PHYS_OFFSET); in __pfn_to_bus()
|
/linux-4.1.27/arch/frv/mm/ |
D | init.c | 123 unsigned long pfn; in mem_init() local 125 for (pfn = get_num_physpages() - 1; in mem_init() 126 pfn >= num_mappedpages; pfn--) in mem_init() 127 free_highmem_page(&mem_map[pfn]); in mem_init()
|
/linux-4.1.27/arch/arm/include/asm/ |
D | dma-mapping.h | 59 static inline dma_addr_t pfn_to_dma(struct device *dev, unsigned long pfn) in pfn_to_dma() argument 62 pfn -= dev->dma_pfn_offset; in pfn_to_dma() 63 return (dma_addr_t)__pfn_to_bus(pfn); in pfn_to_dma() 68 unsigned long pfn = __bus_to_pfn(addr); in dma_to_pfn() local 71 pfn += dev->dma_pfn_offset; in dma_to_pfn() 73 return pfn; in dma_to_pfn() 79 unsigned long pfn = dma_to_pfn(dev, addr); in dma_to_virt() local 81 return phys_to_virt(__pfn_to_phys(pfn)); in dma_to_virt() 96 static inline dma_addr_t pfn_to_dma(struct device *dev, unsigned long pfn) in pfn_to_dma() argument 98 return __arch_pfn_to_dma(dev, pfn); in pfn_to_dma()
|
D | kvm_mmu.h | 185 static inline void __coherent_cache_guest_page(struct kvm_vcpu *vcpu, pfn_t pfn, in __coherent_cache_guest_page() argument 215 void *va = kmap_atomic_pfn(pfn); in __coherent_cache_guest_page() 225 pfn++; in __coherent_cache_guest_page() 249 pfn_t pfn = pmd_pfn(pmd); in __kvm_flush_dcache_pmd() local 252 void *va = kmap_atomic_pfn(pfn); in __kvm_flush_dcache_pmd() 256 pfn++; in __kvm_flush_dcache_pmd()
|
D | memory.h | 127 #define __pfn_to_phys(pfn) ((phys_addr_t)(pfn) << PAGE_SHIFT) argument 294 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
|
/linux-4.1.27/arch/arm/mach-tegra/ |
D | io.c | 36 .pfn = __phys_to_pfn(IO_PPSB_PHYS), 42 .pfn = __phys_to_pfn(IO_APB_PHYS), 48 .pfn = __phys_to_pfn(IO_CPU_PHYS), 54 .pfn = __phys_to_pfn(IO_IRAM_PHYS),
|
/linux-4.1.27/arch/mips/include/asm/ |
D | page.h | 195 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument 199 static inline int pfn_valid(unsigned long pfn) in pfn_valid() argument 204 return pfn >= ARCH_PFN_OFFSET && pfn < max_mapnr; in pfn_valid() 213 #define pfn_valid(pfn) \ argument 215 unsigned long __pfn = (pfn); \
|
D | pgtable-32.h | 110 pfn_pte(unsigned long pfn, pgprot_t prot) in pfn_pte() argument 114 pte.pte_low = (pfn >> _PAGE_PRESENT_SHIFT) | in pfn_pte() 116 pte.pte_high = (pfn << _PFN_SHIFT) | in pfn_pte() 127 #define pfn_pte(pfn, prot) __pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot)) argument 130 #define pfn_pte(pfn, prot) __pte(((unsigned long long)(pfn) << _PFN_SHIFT) | pgprot_val(prot)) argument
|
D | pgtable-64.h | 229 #define pfn_pte(pfn, prot) __pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot)) argument 232 #define pfn_pte(pfn, prot) __pte(((pfn) << _PFN_SHIFT) | pgprot_val(prot)) argument 233 #define pfn_pmd(pfn, prot) __pmd(((pfn) << _PFN_SHIFT) | pgprot_val(prot)) argument
|
D | mmzone.h | 13 #define pfn_to_nid(pfn) pa_to_nid((pfn) << PAGE_SHIFT) argument
|
/linux-4.1.27/arch/arm/kvm/ |
D | mmu.c | 101 static bool kvm_is_device_pfn(unsigned long pfn) in kvm_is_device_pfn() argument 103 return !pfn_valid(pfn); in kvm_is_device_pfn() 452 unsigned long end, unsigned long pfn, in create_hyp_pte_mappings() argument 461 kvm_set_pte(pte, pfn_pte(pfn, prot)); in create_hyp_pte_mappings() 464 pfn++; in create_hyp_pte_mappings() 469 unsigned long end, unsigned long pfn, in create_hyp_pmd_mappings() argument 495 create_hyp_pte_mappings(pmd, addr, next, pfn, prot); in create_hyp_pmd_mappings() 496 pfn += (next - addr) >> PAGE_SHIFT; in create_hyp_pmd_mappings() 503 unsigned long end, unsigned long pfn, in create_hyp_pud_mappings() argument 527 ret = create_hyp_pmd_mappings(pud, addr, next, pfn, prot); in create_hyp_pud_mappings() [all …]
|
/linux-4.1.27/arch/avr32/include/asm/ |
D | pgtable-2level.h | 42 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument 43 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument
|
D | page.h | 88 #define pfn_to_page(pfn) (mem_map + ((pfn) - PHYS_PFN_OFFSET)) argument 90 #define pfn_valid(pfn) ((pfn) >= PHYS_PFN_OFFSET && (pfn) < (PHYS_PFN_OFFSET + max_mapnr)) argument
|
/linux-4.1.27/arch/arm/mach-cns3xxx/ |
D | core.c | 32 .pfn = __phys_to_pfn(CNS3XXX_TC11MP_SCU_BASE), 37 .pfn = __phys_to_pfn(CNS3XXX_TIMER1_2_3_BASE), 42 .pfn = __phys_to_pfn(CNS3XXX_MISC_BASE), 47 .pfn = __phys_to_pfn(CNS3XXX_PM_BASE), 53 .pfn = __phys_to_pfn(CNS3XXX_PCIE0_HOST_BASE), 58 .pfn = __phys_to_pfn(CNS3XXX_PCIE0_CFG0_BASE), 63 .pfn = __phys_to_pfn(CNS3XXX_PCIE0_CFG1_BASE), 68 .pfn = __phys_to_pfn(CNS3XXX_PCIE1_HOST_BASE), 73 .pfn = __phys_to_pfn(CNS3XXX_PCIE1_CFG0_BASE), 78 .pfn = __phys_to_pfn(CNS3XXX_PCIE1_CFG1_BASE),
|
/linux-4.1.27/arch/microblaze/mm/ |
D | init.c | 76 unsigned long pfn; in highmem_setup() local 78 for (pfn = max_low_pfn; pfn < max_pfn; ++pfn) { in highmem_setup() 79 struct page *page = pfn_to_page(pfn); in highmem_setup() 82 if (!memblock_is_reserved(pfn << PAGE_SHIFT)) in highmem_setup() 267 int page_is_ram(unsigned long pfn) in page_is_ram() argument 269 return __range_ok(pfn, 0); in page_is_ram() 272 int page_is_ram(unsigned long pfn) in page_is_ram() argument 274 return pfn < max_low_pfn; in page_is_ram()
|
D | consistent.c | 204 unsigned long pfn; in consistent_free() local 207 pfn = pte_pfn(*ptep); in consistent_free() 209 if (pfn_valid(pfn)) { in consistent_free() 210 page = pfn_to_page(pfn); in consistent_free()
|
/linux-4.1.27/arch/arm/mach-realview/ |
D | realview_pbx.c | 53 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE), 58 .pfn = __phys_to_pfn(REALVIEW_PBX_GIC_CPU_BASE), 63 .pfn = __phys_to_pfn(REALVIEW_PBX_GIC_DIST_BASE), 68 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE), 73 .pfn = __phys_to_pfn(REALVIEW_PBX_TIMER0_1_BASE), 78 .pfn = __phys_to_pfn(REALVIEW_PBX_TIMER2_3_BASE), 85 .pfn = __phys_to_pfn(REALVIEW_PBX_PCI_BASE), 93 .pfn = __phys_to_pfn(REALVIEW_PBX_UART0_BASE), 103 .pfn = __phys_to_pfn(REALVIEW_PBX_TILE_SCU_BASE), 108 .pfn = __phys_to_pfn(REALVIEW_PBX_TILE_GIC_DIST_BASE), [all …]
|
D | realview_pb1176.c | 56 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE), 61 .pfn = __phys_to_pfn(REALVIEW_PB1176_GIC_CPU_BASE), 66 .pfn = __phys_to_pfn(REALVIEW_PB1176_GIC_DIST_BASE), 71 .pfn = __phys_to_pfn(REALVIEW_DC1176_GIC_CPU_BASE), 76 .pfn = __phys_to_pfn(REALVIEW_DC1176_GIC_DIST_BASE), 81 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE), 86 .pfn = __phys_to_pfn(REALVIEW_PB1176_TIMER0_1_BASE), 91 .pfn = __phys_to_pfn(REALVIEW_PB1176_TIMER2_3_BASE), 96 .pfn = __phys_to_pfn(REALVIEW_PB1176_L220_BASE), 103 .pfn = __phys_to_pfn(REALVIEW_PB1176_UART0_BASE),
|
D | realview_pba8.c | 51 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE), 56 .pfn = __phys_to_pfn(REALVIEW_PBA8_GIC_CPU_BASE), 61 .pfn = __phys_to_pfn(REALVIEW_PBA8_GIC_DIST_BASE), 66 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE), 71 .pfn = __phys_to_pfn(REALVIEW_PBA8_TIMER0_1_BASE), 76 .pfn = __phys_to_pfn(REALVIEW_PBA8_TIMER2_3_BASE), 83 .pfn = __phys_to_pfn(REALVIEW_PBA8_PCI_BASE), 91 .pfn = __phys_to_pfn(REALVIEW_PBA8_UART0_BASE),
|
D | realview_pb11mp.c | 54 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE), 59 .pfn = __phys_to_pfn(REALVIEW_PB11MP_GIC_CPU_BASE), 64 .pfn = __phys_to_pfn(REALVIEW_PB11MP_GIC_DIST_BASE), 69 .pfn = __phys_to_pfn(REALVIEW_TC11MP_PRIV_MEM_BASE), 74 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE), 79 .pfn = __phys_to_pfn(REALVIEW_PB11MP_TIMER0_1_BASE), 84 .pfn = __phys_to_pfn(REALVIEW_PB11MP_TIMER2_3_BASE), 89 .pfn = __phys_to_pfn(REALVIEW_TC11MP_L220_BASE), 96 .pfn = __phys_to_pfn(REALVIEW_PB11MP_UART0_BASE),
|
D | realview_eb.c | 54 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE), 59 .pfn = __phys_to_pfn(REALVIEW_EB_GIC_CPU_BASE), 64 .pfn = __phys_to_pfn(REALVIEW_EB_GIC_DIST_BASE), 69 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE), 74 .pfn = __phys_to_pfn(REALVIEW_EB_TIMER0_1_BASE), 79 .pfn = __phys_to_pfn(REALVIEW_EB_TIMER2_3_BASE), 86 .pfn = __phys_to_pfn(REALVIEW_EB_UART0_BASE), 96 .pfn = __phys_to_pfn(REALVIEW_EB11MP_PRIV_MEM_BASE), 101 .pfn = __phys_to_pfn(REALVIEW_EB11MP_L220_BASE),
|
/linux-4.1.27/arch/um/include/asm/ |
D | pgtable-2level.h | 41 #define pfn_pte(pfn, prot) __pte(pfn_to_phys(pfn) | pgprot_val(prot)) argument 42 #define pfn_pmd(pfn, prot) __pmd(pfn_to_phys(pfn) | pgprot_val(prot)) argument
|
D | page.h | 113 #define pfn_to_phys(pfn) ((phys_t) ((pfn) << PAGE_SHIFT)) argument 115 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
|
/linux-4.1.27/arch/nios2/include/asm/ |
D | page.h | 89 # define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument 90 # define pfn_valid(pfn) ((pfn) >= ARCH_PFN_OFFSET && \ argument 91 (pfn) < max_mapnr)
|
/linux-4.1.27/arch/sparc/include/asm/ |
D | pgtable_32.h | 262 #define pfn_pte(pfn, prot) mk_pte(pfn_to_page(pfn), prot) argument 405 #define MK_IOSPACE_PFN(space, pfn) (pfn | (space << (BITS_PER_LONG - 4))) argument 406 #define GET_IOSPACE(pfn) (pfn >> (BITS_PER_LONG - 4)) argument 407 #define GET_PFN(pfn) (pfn & 0x0fffffffUL) argument 413 unsigned long from, unsigned long pfn, in io_remap_pfn_range() argument 418 offset = ((unsigned long long) GET_PFN(pfn)) << PAGE_SHIFT; in io_remap_pfn_range() 419 space = GET_IOSPACE(pfn); in io_remap_pfn_range()
|
D | pgtable_64.h | 237 static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot) in pfn_pte() argument 239 unsigned long paddr = pfn << PAGE_SHIFT; in pfn_pte() 801 unsigned long pfn; in __pmd_page() local 803 pfn = pte_pfn(pte); in __pmd_page() 805 return ((unsigned long) __va(pfn << PAGE_SHIFT)); in __pmd_page() 966 #define MK_IOSPACE_PFN(space, pfn) (pfn | (space << (BITS_PER_LONG - 4))) argument 967 #define GET_IOSPACE(pfn) (pfn >> (BITS_PER_LONG - 4)) argument 968 #define GET_PFN(pfn) (pfn & 0x0fffffffffffffffUL) argument 974 unsigned long from, unsigned long pfn, in io_remap_pfn_range() argument 977 unsigned long offset = GET_PFN(pfn) << PAGE_SHIFT; in io_remap_pfn_range() [all …]
|
/linux-4.1.27/arch/arm/mach-s3c64xx/ |
D | common.c | 109 .pfn = __phys_to_pfn(S3C64XX_PA_SYSCON), 114 .pfn = __phys_to_pfn(S3C64XX_PA_SROM), 119 .pfn = __phys_to_pfn(S3C_PA_UART), 124 .pfn = __phys_to_pfn(S3C64XX_PA_VIC0), 129 .pfn = __phys_to_pfn(S3C64XX_PA_VIC1), 134 .pfn = __phys_to_pfn(S3C_PA_TIMER), 139 .pfn = __phys_to_pfn(S3C64XX_PA_GPIO), 144 .pfn = __phys_to_pfn(S3C64XX_PA_MODEM), 149 .pfn = __phys_to_pfn(S3C64XX_PA_WATCHDOG), 154 .pfn = __phys_to_pfn(S3C64XX_PA_USB_HSPHY),
|
/linux-4.1.27/drivers/media/v4l2-core/ |
D | videobuf2-dma-contig.c | 440 unsigned long pfn, start_pfn, prev_pfn; in vb2_dc_get_user_pfn() local 447 ret = follow_pfn(vma, start, &pfn); in vb2_dc_get_user_pfn() 451 start_pfn = pfn; in vb2_dc_get_user_pfn() 455 prev_pfn = pfn; in vb2_dc_get_user_pfn() 456 ret = follow_pfn(vma, start, &pfn); in vb2_dc_get_user_pfn() 462 if (pfn != prev_pfn + 1) in vb2_dc_get_user_pfn() 478 unsigned long pfn; in vb2_dc_get_user_pages() local 479 int ret = follow_pfn(vma, start, &pfn); in vb2_dc_get_user_pages() 481 if (!pfn_valid(pfn)) in vb2_dc_get_user_pages() 488 pages[i] = pfn_to_page(pfn); in vb2_dc_get_user_pages() [all …]
|
/linux-4.1.27/arch/arm/mach-integrator/ |
D | integrator_cp.c | 69 .pfn = __phys_to_pfn(INTEGRATOR_HDR_BASE), 74 .pfn = __phys_to_pfn(INTEGRATOR_CT_BASE), 79 .pfn = __phys_to_pfn(INTEGRATOR_IC_BASE), 84 .pfn = __phys_to_pfn(INTEGRATOR_UART0_BASE), 89 .pfn = __phys_to_pfn(INTEGRATOR_DBG_BASE), 94 .pfn = __phys_to_pfn(INTEGRATOR_CP_GPIO_BASE), 99 .pfn = __phys_to_pfn(INTEGRATOR_CP_SIC_BASE),
|
/linux-4.1.27/arch/sh/kernel/ |
D | swsusp.c | 22 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 27 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
|
D | crash_dump.c | 25 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument 33 vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); in copy_oldmem_page()
|
/linux-4.1.27/arch/unicore32/include/asm/ |
D | memory.h | 67 #define __pfn_to_phys(pfn) ((pfn) << PAGE_SHIFT) argument 96 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
|
D | io.h | 63 static inline int devmem_is_allowed(unsigned long pfn) in devmem_is_allowed() argument 65 if (iomem_is_exclusive(pfn << PAGE_SHIFT)) in devmem_is_allowed() 67 if (!page_is_ram(pfn)) in devmem_is_allowed()
|
/linux-4.1.27/arch/sparc/power/ |
D | hibernate.c | 21 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 26 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
|
/linux-4.1.27/arch/arm/mach-spear/ |
D | spear13xx.c | 63 .pfn = __phys_to_pfn(PERIP_GRP2_BASE), 68 .pfn = __phys_to_pfn(PERIP_GRP1_BASE), 73 .pfn = __phys_to_pfn(A9SM_AND_MPMC_BASE), 78 .pfn = __phys_to_pfn(L2CC_BASE),
|
/linux-4.1.27/arch/powerpc/include/asm/ |
D | page.h | 127 #define pfn_valid(pfn) ((pfn) >= ARCH_PFN_OFFSET && (pfn) < max_mapnr) argument 131 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument 395 extern int page_is_ram(unsigned long pfn); 396 extern int devmem_is_allowed(unsigned long pfn);
|
D | pte-hash64-64k.h | 97 #define remap_4k_pfn(vma, addr, pfn, prot) \ argument 98 (WARN_ON(((pfn) >= (1UL << (64 - PTE_RPN_SHIFT)))) ? -EINVAL : \ 99 remap_pfn_range((vma), (addr), (pfn), PAGE_SIZE, \
|
/linux-4.1.27/arch/mips/power/ |
D | cpu.c | 37 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 42 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
|
/linux-4.1.27/arch/arm/mach-s3c24xx/ |
D | mach-smdk2443.c | 53 .pfn = __phys_to_pfn(S3C2410_CS2), 58 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)), 63 .pfn = __phys_to_pfn(S3C2410_CS2), 68 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
|
D | mach-smdk2440.c | 53 .pfn = __phys_to_pfn(S3C2410_CS2), 58 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)), 63 .pfn = __phys_to_pfn(S3C2410_CS2), 68 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
|
D | mach-bast.c | 83 .pfn = PA_CS2(BAST_PA_ISAIO), 88 .pfn = PA_CS3(BAST_PA_ISAIO), 95 .pfn = __phys_to_pfn(BAST_PA_CTRL1), 100 .pfn = __phys_to_pfn(BAST_PA_CTRL2), 105 .pfn = __phys_to_pfn(BAST_PA_CTRL3), 110 .pfn = __phys_to_pfn(BAST_PA_CTRL4), 117 .pfn = __phys_to_pfn(BAST_PA_PC104_IRQREQ), 122 .pfn = __phys_to_pfn(BAST_PA_PC104_IRQRAW), 127 .pfn = __phys_to_pfn(BAST_PA_PC104_IRQMASK),
|
D | mach-vr1000.c | 72 .pfn = PA_CS2(BAST_PA_ISAIO), 77 .pfn = PA_CS3(BAST_PA_ISAIO), 85 .pfn = __phys_to_pfn(VR1000_PA_CTRL1), 90 .pfn = __phys_to_pfn(VR1000_PA_CTRL2), 95 .pfn = __phys_to_pfn(VR1000_PA_CTRL3), 100 .pfn = __phys_to_pfn(VR1000_PA_CTRL4),
|
D | mach-osiris.c | 65 .pfn = __phys_to_pfn(S3C2410_CS5), 70 .pfn = __phys_to_pfn(S3C2410_CS5), 79 .pfn = __phys_to_pfn(OSIRIS_PA_CTRL0), 84 .pfn = __phys_to_pfn(OSIRIS_PA_CTRL1), 89 .pfn = __phys_to_pfn(OSIRIS_PA_CTRL2), 94 .pfn = __phys_to_pfn(OSIRIS_PA_IDREG),
|
D | mach-smdk2416.c | 65 .pfn = __phys_to_pfn(S3C2410_CS2), 70 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)), 75 .pfn = __phys_to_pfn(S3C2410_CS2), 80 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
|
/linux-4.1.27/arch/arm/mach-ebsa110/ |
D | core.c | 78 .pfn = __phys_to_pfn(TRICK4_PHYS), 83 .pfn = __phys_to_pfn(TRICK3_PHYS), 88 .pfn = __phys_to_pfn(TRICK1_PHYS), 93 .pfn = __phys_to_pfn(TRICK0_PHYS), 103 .pfn = __phys_to_pfn(ISAIO_PHYS), 108 .pfn = __phys_to_pfn(ISAMEM_PHYS),
|
/linux-4.1.27/arch/score/mm/ |
D | cache.c | 82 unsigned long pfn, addr; in __update_cache() local 85 pfn = pte_pfn(pte); in __update_cache() 86 if (unlikely(!pfn_valid(pfn))) in __update_cache() 88 page = pfn_to_page(pfn); in __update_cache() 212 unsigned long addr, unsigned long pfn) in flush_cache_page() argument 215 unsigned long kaddr = 0xa0000000 | (pfn << PAGE_SHIFT); in flush_cache_page()
|
/linux-4.1.27/arch/openrisc/include/asm/ |
D | page.h | 83 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument 92 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
|
/linux-4.1.27/arch/tile/gxio/ |
D | kiorpc.c | 34 unsigned long pfn; in iorpc_ioremap() local 56 pfn = pte_pfn(mmio_base) + PFN_DOWN(offset); in iorpc_ioremap() 58 return ioremap_prot(PFN_PHYS(pfn), size, prot); in iorpc_ioremap()
|
/linux-4.1.27/drivers/infiniband/hw/mlx5/ |
D | mem.c | 56 u64 pfn; in mlx5_ib_cont_pages() local 80 pfn = sg_dma_address(sg) >> page_shift; in mlx5_ib_cont_pages() 83 tmp = (unsigned long)pfn; in mlx5_ib_cont_pages() 87 base = pfn; in mlx5_ib_cont_pages() 90 if (base + p != pfn) { in mlx5_ib_cont_pages() 95 base = pfn; in mlx5_ib_cont_pages()
|
/linux-4.1.27/arch/hexagon/include/asm/ |
D | page.h | 112 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument 144 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument
|
/linux-4.1.27/arch/arm64/include/asm/ |
D | memory.h | 87 #define __pfn_to_phys(pfn) ((phys_addr_t)(pfn) << PAGE_SHIFT) argument 156 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
|
/linux-4.1.27/arch/arm/kernel/ |
D | hibernate.c | 27 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument 32 return (pfn >= nosave_begin_pfn) && (pfn <= nosave_end_pfn); in pfn_is_nosave()
|
D | crash_dump.c | 33 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument 42 vaddr = ioremap(__pfn_to_phys(pfn), PAGE_SIZE); in copy_oldmem_page()
|
/linux-4.1.27/arch/arm/mach-ep93xx/ |
D | ts72xx.c | 35 .pfn = __phys_to_pfn(TS72XX_MODEL_PHYS_BASE), 40 .pfn = __phys_to_pfn(TS72XX_OPTIONS_PHYS_BASE), 45 .pfn = __phys_to_pfn(TS72XX_OPTIONS2_PHYS_BASE), 50 .pfn = __phys_to_pfn(TS72XX_RTC_INDEX_PHYS_BASE), 55 .pfn = __phys_to_pfn(TS72XX_RTC_DATA_PHYS_BASE),
|
/linux-4.1.27/drivers/char/ |
D | mem.c | 55 static inline int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) in valid_mmap_phys_addr_range() argument 62 static inline int range_is_allowed(unsigned long pfn, unsigned long size) in range_is_allowed() argument 64 u64 from = ((u64)pfn) << PAGE_SHIFT; in range_is_allowed() 69 if (!devmem_is_allowed(pfn)) { in range_is_allowed() 76 pfn++; in range_is_allowed() 81 static inline int range_is_allowed(unsigned long pfn, unsigned long size) in range_is_allowed() argument 224 unsigned long pfn, unsigned long size, pgprot_t *vma_prot) in phys_mem_access_prot_allowed() argument 265 static pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument 269 phys_addr_t offset = pfn << PAGE_SHIFT; in phys_mem_access_prot() 357 unsigned long pfn; in mmap_kmem() local [all …]
|
/linux-4.1.27/drivers/net/ethernet/ibm/ehea/ |
D | ehea_qmr.c | 628 static int ehea_update_busmap(unsigned long pfn, unsigned long nr_pages, int add) in ehea_update_busmap() argument 641 start_section = (pfn * PAGE_SIZE) / EHEA_SECTSIZE; in ehea_update_busmap() 671 int ehea_add_sect_bmap(unsigned long pfn, unsigned long nr_pages) in ehea_add_sect_bmap() argument 676 ret = ehea_update_busmap(pfn, nr_pages, EHEA_BUSMAP_ADD_SECT); in ehea_add_sect_bmap() 681 int ehea_rem_sect_bmap(unsigned long pfn, unsigned long nr_pages) in ehea_rem_sect_bmap() argument 686 ret = ehea_update_busmap(pfn, nr_pages, EHEA_BUSMAP_REM_SECT); in ehea_rem_sect_bmap() 691 static int ehea_is_hugepage(unsigned long pfn) in ehea_is_hugepage() argument 695 if (pfn & EHEA_HUGEPAGE_PFN_MASK) in ehea_is_hugepage() 698 page_order = compound_order(pfn_to_page(pfn)); in ehea_is_hugepage() 709 unsigned long pfn, start_pfn, end_pfn, nr_pages; in ehea_create_busmap_callback() local [all …]
|
/linux-4.1.27/arch/arm/mach-sa1100/ |
D | nanoengine.c | 70 .pfn = __phys_to_pfn(0x10000000), 76 .pfn = __phys_to_pfn(NANO_PCI_MEM_RW_PHYS), 82 .pfn = __phys_to_pfn(NANO_PCI_CONFIG_SPACE_PHYS),
|
/linux-4.1.27/drivers/misc/ |
D | vmw_balloon.c | 318 static int vmballoon_send_lock_page(struct vmballoon *b, unsigned long pfn, in vmballoon_send_lock_page() argument 324 pfn32 = (u32)pfn; in vmballoon_send_lock_page() 325 if (pfn32 != pfn) in vmballoon_send_lock_page() 330 *hv_status = status = VMWARE_BALLOON_CMD(LOCK, pfn, dummy); in vmballoon_send_lock_page() 334 pr_debug("%s - ppn %lx, hv returns %ld\n", __func__, pfn, status); in vmballoon_send_lock_page() 343 static bool vmballoon_send_unlock_page(struct vmballoon *b, unsigned long pfn) in vmballoon_send_unlock_page() argument 348 pfn32 = (u32)pfn; in vmballoon_send_unlock_page() 349 if (pfn32 != pfn) in vmballoon_send_unlock_page() 354 status = VMWARE_BALLOON_CMD(UNLOCK, pfn, dummy); in vmballoon_send_unlock_page() 358 pr_debug("%s - ppn %lx, hv returns %ld\n", __func__, pfn, status); in vmballoon_send_unlock_page()
|
/linux-4.1.27/arch/mips/include/asm/mach-ip27/ |
D | kernel-entry-init.h | 51 dsrl t1, 12 # 4K pfn 52 dsrl t2, 12 # 4K pfn 53 dsll t1, 6 # Get pfn into place 54 dsll t2, 6 # Get pfn into place
|
/linux-4.1.27/lib/ |
D | ioremap.c | 57 u64 pfn; in ioremap_pte_range() local 59 pfn = phys_addr >> PAGE_SHIFT; in ioremap_pte_range() 65 set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot)); in ioremap_pte_range() 66 pfn++; in ioremap_pte_range()
|
/linux-4.1.27/arch/mips/mm/ |
D | cache.c | 31 unsigned long pfn); 131 unsigned long pfn, addr; in __update_cache() local 134 pfn = pte_pfn(pte); in __update_cache() 135 if (unlikely(!pfn_valid(pfn))) in __update_cache() 137 page = pfn_to_page(pfn); in __update_cache()
|
D | ioremap.c | 23 unsigned long pfn; in remap_area_pte() local 32 pfn = phys_addr >> PAGE_SHIFT; in remap_area_pte() 38 set_pte(pte, pfn_pte(pfn, pgprot)); in remap_area_pte() 40 pfn++; in remap_area_pte()
|
D | init.c | 396 unsigned long pfn; in free_init_pages() local 398 for (pfn = PFN_UP(begin); pfn < PFN_DOWN(end); pfn++) { in free_init_pages() 399 struct page *page = pfn_to_page(pfn); in free_init_pages() 400 void *addr = phys_to_virt(PFN_PHYS(pfn)); in free_init_pages()
|
/linux-4.1.27/arch/arm/mach-socfpga/ |
D | socfpga.c | 37 .pfn = 0, /* run-time */ 44 .pfn = __phys_to_pfn(0xffc02000), 56 scu_io_desc.pfn = __phys_to_pfn(base); in socfpga_scu_map_io()
|
/linux-4.1.27/tools/testing/selftests/vm/ |
D | transhuge-stress.c | 110 int64_t pfn; in main() local 112 pfn = allocate_transhuge(p); in main() 114 if (pfn < 0) { in main() 117 size_t idx = pfn >> (HPAGE_SHIFT - PAGE_SHIFT); in main()
|
/linux-4.1.27/arch/xtensa/mm/ |
D | init.c | 238 unsigned long pfn; in bootmem_init() local 258 pfn = PAGE_ALIGN(sysmem.bank[i].start) >> PAGE_SHIFT; in bootmem_init() 259 if (pfn < min_low_pfn) in bootmem_init() 260 min_low_pfn = pfn; in bootmem_init() 261 pfn = PAGE_ALIGN(sysmem.bank[i].end - 1) >> PAGE_SHIFT; in bootmem_init() 262 if (pfn > max_pfn) in bootmem_init() 263 max_pfn = pfn; in bootmem_init()
|
D | cache.c | 200 unsigned long pfn) in local_flush_cache_page() argument 204 unsigned long phys = page_to_phys(pfn_to_page(pfn)); in local_flush_cache_page() 216 unsigned long pfn = pte_pfn(*ptep); in update_mmu_cache() local 219 if (!pfn_valid(pfn)) in update_mmu_cache() 222 page = pfn_to_page(pfn); in update_mmu_cache()
|
/linux-4.1.27/arch/arm/mach-lpc32xx/ |
D | common.c | 168 .pfn = __phys_to_pfn(LPC32XX_AHB0_START), 174 .pfn = __phys_to_pfn(LPC32XX_AHB1_START), 180 .pfn = __phys_to_pfn(LPC32XX_FABAPB_START), 186 .pfn = __phys_to_pfn(LPC32XX_IRAM_BASE),
|
/linux-4.1.27/Documentation/virtual/kvm/ |
D | locking.txt | 31 1): The mapping from gfn to pfn 32 The mapping from gfn to pfn may be changed since we can only ensure the pfn 63 to pin gfn to pfn, because after gfn_to_pfn_atomic(): 64 - We have held the refcount of pfn that means the pfn can not be freed and 66 - The pfn is writable that means it can not be shared between different gfns 106 kvm_set_pfn_accessed(spte.pfn); 108 kvm_set_pfn_dirty(spte.pfn);
|
/linux-4.1.27/drivers/gpu/drm/gma500/ |
D | gem.c | 179 unsigned long pfn; in psb_gem_fault() local 212 pfn = (dev_priv->stolen_base + r->offset) >> PAGE_SHIFT; in psb_gem_fault() 214 pfn = page_to_pfn(r->pages[page_offset]); in psb_gem_fault() 215 ret = vm_insert_pfn(vma, (unsigned long)vmf->virtual_address, pfn); in psb_gem_fault()
|
/linux-4.1.27/arch/nios2/mm/ |
D | ioremap.c | 27 unsigned long pfn; in remap_area_pte() local 37 pfn = PFN_DOWN(phys_addr); in remap_area_pte() 43 set_pte(pte, pfn_pte(pfn, pgprot)); in remap_area_pte() 45 pfn++; in remap_area_pte()
|
/linux-4.1.27/drivers/edac/ |
D | cell_edac.c | 38 unsigned long address, pfn, offset, syndrome; in cell_edac_count_ce() local 47 pfn = address >> PAGE_SHIFT; in cell_edac_count_ce() 53 csrow->first_page + pfn, offset, syndrome, in cell_edac_count_ce() 61 unsigned long address, pfn, offset; in cell_edac_count_ue() local 70 pfn = address >> PAGE_SHIFT; in cell_edac_count_ue() 75 csrow->first_page + pfn, offset, 0, in cell_edac_count_ue()
|
/linux-4.1.27/arch/ia64/kernel/ |
D | crash_dump.c | 34 copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument 41 vaddr = __va(pfn<<PAGE_SHIFT); in copy_oldmem_page()
|
/linux-4.1.27/drivers/scsi/csiostor/ |
D | csio_mb.c | 281 (FW_LDST_CMD_LC_F | FW_LDST_CMD_FN_V(hw->pfn)); in csio_mb_ldst() 449 FW_IQ_CMD_PFN_V(iq_params->pfn) | in csio_mb_iq_alloc() 505 FW_IQ_CMD_PFN_V(iq_params->pfn) | in csio_mb_iq_write() 628 FW_IQ_CMD_PFN_V(iq_params->pfn) | in csio_mb_iq_free() 663 FW_EQ_OFLD_CMD_PFN_V(eq_ofld_params->pfn) | in csio_mb_eq_ofld_alloc() 711 FW_EQ_OFLD_CMD_PFN_V(eq_ofld_params->pfn) | in csio_mb_eq_ofld_write() 814 FW_EQ_OFLD_CMD_PFN_V(eq_ofld_params->pfn) | in csio_mb_eq_ofld_free() 1158 uint32_t ctl_reg = PF_REG(hw->pfn, CIM_PF_MAILBOX_CTRL_A); in csio_mb_debug_cmd_handler() 1159 uint32_t data_reg = PF_REG(hw->pfn, CIM_PF_MAILBOX_DATA_A); in csio_mb_debug_cmd_handler() 1192 uint32_t ctl_reg = PF_REG(hw->pfn, CIM_PF_MAILBOX_CTRL_A); in csio_mb_issue() [all …]
|
/linux-4.1.27/virt/kvm/ |
D | kvm_main.c | 105 static void kvm_release_pfn_dirty(pfn_t pfn); 114 bool kvm_is_reserved_pfn(pfn_t pfn) in kvm_is_reserved_pfn() argument 116 if (pfn_valid(pfn)) in kvm_is_reserved_pfn() 117 return PageReserved(pfn_to_page(pfn)); in kvm_is_reserved_pfn() 1217 bool write_fault, bool *writable, pfn_t *pfn) in hva_to_pfn_fast() argument 1235 *pfn = page_to_pfn(page[0]); in hva_to_pfn_fast() 1250 bool *writable, pfn_t *pfn) in hva_to_pfn_slow() argument 1285 *pfn = page_to_pfn(page[0]); in hva_to_pfn_slow() 1318 pfn_t pfn = 0; in hva_to_pfn() local 1324 if (hva_to_pfn_fast(addr, atomic, async, write_fault, writable, &pfn)) in hva_to_pfn() [all …]
|
/linux-4.1.27/arch/arm/mach-mmp/ |
D | common.c | 30 .pfn = __phys_to_pfn(APB_PHYS_BASE), 35 .pfn = __phys_to_pfn(AXI_PHYS_BASE),
|
/linux-4.1.27/arch/microblaze/kernel/ |
D | dma.c | 165 unsigned long pfn; in dma_direct_mmap_coherent() local 172 pfn = consistent_virt_to_pfn(cpu_addr); in dma_direct_mmap_coherent() 174 pfn = virt_to_pfn(cpu_addr); in dma_direct_mmap_coherent() 176 return remap_pfn_range(vma, vma->vm_start, pfn + off, in dma_direct_mmap_coherent()
|
/linux-4.1.27/arch/arm/mach-ux500/ |
D | setup.h | 31 .pfn = __phys_to_pfn(x), \ 38 .pfn = __phys_to_pfn(x), \
|
/linux-4.1.27/arch/mips/kernel/ |
D | crash_dump.c | 26 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument 34 vaddr = kmap_atomic_pfn(pfn); in copy_oldmem_page()
|
/linux-4.1.27/tools/vm/ |
D | page-types.c | 292 unsigned long pfn; in pagemap_pfn() local 295 pfn = PM_PFRAME(val); in pagemap_pfn() 297 pfn = 0; in pagemap_pfn() 299 return pfn; in pagemap_pfn() 620 unsigned long pfn; in walk_vma() local 630 pfn = pagemap_pfn(buf[i]); in walk_vma() 631 if (pfn) in walk_vma() 632 walk_pfn(index + i, pfn, 1, buf[i]); in walk_vma() 852 unsigned long nr_pages, pfn, i; in walk_file() local 904 pfn = pagemap_pfn(buf[i]); in walk_file() [all …]
|
/linux-4.1.27/arch/mn10300/mm/ |
D | pgtable.c | 35 void set_pmd_pfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags) in set_pmd_pfn() argument 45 if (pfn & (PTRS_PER_PTE-1)) { /* pfn is misaligned */ in set_pmd_pfn() 56 set_pmd(pmd, pfn_pmd(pfn, flags)); in set_pmd_pfn()
|
/linux-4.1.27/drivers/gpu/drm/ttm/ |
D | ttm_bo_vm.c | 93 unsigned long pfn; in ttm_bo_vm_fault() local 216 pfn = ((bo->mem.bus.base + bo->mem.bus.offset) >> PAGE_SHIFT) + page_offset; in ttm_bo_vm_fault() 228 pfn = page_to_pfn(page); in ttm_bo_vm_fault() 232 ret = vm_insert_mixed(&cvma, address, pfn); in ttm_bo_vm_fault() 234 ret = vm_insert_pfn(&cvma, address, pfn); in ttm_bo_vm_fault()
|
/linux-4.1.27/arch/arm/mach-versatile/ |
D | core.c | 134 .pfn = __phys_to_pfn(VERSATILE_SYS_BASE), 139 .pfn = __phys_to_pfn(VERSATILE_SIC_BASE), 144 .pfn = __phys_to_pfn(VERSATILE_VIC_BASE), 149 .pfn = __phys_to_pfn(VERSATILE_SCTL_BASE), 156 .pfn = __phys_to_pfn(VERSATILE_IB2_BASE), 164 .pfn = __phys_to_pfn(VERSATILE_UART0_BASE), 172 .pfn = __phys_to_pfn(VERSATILE_PCI_CORE_BASE), 177 .pfn = __phys_to_pfn(VERSATILE_PCI_BASE), 182 .pfn = __phys_to_pfn(VERSATILE_PCI_CFG_BASE),
|
/linux-4.1.27/drivers/iommu/ |
D | iova.c | 291 struct iova *find_iova(struct iova_domain *iovad, unsigned long pfn) in find_iova() argument 303 if ((pfn >= iova->pfn_lo) && (pfn <= iova->pfn_hi)) { in find_iova() 314 if (pfn < iova->pfn_lo) in find_iova() 316 else if (pfn > iova->pfn_lo) in find_iova() 350 free_iova(struct iova_domain *iovad, unsigned long pfn) in free_iova() argument 352 struct iova *iova = find_iova(iovad, pfn); in free_iova()
|