Home
last modified time | relevance | path

Searched refs:pfn (Results 1 – 200 of 558) sorted by relevance

123

/linux-4.1.27/arch/x86/xen/
Dp2m.c110 static inline unsigned p2m_top_index(unsigned long pfn) in p2m_top_index() argument
112 BUG_ON(pfn >= MAX_P2M_PFN); in p2m_top_index()
113 return pfn / (P2M_MID_PER_PAGE * P2M_PER_PAGE); in p2m_top_index()
116 static inline unsigned p2m_mid_index(unsigned long pfn) in p2m_mid_index() argument
118 return (pfn / P2M_PER_PAGE) % P2M_MID_PER_PAGE; in p2m_mid_index()
121 static inline unsigned p2m_index(unsigned long pfn) in p2m_index() argument
123 return pfn % P2M_PER_PAGE; in p2m_index()
158 static void p2m_init_identity(unsigned long *p2m, unsigned long pfn) in p2m_init_identity() argument
163 p2m[i] = IDENTITY_FRAME(pfn + i); in p2m_init_identity()
196 unsigned long pfn, mfn; in xen_build_mfn_list_list() local
[all …]
Dsetup.c130 unsigned long __ref xen_chk_extra_mem(unsigned long pfn) in xen_chk_extra_mem() argument
133 phys_addr_t addr = PFN_PHYS(pfn); in xen_chk_extra_mem()
141 return IDENTITY_FRAME(pfn); in xen_chk_extra_mem()
149 unsigned long pfn, pfn_s, pfn_e; in xen_inv_extra_mem() local
157 for (pfn = pfn_s; pfn < pfn_e; pfn++) in xen_inv_extra_mem()
158 set_phys_to_machine(pfn, INVALID_P2M_ENTRY); in xen_inv_extra_mem()
226 unsigned long pfn, end; in xen_set_identity_and_release_chunk() local
233 for (pfn = start_pfn; pfn < end; pfn++) { in xen_set_identity_and_release_chunk()
234 unsigned long mfn = pfn_to_mfn(pfn); in xen_set_identity_and_release_chunk()
237 if (mfn == INVALID_P2M_ENTRY || mfn_to_pfn(mfn) != pfn) in xen_set_identity_and_release_chunk()
[all …]
Dmmu.c370 unsigned long pfn = mfn_to_pfn(mfn); in pte_mfn_to_pfn() local
373 if (unlikely(pfn == ~0)) in pte_mfn_to_pfn()
376 val = ((pteval_t)pfn << PAGE_SHIFT) | flags; in pte_mfn_to_pfn()
385 unsigned long pfn = (val & PTE_PFN_MASK) >> PAGE_SHIFT; in pte_pfn_to_mfn() local
390 mfn = __pfn_to_mfn(pfn); in pte_pfn_to_mfn()
392 mfn = pfn; in pte_pfn_to_mfn()
728 static void xen_do_pin(unsigned level, unsigned long pfn) in xen_do_pin() argument
733 op.arg1.mfn = pfn_to_mfn(pfn); in xen_do_pin()
752 unsigned long pfn = page_to_pfn(page); in xen_pin_page() local
783 pfn_pte(pfn, PAGE_KERNEL_RO), in xen_pin_page()
[all …]
Dmmu.h14 bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn);
16 void set_pte_mfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags);
Denlighten.c488 unsigned long pfn; in set_aliased_prot() local
495 pfn = pte_pfn(*ptep); in set_aliased_prot()
496 page = pfn_to_page(pfn); in set_aliased_prot()
498 pte = pfn_pte(pfn, prot); in set_aliased_prot()
530 void *av = __va(PFN_PHYS(pfn)); in set_aliased_prot()
606 unsigned long pfn, mfn; in xen_load_gdt() local
619 pfn = pte_pfn(*ptep); in xen_load_gdt()
620 mfn = pfn_to_mfn(pfn); in xen_load_gdt()
621 virt = __va(PFN_PHYS(pfn)); in xen_load_gdt()
654 unsigned long pfn, mfn; in xen_load_gdt_boot() local
[all …]
/linux-4.1.27/mm/
Dpage_isolation.c15 unsigned long flags, pfn; in set_migratetype_isolate() local
24 pfn = page_to_pfn(page); in set_migratetype_isolate()
25 arg.start_pfn = pfn; in set_migratetype_isolate()
132 __first_valid_page(unsigned long pfn, unsigned long nr_pages) in __first_valid_page() argument
136 if (pfn_valid_within(pfn + i)) in __first_valid_page()
140 return pfn_to_page(pfn + i); in __first_valid_page()
160 unsigned long pfn; in start_isolate_page_range() local
167 for (pfn = start_pfn; in start_isolate_page_range()
168 pfn < end_pfn; in start_isolate_page_range()
169 pfn += pageblock_nr_pages) { in start_isolate_page_range()
[all …]
Dpage_ext.c100 unsigned long pfn = page_to_pfn(page); in lookup_page_ext() local
115 offset = pfn - round_down(node_start_pfn(page_to_nid(page)), in lookup_page_ext()
177 unsigned long pfn = page_to_pfn(page); in lookup_page_ext() local
178 struct mem_section *section = __pfn_to_section(pfn); in lookup_page_ext()
189 return section->page_ext + pfn; in lookup_page_ext()
211 static int __meminit init_section_page_ext(unsigned long pfn, int nid) in init_section_page_ext() argument
217 section = __pfn_to_section(pfn); in init_section_page_ext()
241 pfn &= PAGE_SECTION_MASK; in init_section_page_ext()
242 section->page_ext = base - pfn; in init_section_page_ext()
262 static void __free_page_ext(unsigned long pfn) in __free_page_ext() argument
[all …]
Dpage_owner.c80 print_page_owner(char __user *buf, size_t count, unsigned long pfn, in print_page_owner() argument
103 pageblock_mt = get_pfnblock_migratetype(page, pfn); in print_page_owner()
107 pfn, in print_page_owner()
108 pfn >> pageblock_order, in print_page_owner()
149 unsigned long pfn; in read_page_owner() local
157 pfn = min_low_pfn + *ppos; in read_page_owner()
160 while (!pfn_valid(pfn) && (pfn & (MAX_ORDER_NR_PAGES - 1)) != 0) in read_page_owner()
161 pfn++; in read_page_owner()
166 for (; pfn < max_pfn; pfn++) { in read_page_owner()
171 if ((pfn & (MAX_ORDER_NR_PAGES - 1)) == 0 && !pfn_valid(pfn)) { in read_page_owner()
[all …]
Dmemory-failure.c191 unsigned long pfn, struct page *page, int flags) in kill_proc() argument
198 pfn, t->comm, t->pid); in kill_proc()
335 int fail, struct page *page, unsigned long pfn, in kill_procs() argument
350 pfn, tk->tsk->comm, tk->tsk->pid); in kill_procs()
361 pfn, page, flags) < 0) in kill_procs()
364 pfn, tk->tsk->comm, tk->tsk->pid); in kill_procs()
599 static int me_kernel(struct page *p, unsigned long pfn) in me_kernel() argument
607 static int me_unknown(struct page *p, unsigned long pfn) in me_unknown() argument
609 printk(KERN_ERR "MCE %#lx: Unknown page state\n", pfn); in me_unknown()
616 static int me_pagecache_clean(struct page *p, unsigned long pfn) in me_pagecache_clean() argument
[all …]
Dmemory_hotplug.c247 unsigned long i, pfn, end_pfn, nr_pages; in register_page_bootmem_info_node() local
271 pfn = pgdat->node_start_pfn; in register_page_bootmem_info_node()
275 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node()
282 if (pfn_valid(pfn) && (pfn_to_nid(pfn) == node)) in register_page_bootmem_info_node()
283 register_page_bootmem_info_section(pfn); in register_page_bootmem_info_node()
330 unsigned long pfn; in fix_zone_id() local
332 for (pfn = start_pfn; pfn < end_pfn; pfn++) in fix_zone_id()
333 set_page_links(pfn_to_page(pfn), zid, nid, pfn); in fix_zone_id()
553 unsigned long pfn; in find_biggest_section_pfn() local
556 pfn = end_pfn - 1; in find_biggest_section_pfn()
[all …]
Dcma.c84 static void cma_clear_bitmap(struct cma *cma, unsigned long pfn, in cma_clear_bitmap() argument
89 bitmap_no = (pfn - cma->base_pfn) >> cma->order_per_bit; in cma_clear_bitmap()
100 unsigned long base_pfn = cma->base_pfn, pfn = base_pfn; in cma_activate_area() local
109 WARN_ON_ONCE(!pfn_valid(pfn)); in cma_activate_area()
110 zone = page_zone(pfn_to_page(pfn)); in cma_activate_area()
115 base_pfn = pfn; in cma_activate_area()
116 for (j = pageblock_nr_pages; j; --j, pfn++) { in cma_activate_area()
117 WARN_ON_ONCE(!pfn_valid(pfn)); in cma_activate_area()
124 if (page_zone(pfn_to_page(pfn)) != zone) in cma_activate_area()
364 unsigned long mask, offset, pfn, start = 0; in cma_alloc() local
[all …]
Dcompaction.c59 unsigned long pfn = page_to_pfn(page); in release_freepages() local
62 if (pfn > high_pfn) in release_freepages()
63 high_pfn = pfn; in release_freepages()
219 unsigned long pfn; in __reset_isolation_suitable() local
227 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in __reset_isolation_suitable()
232 if (!pfn_valid(pfn)) in __reset_isolation_suitable()
235 page = pfn_to_page(pfn); in __reset_isolation_suitable()
267 unsigned long pfn; in update_pageblock_skip() local
280 pfn = page_to_pfn(page); in update_pageblock_skip()
284 if (pfn > zone->compact_cached_migrate_pfn[0]) in update_pageblock_skip()
[all …]
Dhwpoison-inject.c15 unsigned long pfn = val; in hwpoison_inject() local
23 if (!pfn_valid(pfn)) in hwpoison_inject()
26 p = pfn_to_page(pfn); in hwpoison_inject()
58 pr_info("Injecting memory failure at pfn %#lx\n", pfn); in hwpoison_inject()
59 return memory_failure(pfn, 18, MF_COUNT_INCREASED); in hwpoison_inject()
Dpage_alloc.c253 unsigned long pfn = page_to_pfn(page); in page_outside_zone_boundaries() local
260 if (!zone_spans_pfn(zone, pfn)) in page_outside_zone_boundaries()
266 pfn, zone_to_nid(zone), zone->name, in page_outside_zone_boundaries()
574 unsigned long pfn, in __free_one_page() argument
593 page_idx = pfn & ((1 << MAX_ORDER) - 1); in __free_one_page()
768 struct page *page, unsigned long pfn, in free_one_page() argument
780 migratetype = get_pfnblock_migratetype(page, pfn); in free_one_page()
782 __free_one_page(page, pfn, zone, order, migratetype); in free_one_page()
842 unsigned long pfn = page_to_pfn(page); in __free_pages_ok() local
847 migratetype = get_pfnblock_migratetype(page, pfn); in __free_pages_ok()
[all …]
Dvmstat.c980 unsigned long pfn; in pagetypeinfo_showblockcount_print() local
985 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in pagetypeinfo_showblockcount_print()
988 if (!pfn_valid(pfn)) in pagetypeinfo_showblockcount_print()
991 page = pfn_to_page(pfn); in pagetypeinfo_showblockcount_print()
994 if (!memmap_valid_within(pfn, page, zone)) in pagetypeinfo_showblockcount_print()
1032 unsigned long pfn = zone->zone_start_pfn, block_end_pfn; in pagetypeinfo_showmixedcount_print() local
1033 unsigned long end_pfn = pfn + zone->spanned_pages; in pagetypeinfo_showmixedcount_print()
1039 pfn = zone->zone_start_pfn; in pagetypeinfo_showmixedcount_print()
1046 for (; pfn < end_pfn; ) { in pagetypeinfo_showmixedcount_print()
1047 if (!pfn_valid(pfn)) { in pagetypeinfo_showmixedcount_print()
[all …]
Dinternal.h158 extern void __free_pages_bootmem(struct page *page, unsigned long pfn,
323 unsigned long pfn = page_to_pfn(base) + offset; in mem_map_next() local
324 if (!pfn_valid(pfn)) in mem_map_next()
326 return pfn_to_page(pfn); in mem_map_next()
366 enum zone_type zone, unsigned long nid, unsigned long pfn);
381 enum zone_type zone, unsigned long nid, unsigned long pfn) in mminit_verify_page_links() argument
Dsparse.c172 unsigned long pfn; in memory_present() local
176 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { in memory_present()
177 unsigned long section = pfn_to_section_nr(pfn); in memory_present()
197 unsigned long pfn; in node_memmap_size_bytes() local
201 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in node_memmap_size_bytes()
202 if (nid != early_pfn_to_nid(pfn)) in node_memmap_size_bytes()
205 if (pfn_present(pfn)) in node_memmap_size_bytes()
/linux-4.1.27/arch/x86/include/asm/xen/
Dpage.h48 extern unsigned long get_phys_to_machine(unsigned long pfn);
49 extern bool set_phys_to_machine(unsigned long pfn, unsigned long mfn);
50 extern bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn);
85 static inline unsigned long __pfn_to_mfn(unsigned long pfn) in __pfn_to_mfn() argument
89 if (pfn < xen_p2m_size) in __pfn_to_mfn()
90 mfn = xen_p2m_addr[pfn]; in __pfn_to_mfn()
91 else if (unlikely(pfn < xen_max_p2m_pfn)) in __pfn_to_mfn()
92 return get_phys_to_machine(pfn); in __pfn_to_mfn()
94 return IDENTITY_FRAME(pfn); in __pfn_to_mfn()
97 return get_phys_to_machine(pfn); in __pfn_to_mfn()
[all …]
Dinterface_32.h99 #define xen_pfn_to_cr3(pfn) (((unsigned)(pfn) << 12) | ((unsigned)(pfn) >> 20)) argument
/linux-4.1.27/arch/arm/xen/
Dp2m.c21 unsigned long pfn; member
42 if (new->pfn == entry->pfn) in xen_add_phys_to_mach_entry()
45 if (new->pfn < entry->pfn) in xen_add_phys_to_mach_entry()
57 __func__, &new->pfn, &new->mfn, &entry->pfn, &entry->mfn); in xen_add_phys_to_mach_entry()
62 unsigned long __pfn_to_mfn(unsigned long pfn) in __pfn_to_mfn() argument
71 if (entry->pfn <= pfn && in __pfn_to_mfn()
72 entry->pfn + entry->nr_pages > pfn) { in __pfn_to_mfn()
74 return entry->mfn + (pfn - entry->pfn); in __pfn_to_mfn()
76 if (pfn < entry->pfn) in __pfn_to_mfn()
119 bool __set_phys_to_machine_multi(unsigned long pfn, in __set_phys_to_machine_multi() argument
[all …]
Dmm.c51 unsigned long pfn; in dma_cache_maint() local
54 pfn = (handle >> PAGE_SHIFT) + offset / PAGE_SIZE; in dma_cache_maint()
66 cflush.a.dev_bus_addr = pfn << PAGE_SHIFT; in dma_cache_maint()
82 pfn++; in dma_cache_maint()
141 unsigned long pfn, in xen_arch_need_swiotlb() argument
144 return (!hypercall_cflush && (pfn != mfn) && !is_device_dma_coherent(dev)); in xen_arch_need_swiotlb()
/linux-4.1.27/include/trace/events/
Dcma.h12 TP_PROTO(unsigned long pfn, const struct page *page,
15 TP_ARGS(pfn, page, count, align),
18 __field(unsigned long, pfn)
25 __entry->pfn = pfn;
32 __entry->pfn,
40 TP_PROTO(unsigned long pfn, const struct page *page,
43 TP_ARGS(pfn, page, count),
46 __field(unsigned long, pfn)
52 __entry->pfn = pfn;
58 __entry->pfn,
Dkmem.h180 __field( unsigned long, pfn )
185 __entry->pfn = page_to_pfn(page);
190 pfn_to_page(__entry->pfn),
191 __entry->pfn,
202 __field( unsigned long, pfn )
207 __entry->pfn = page_to_pfn(page);
212 pfn_to_page(__entry->pfn),
213 __entry->pfn,
225 __field( unsigned long, pfn )
232 __entry->pfn = page ? page_to_pfn(page) : -1UL;
[all …]
Dpagemap.h38 __field(unsigned long, pfn )
45 __entry->pfn = page_to_pfn(page);
53 __entry->pfn,
71 __field(unsigned long, pfn )
76 __entry->pfn = page_to_pfn(page);
80 TP_printk("page=%p pfn=%lu", __entry->page, __entry->pfn)
Dfilemap.h21 __field(unsigned long, pfn)
28 __entry->pfn = page_to_pfn(page);
40 pfn_to_page(__entry->pfn),
41 __entry->pfn,
Dxen.h328 TP_PROTO(struct mm_struct *mm, unsigned long pfn, unsigned level, bool pinned),
329 TP_ARGS(mm, pfn, level, pinned),
332 __field(unsigned long, pfn)
337 __entry->pfn = pfn;
341 __entry->mm, __entry->pfn, __entry->level,
346 TP_PROTO(unsigned long pfn, unsigned level, bool pinned),
347 TP_ARGS(pfn, level, pinned),
349 __field(unsigned long, pfn)
353 TP_fast_assign(__entry->pfn = pfn;
357 __entry->pfn, __entry->level,
/linux-4.1.27/include/asm-generic/
Dmemory_model.h15 #define arch_pfn_to_nid(pfn) pfn_to_nid(pfn) argument
19 #define arch_local_page_offset(pfn, nid) \ argument
20 ((pfn) - NODE_DATA(nid)->node_start_pfn)
30 #define __pfn_to_page(pfn) (mem_map + ((pfn) - ARCH_PFN_OFFSET)) argument
35 #define __pfn_to_page(pfn) \ argument
36 ({ unsigned long __pfn = (pfn); \
51 #define __pfn_to_page(pfn) (vmemmap + (pfn)) argument
65 #define __pfn_to_page(pfn) \ argument
66 ({ unsigned long __pfn = (pfn); \
Dpage.h84 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument
93 #define pfn_valid(pfn) ((pfn) >= ARCH_PFN_OFFSET && ((pfn) - ARCH_PFN_OFFSET) < max_mapnr) argument
Dpgtable.h493 unsigned long pfn, unsigned long addr, in track_pfn_remap() argument
504 unsigned long pfn) in track_pfn_insert() argument
524 unsigned long pfn, unsigned long size) in untrack_pfn() argument
529 unsigned long pfn, unsigned long addr,
532 unsigned long pfn);
534 extern void untrack_pfn(struct vm_area_struct *vma, unsigned long pfn,
539 static inline int is_zero_pfn(unsigned long pfn) in is_zero_pfn() argument
542 unsigned long offset_from_zero_pfn = pfn - zero_pfn; in is_zero_pfn()
549 static inline int is_zero_pfn(unsigned long pfn) in is_zero_pfn() argument
552 return pfn == zero_pfn; in is_zero_pfn()
/linux-4.1.27/arch/m32r/include/asm/
Dmmzone.h16 #define node_localnr(pfn, nid) ((pfn) - NODE_DATA(nid)->node_start_pfn) argument
27 #define pfn_valid(pfn) (1) argument
29 #define pfn_valid(pfn) ((pfn) < num_physpages) argument
36 static __inline__ int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument
41 if (pfn >= node_start_pfn(node) && pfn < node_end_pfn(node)) in pfn_to_nid()
47 static __inline__ struct pglist_data *pfn_to_pgdat(unsigned long pfn) in pfn_to_pgdat() argument
49 return(NODE_DATA(pfn_to_nid(pfn))); in pfn_to_pgdat()
Dpgtable-2level.h71 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument
72 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument
Dcacheflush.h14 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) argument
36 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) argument
50 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) argument
/linux-4.1.27/arch/parisc/include/asm/
Dmmzone.h33 #define pfn_is_io(pfn) ((pfn & (0xf0000000UL >> PAGE_SHIFT)) == (0xf0000000UL >> PAGE_SHIFT)) argument
36 #define pfn_is_io(pfn) ((pfn & (0xf000000000000000UL >> PAGE_SHIFT)) == (0xf000000000000000UL >> PA… argument
39 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument
43 if (unlikely(pfn_is_io(pfn))) in pfn_to_nid()
46 i = pfn >> PFNNID_SHIFT; in pfn_to_nid()
52 static inline int pfn_valid(int pfn) in pfn_valid() argument
54 int nid = pfn_to_nid(pfn); in pfn_valid()
57 return (pfn < node_end_pfn(nid)); in pfn_valid()
/linux-4.1.27/arch/arm/mach-gemini/
Dmm.c23 .pfn =__phys_to_pfn(GEMINI_GLOBAL_BASE),
28 .pfn = __phys_to_pfn(GEMINI_UART_BASE),
33 .pfn = __phys_to_pfn(GEMINI_TIMER_BASE),
38 .pfn = __phys_to_pfn(GEMINI_INTERRUPT_BASE),
43 .pfn = __phys_to_pfn(GEMINI_POWER_CTRL_BASE),
48 .pfn = __phys_to_pfn(GEMINI_GPIO_BASE(0)),
53 .pfn = __phys_to_pfn(GEMINI_GPIO_BASE(1)),
58 .pfn = __phys_to_pfn(GEMINI_GPIO_BASE(2)),
63 .pfn = __phys_to_pfn(GEMINI_FLASH_CTRL_BASE),
68 .pfn = __phys_to_pfn(GEMINI_DRAM_CTRL_BASE),
[all …]
/linux-4.1.27/arch/s390/kernel/
Dsuspend.c99 void page_key_read(unsigned long *pfn) in page_key_read() argument
103 addr = (unsigned long) page_address(pfn_to_page(*pfn)); in page_key_read()
104 *(unsigned char *) pfn = (unsigned char) page_get_storage_key(addr); in page_key_read()
111 void page_key_memorize(unsigned long *pfn) in page_key_memorize() argument
113 page_key_wp->data[page_key_wx] = *(unsigned char *) pfn; in page_key_memorize()
114 *(unsigned char *) pfn = 0; in page_key_memorize()
137 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
145 if (pfn <= LC_PAGES) in pfn_is_nosave()
147 if (pfn >= nosave_begin_pfn && pfn < nosave_end_pfn) in pfn_is_nosave()
150 if (pfn >= stext_pfn && pfn <= eshared_pfn) in pfn_is_nosave()
[all …]
Dcrash_dump.c166 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, size_t csize, in copy_oldmem_page() argument
173 src = (pfn << PAGE_SHIFT) + offset; in copy_oldmem_page()
187 unsigned long from, unsigned long pfn, in remap_oldmem_pfn_range_kdump() argument
193 if (pfn < OLDMEM_SIZE >> PAGE_SHIFT) { in remap_oldmem_pfn_range_kdump()
194 size_old = min(size, OLDMEM_SIZE - (pfn << PAGE_SHIFT)); in remap_oldmem_pfn_range_kdump()
196 pfn + (OLDMEM_BASE >> PAGE_SHIFT), in remap_oldmem_pfn_range_kdump()
202 pfn += size_old >> PAGE_SHIFT; in remap_oldmem_pfn_range_kdump()
204 return remap_pfn_range(vma, from, pfn, size, prot); in remap_oldmem_pfn_range_kdump()
215 unsigned long pfn, in remap_oldmem_pfn_range_zfcpdump() argument
221 if (pfn < hsa_end >> PAGE_SHIFT) { in remap_oldmem_pfn_range_zfcpdump()
[all …]
/linux-4.1.27/arch/x86/power/
Dhibernate_32.c83 unsigned long pfn; in resume_physical_mapping_init() local
91 pfn = 0; in resume_physical_mapping_init()
98 if (pfn >= max_low_pfn) in resume_physical_mapping_init()
102 if (pfn >= max_low_pfn) in resume_physical_mapping_init()
110 set_pmd(pmd, pfn_pmd(pfn, PAGE_KERNEL_LARGE_EXEC)); in resume_physical_mapping_init()
111 pfn += PTRS_PER_PTE; in resume_physical_mapping_init()
120 for (; pte < max_pte; pte++, pfn++) { in resume_physical_mapping_init()
121 if (pfn >= max_low_pfn) in resume_physical_mapping_init()
124 set_pte(pte, pfn_pte(pfn, PAGE_KERNEL_EXEC)); in resume_physical_mapping_init()
167 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
[all …]
Dhibernate_64.c103 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
107 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
/linux-4.1.27/arch/alpha/include/asm/
Dmmzone.h34 #define node_localnr(pfn, nid) ((pfn) - NODE_DATA(nid)->node_start_pfn) argument
81 unsigned long pfn; \
83 pfn = page_to_pfn(page) << 32; \
84 pte_val(pte) = pfn | pgprot_val(pgprot); \
103 #define pfn_to_nid(pfn) pa_to_nid(((u64)(pfn) << PAGE_SHIFT)) argument
104 #define pfn_valid(pfn) \ argument
105 (((pfn) - node_start_pfn(pfn_to_nid(pfn))) < \
106 node_spanned_pages(pfn_to_nid(pfn))) \
Dpgtable.h194 #define PHYS_TWIDDLE(pfn) \ argument
195 ((((pfn) & KSEG_PFN) == (0x40000000000UL >> PAGE_SHIFT)) \
196 ? ((pfn) ^= KSEG_PFN) : (pfn))
198 #define PHYS_TWIDDLE(pfn) (pfn) argument
/linux-4.1.27/arch/unicore32/mm/
Dioremap.c102 remap_area_sections(unsigned long virt, unsigned long pfn, in remap_area_sections() argument
118 set_pmd(pmd, __pmd(__pfn_to_phys(pfn) | type->prot_sect)); in remap_area_sections()
119 pfn += SZ_4M >> PAGE_SHIFT; in remap_area_sections()
129 void __iomem *__uc32_ioremap_pfn_caller(unsigned long pfn, in __uc32_ioremap_pfn_caller() argument
140 if (pfn >= 0x100000 && (__pfn_to_phys(pfn) & ~SECTION_MASK)) in __uc32_ioremap_pfn_caller()
146 if (pfn_valid(pfn)) { in __uc32_ioremap_pfn_caller()
168 if (!((__pfn_to_phys(pfn) | size | addr) & ~PMD_MASK)) { in __uc32_ioremap_pfn_caller()
170 err = remap_area_sections(addr, pfn, size, type); in __uc32_ioremap_pfn_caller()
172 err = ioremap_page_range(addr, addr + size, __pfn_to_phys(pfn), in __uc32_ioremap_pfn_caller()
189 unsigned long pfn = __phys_to_pfn(phys_addr); in __uc32_ioremap_caller() local
[all …]
Dmmu.c81 unsigned long pfn; member
164 unsigned long end, unsigned long pfn, in alloc_init_pte() argument
169 set_pte(pte, pfn_pte(pfn, __pgprot(type->prot_pte))); in alloc_init_pte()
170 pfn++; in alloc_init_pte()
217 __pfn_to_phys((u64)md->pfn), md->virtual); in create_mapping()
225 __pfn_to_phys((u64)md->pfn), md->virtual); in create_mapping()
231 phys = (unsigned long)__pfn_to_phys(md->pfn); in create_mapping()
237 __pfn_to_phys(md->pfn), addr); in create_mapping()
368 map.pfn = __phys_to_pfn(virt_to_phys(vectors)); in devicemaps_init()
378 map.pfn = __phys_to_pfn(virt_to_phys(vectors)); in devicemaps_init()
[all …]
/linux-4.1.27/arch/unicore32/kernel/
Dhibernate.c75 unsigned long pfn; in resume_physical_mapping_init() local
83 pfn = 0; in resume_physical_mapping_init()
90 if (pfn >= max_low_pfn) in resume_physical_mapping_init()
96 if (pfn >= max_low_pfn) in resume_physical_mapping_init()
107 for (; pte < max_pte; pte++, pfn++) { in resume_physical_mapping_init()
108 if (pfn >= max_low_pfn) in resume_physical_mapping_init()
111 set_pte(pte, pfn_pte(pfn, PAGE_KERNEL_EXEC)); in resume_physical_mapping_init()
145 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
150 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
/linux-4.1.27/arch/x86/include/asm/
Dmmzone_32.h33 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument
36 return((int) physnode_map[(pfn) / PAGES_PER_SECTION]); in pfn_to_nid()
42 static inline int pfn_valid(int pfn) in pfn_valid() argument
44 int nid = pfn_to_nid(pfn); in pfn_valid()
47 return (pfn < node_end_pfn(nid)); in pfn_valid()
51 #define early_pfn_valid(pfn) pfn_valid((pfn)) argument
Dpgalloc.h15 static inline void paravirt_alloc_pte(struct mm_struct *mm, unsigned long pfn) {} in paravirt_alloc_pte() argument
16 static inline void paravirt_alloc_pmd(struct mm_struct *mm, unsigned long pfn) {} in paravirt_alloc_pmd() argument
17 static inline void paravirt_alloc_pmd_clone(unsigned long pfn, unsigned long clonepfn, in paravirt_alloc_pmd_clone() argument
19 static inline void paravirt_alloc_pud(struct mm_struct *mm, unsigned long pfn) {} in paravirt_alloc_pud() argument
20 static inline void paravirt_release_pte(unsigned long pfn) {} in paravirt_release_pte() argument
21 static inline void paravirt_release_pmd(unsigned long pfn) {} in paravirt_release_pmd() argument
22 static inline void paravirt_release_pud(unsigned long pfn) {} in paravirt_release_pud() argument
72 unsigned long pfn = page_to_pfn(pte); in pmd_populate() local
74 paravirt_alloc_pte(mm, pfn); in pmd_populate()
75 set_pmd(pmd, __pmd(((pteval_t)pfn << PAGE_SHIFT) | _PAGE_TABLE)); in pmd_populate()
Dpage_64.h34 #define pfn_valid(pfn) ((pfn) < max_pfn) argument
Dpage_32.h18 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
Dhighmem.h69 void *kmap_atomic_pfn(unsigned long pfn);
70 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot);
/linux-4.1.27/include/linux/
Dmmzone.h85 static inline int get_pfnblock_migratetype(struct page *page, unsigned long pfn) in get_pfnblock_migratetype() argument
88 return get_pfnblock_flags_mask(page, pfn, PB_migrate_end, in get_pfnblock_migratetype()
553 static inline bool zone_spans_pfn(const struct zone *zone, unsigned long pfn) in zone_spans_pfn() argument
555 return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone); in zone_spans_pfn()
1041 static inline unsigned long early_pfn_to_nid(unsigned long pfn) in early_pfn_to_nid() argument
1048 #define pfn_to_nid(pfn) (0) argument
1074 #define pfn_to_section_nr(pfn) ((pfn) >> PFN_SECTION_SHIFT) argument
1077 #define SECTION_ALIGN_UP(pfn) (((pfn) + PAGES_PER_SECTION - 1) & PAGE_SECTION_MASK) argument
1078 #define SECTION_ALIGN_DOWN(pfn) ((pfn) & PAGE_SECTION_MASK) argument
1176 static inline struct mem_section *__pfn_to_section(unsigned long pfn) in __pfn_to_section() argument
[all …]
Dio-mapping.h88 unsigned long pfn; in io_mapping_map_atomic_wc() local
92 pfn = (unsigned long) (phys_addr >> PAGE_SHIFT); in io_mapping_map_atomic_wc()
93 return iomap_atomic_prot_pfn(pfn, mapping->prot); in io_mapping_map_atomic_wc()
Dsuspend.h477 void page_key_read(unsigned long *pfn);
478 void page_key_memorize(unsigned long *pfn);
494 static inline void page_key_read(unsigned long *pfn) {} in page_key_read() argument
495 static inline void page_key_memorize(unsigned long *pfn) {} in page_key_memorize() argument
Dcrash_dump.h22 unsigned long from, unsigned long pfn,
83 extern int register_oldmem_pfn_is_ram(int (*fn)(unsigned long pfn));
/linux-4.1.27/arch/arm/include/asm/xen/
Dpage.h14 #define phys_to_machine_mapping_valid(pfn) (1) argument
35 unsigned long __pfn_to_mfn(unsigned long pfn);
38 static inline unsigned long pfn_to_mfn(unsigned long pfn) in pfn_to_mfn() argument
43 mfn = __pfn_to_mfn(pfn); in pfn_to_mfn()
48 return pfn; in pfn_to_mfn()
98 bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn);
99 bool __set_phys_to_machine_multi(unsigned long pfn, unsigned long mfn,
102 static inline bool set_phys_to_machine(unsigned long pfn, unsigned long mfn) in set_phys_to_machine() argument
104 return __set_phys_to_machine(pfn, mfn); in set_phys_to_machine()
111 unsigned long pfn,
Dpage-coherent.h53 unsigned long pfn = PFN_DOWN(handle); in xen_dma_unmap_page() local
58 if (pfn_valid(pfn)) { in xen_dma_unmap_page()
68 unsigned long pfn = PFN_DOWN(handle); in xen_dma_sync_single_for_cpu() local
69 if (pfn_valid(pfn)) { in xen_dma_sync_single_for_cpu()
79 unsigned long pfn = PFN_DOWN(handle); in xen_dma_sync_single_for_device() local
80 if (pfn_valid(pfn)) { in xen_dma_sync_single_for_device()
/linux-4.1.27/kernel/power/
Dsnapshot.c634 static int memory_bm_find_bit(struct memory_bitmap *bm, unsigned long pfn, in memory_bm_find_bit() argument
643 if (pfn >= zone->start_pfn && pfn < zone->end_pfn) in memory_bm_find_bit()
650 if (pfn >= curr->start_pfn && pfn < curr->end_pfn) { in memory_bm_find_bit()
666 if (((pfn - zone->start_pfn) & ~BM_BLOCK_MASK) == bm->cur.node_pfn) in memory_bm_find_bit()
670 block_nr = (pfn - zone->start_pfn) >> BM_BLOCK_SHIFT; in memory_bm_find_bit()
685 bm->cur.node_pfn = (pfn - zone->start_pfn) & ~BM_BLOCK_MASK; in memory_bm_find_bit()
689 *bit_nr = (pfn - zone->start_pfn) & BM_BLOCK_MASK; in memory_bm_find_bit()
694 static void memory_bm_set_bit(struct memory_bitmap *bm, unsigned long pfn) in memory_bm_set_bit() argument
700 error = memory_bm_find_bit(bm, pfn, &addr, &bit); in memory_bm_set_bit()
705 static int mem_bm_set_bit_check(struct memory_bitmap *bm, unsigned long pfn) in mem_bm_set_bit_check() argument
[all …]
/linux-4.1.27/arch/arm/mach-omap2/
Dio.c75 .pfn = __phys_to_pfn(L3_24XX_PHYS),
81 .pfn = __phys_to_pfn(L4_24XX_PHYS),
91 .pfn = __phys_to_pfn(DSP_MEM_2420_PHYS),
97 .pfn = __phys_to_pfn(DSP_IPI_2420_PHYS),
103 .pfn = __phys_to_pfn(DSP_MMU_2420_PHYS),
115 .pfn = __phys_to_pfn(L4_WK_243X_PHYS),
121 .pfn = __phys_to_pfn(OMAP243X_GPMC_PHYS),
127 .pfn = __phys_to_pfn(OMAP243X_SDRC_PHYS),
133 .pfn = __phys_to_pfn(OMAP243X_SMS_PHYS),
145 .pfn = __phys_to_pfn(L3_34XX_PHYS),
[all …]
/linux-4.1.27/arch/arm/mm/
Dfault-armv.c41 unsigned long pfn, pte_t *ptep) in do_adjust_pte() argument
56 flush_cache_page(vma, address, pfn); in do_adjust_pte()
57 outer_flush_range((pfn << PAGE_SHIFT), in do_adjust_pte()
58 (pfn << PAGE_SHIFT) + PAGE_SIZE); in do_adjust_pte()
93 unsigned long pfn) in adjust_pte() argument
123 ret = do_adjust_pte(vma, address, pfn, pte); in adjust_pte()
133 unsigned long addr, pte_t *ptep, unsigned long pfn) in make_coherent() argument
160 aliases += adjust_pte(mpnt, mpnt->vm_start + offset, pfn); in make_coherent()
164 do_adjust_pte(vma, addr, pfn, ptep); in make_coherent()
183 unsigned long pfn = pte_pfn(*ptep); in update_mmu_cache() local
[all …]
Dflush.c26 static void flush_pfn_alias(unsigned long pfn, unsigned long vaddr) in flush_pfn_alias() argument
31 set_top_pte(to, pfn_pte(pfn, PAGE_KERNEL)); in flush_pfn_alias()
40 static void flush_icache_alias(unsigned long pfn, unsigned long vaddr, unsigned long len) in flush_icache_alias() argument
46 set_top_pte(va, pfn_pte(pfn, PAGE_KERNEL)); in flush_icache_alias()
86 void flush_cache_page(struct vm_area_struct *vma, unsigned long user_addr, unsigned long pfn) in flush_cache_page() argument
89 vivt_flush_cache_page(vma, user_addr, pfn); in flush_cache_page()
94 flush_pfn_alias(pfn, user_addr); in flush_cache_page()
103 #define flush_pfn_alias(pfn,vaddr) do { } while (0) argument
104 #define flush_icache_alias(pfn,vaddr,len) do { } while (0) argument
260 unsigned long pfn; in __sync_icache_dcache() local
[all …]
Dioremap.c186 remap_area_sections(unsigned long virt, unsigned long pfn, in remap_area_sections() argument
204 pmd[0] = __pmd(__pfn_to_phys(pfn) | type->prot_sect); in remap_area_sections()
205 pfn += SZ_1M >> PAGE_SHIFT; in remap_area_sections()
206 pmd[1] = __pmd(__pfn_to_phys(pfn) | type->prot_sect); in remap_area_sections()
207 pfn += SZ_1M >> PAGE_SHIFT; in remap_area_sections()
218 remap_area_supersections(unsigned long virt, unsigned long pfn, in remap_area_supersections() argument
238 super_pmd_val = __pfn_to_phys(pfn) | type->prot_sect | in remap_area_supersections()
240 super_pmd_val |= ((pfn >> (32 - PAGE_SHIFT)) & 0xf) << 20; in remap_area_supersections()
251 pfn += SUPERSECTION_SIZE >> PAGE_SHIFT; in remap_area_supersections()
258 void __iomem * __arm_ioremap_pfn_caller(unsigned long pfn, in __arm_ioremap_pfn_caller() argument
[all …]
Dmmu.c659 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument
662 if (!pfn_valid(pfn)) in phys_mem_access_prot()
696 unsigned long end, unsigned long pfn, in alloc_init_pte() argument
701 set_pte_ext(pte, pfn_pte(pfn, __pgprot(type->prot_pte)), 0); in alloc_init_pte()
702 pfn++; in alloc_init_pte()
787 phys = __pfn_to_phys(md->pfn); in create_36bit_mapping()
792 (long long)__pfn_to_phys((u64)md->pfn), addr); in create_36bit_mapping()
804 (long long)__pfn_to_phys((u64)md->pfn), addr); in create_36bit_mapping()
808 if ((addr | length | __pfn_to_phys(md->pfn)) & ~SUPERSECTION_MASK) { in create_36bit_mapping()
810 (long long)__pfn_to_phys((u64)md->pfn), addr); in create_36bit_mapping()
[all …]
Dmmap.c215 int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) in valid_mmap_phys_addr_range() argument
217 return (pfn + (size >> PAGE_SHIFT)) <= (1 + (PHYS_MASK >> PAGE_SHIFT)); in valid_mmap_phys_addr_range()
231 int devmem_is_allowed(unsigned long pfn) in devmem_is_allowed() argument
233 if (iomem_is_exclusive(pfn << PAGE_SHIFT)) in devmem_is_allowed()
235 if (!page_is_ram(pfn)) in devmem_is_allowed()
Dnommu.c354 void __iomem *__arm_ioremap_pfn(unsigned long pfn, unsigned long offset, in __arm_ioremap_pfn() argument
357 if (pfn >= (0x100000000ULL >> PAGE_SHIFT)) in __arm_ioremap_pfn()
359 return (void __iomem *) (offset + (pfn << PAGE_SHIFT)); in __arm_ioremap_pfn()
363 void __iomem *__arm_ioremap_pfn_caller(unsigned long pfn, unsigned long offset, in __arm_ioremap_pfn_caller() argument
366 return __arm_ioremap_pfn(pfn, offset, size, mtype); in __arm_ioremap_pfn_caller()
Dhighmem.c127 void *kmap_atomic_pfn(unsigned long pfn) in kmap_atomic_pfn() argument
131 struct page *page = pfn_to_page(pfn); in kmap_atomic_pfn()
143 set_fixmap_pte(idx, pfn_pte(pfn, kmap_prot)); in kmap_atomic_pfn()
/linux-4.1.27/arch/ia64/include/asm/
Dpage.h98 extern int ia64_pfn_valid (unsigned long pfn);
100 # define ia64_pfn_valid(pfn) 1 argument
107 # define pfn_to_page(pfn) (vmem_map + (pfn)) argument
116 # define pfn_valid(pfn) (((pfn) < max_mapnr) && ia64_pfn_valid(pfn)) argument
120 # define pfn_valid(pfn) (((pfn) >= min_low_pfn) && ((pfn) < max_low_pfn) && ia64_pfn_valid(pfn)) argument
125 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
Dmmzone.h20 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument
23 int nid = paddr_to_nid(pfn << PAGE_SHIFT); in pfn_to_nid()
/linux-4.1.27/arch/x86/kvm/
Diommu.c50 pfn_t pfn; in kvm_pin_pages() local
52 pfn = gfn_to_pfn_memslot(slot, gfn); in kvm_pin_pages()
56 if (is_error_noslot_pfn(pfn)) in kvm_pin_pages()
57 return pfn; in kvm_pin_pages()
62 return pfn; in kvm_pin_pages()
65 static void kvm_unpin_pages(struct kvm *kvm, pfn_t pfn, unsigned long npages) in kvm_unpin_pages() argument
70 kvm_release_pfn_clean(pfn + i); in kvm_unpin_pages()
76 pfn_t pfn; in kvm_iommu_map_pages() local
123 pfn = kvm_pin_pages(slot, gfn, page_size >> PAGE_SHIFT); in kvm_iommu_map_pages()
124 if (is_error_noslot_pfn(pfn)) { in kvm_iommu_map_pages()
[all …]
Dmmu_audit.c100 pfn_t pfn; in audit_mappings() local
117 pfn = gfn_to_pfn_atomic(vcpu->kvm, gfn); in audit_mappings()
119 if (is_error_pfn(pfn)) in audit_mappings()
122 hpa = pfn << PAGE_SHIFT; in audit_mappings()
125 "ent %llxn", vcpu->arch.mmu.root_level, pfn, in audit_mappings()
Dmmu.c262 pfn_t pfn, unsigned access) in set_mmio_spte() argument
264 if (unlikely(is_noslot_pfn(pfn))) { in set_mmio_spte()
590 pfn_t pfn; in mmu_spte_clear_track_bits() local
601 pfn = spte_to_pfn(old_spte); in mmu_spte_clear_track_bits()
608 WARN_ON(!kvm_is_reserved_pfn(pfn) && !page_count(pfn_to_page(pfn))); in mmu_spte_clear_track_bits()
611 kvm_set_pfn_accessed(pfn); in mmu_spte_clear_track_bits()
613 kvm_set_pfn_dirty(pfn); in mmu_spte_clear_track_bits()
2515 gfn_t gfn, pfn_t pfn, bool speculative, in set_spte() argument
2521 if (set_mmio_spte(vcpu->kvm, sptep, gfn, pfn, pte_access)) in set_spte()
2540 kvm_is_reserved_pfn(pfn)); in set_spte()
[all …]
/linux-4.1.27/arch/tile/mm/
Dinit.c93 pte_t *get_prealloc_pte(unsigned long pfn) in get_prealloc_pte() argument
95 int node = pfn_to_nid(pfn); in get_prealloc_pte()
96 pfn &= ~(-1UL << (NR_PA_HIGHBIT_SHIFT - PAGE_SHIFT)); in get_prealloc_pte()
98 BUG_ON(pfn >= num_l2_ptes[node]); in get_prealloc_pte()
99 return &l2_ptes[node][pfn]; in get_prealloc_pte()
408 unsigned long address, pfn; in kernel_physical_mapping_init() local
453 pfn = node_start_pfn[i]; in kernel_physical_mapping_init()
456 init_prealloc_ptes(i, end_pfn - pfn); in kernel_physical_mapping_init()
458 address = (unsigned long) pfn_to_kaddr(pfn); in kernel_physical_mapping_init()
459 while (pfn < end_pfn) { in kernel_physical_mapping_init()
[all …]
/linux-4.1.27/arch/x86/mm/
Dinit_32.c261 unsigned long pfn; in kernel_physical_mapping_init() local
292 pfn = start_pfn; in kernel_physical_mapping_init()
293 pgd_idx = pgd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET); in kernel_physical_mapping_init()
298 if (pfn >= end_pfn) in kernel_physical_mapping_init()
301 pmd_idx = pmd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET); in kernel_physical_mapping_init()
306 for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn; in kernel_physical_mapping_init()
308 unsigned int addr = pfn * PAGE_SIZE + PAGE_OFFSET; in kernel_physical_mapping_init()
325 pfn &= PMD_MASK >> PAGE_SHIFT; in kernel_physical_mapping_init()
326 addr2 = (pfn + PTRS_PER_PTE-1) * PAGE_SIZE + in kernel_physical_mapping_init()
335 set_pmd(pmd, pfn_pmd(pfn, init_prot)); in kernel_physical_mapping_init()
[all …]
Dpat.c336 u64 pfn; in reserve_ram_pages_type() local
344 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in reserve_ram_pages_type()
347 page = pfn_to_page(pfn); in reserve_ram_pages_type()
362 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in reserve_ram_pages_type()
363 page = pfn_to_page(pfn); in reserve_ram_pages_type()
372 u64 pfn; in free_ram_pages_type() local
374 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { in free_ram_pages_type()
375 page = pfn_to_page(pfn); in free_ram_pages_type()
606 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument
614 static inline int range_is_allowed(unsigned long pfn, unsigned long size) in range_is_allowed() argument
[all …]
Dnuma_32.c49 unsigned long pfn; in memory_present() local
57 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { in memory_present()
58 physnode_map[pfn / PAGES_PER_SECTION] = nid; in memory_present()
59 printk(KERN_CONT "%lx ", pfn); in memory_present()
Dinit.c82 unsigned long pfn; in alloc_low_pages() local
103 pfn = ret >> PAGE_SHIFT; in alloc_low_pages()
105 pfn = pgt_buf_end; in alloc_low_pages()
108 pfn << PAGE_SHIFT, (pgt_buf_end << PAGE_SHIFT) - 1); in alloc_low_pages()
114 adr = __va((pfn + i) << PAGE_SHIFT); in alloc_low_pages()
118 return __va(pfn << PAGE_SHIFT); in alloc_low_pages()
266 unsigned long pfn; in split_mem_range() local
272 pfn = start_pfn = PFN_DOWN(start); in split_mem_range()
280 if (pfn == 0) in split_mem_range()
283 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
[all …]
Dpageattr.c38 unsigned long pfn; member
253 unsigned long pfn) in static_protections() argument
262 if (pcibios_enabled && within(pfn, BIOS_BEGIN >> PAGE_SHIFT, BIOS_END >> PAGE_SHIFT)) in static_protections()
278 if (within(pfn, __pa_symbol(__start_rodata) >> PAGE_SHIFT, in static_protections()
464 unsigned long nextpage_addr, numpages, pmask, psize, addr, pfn; in try_preserve_large_page() local
539 pfn = pte_pfn(old_pte) + ((address & (psize - 1)) >> PAGE_SHIFT); in try_preserve_large_page()
540 cpa->pfn = pfn; in try_preserve_large_page()
542 new_prot = static_protections(req_prot, address, pfn); in try_preserve_large_page()
550 pfn = pte_pfn(old_pte); in try_preserve_large_page()
551 for (i = 0; i < (psize >> PAGE_SHIFT); i++, addr += PAGE_SIZE, pfn++) { in try_preserve_large_page()
[all …]
Diomap_32.c57 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) in kmap_atomic_prot_pfn() argument
67 set_pte(kmap_pte - idx, pfn_pte(pfn, prot)); in kmap_atomic_prot_pfn()
77 iomap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) in iomap_atomic_prot_pfn() argument
90 return (void __force __iomem *) kmap_atomic_prot_pfn(pfn, prot); in iomap_atomic_prot_pfn()
Dpageattr-test.c134 unsigned long pfn = prandom_u32() % max_pfn_mapped; in pageattr_test() local
136 addr[i] = (unsigned long)__va(pfn << PAGE_SHIFT); in pageattr_test()
138 len[i] = min_t(unsigned long, len[i], max_pfn_mapped - pfn - 1); in pageattr_test()
162 if (test_bit(pfn + k, bm)) { in pageattr_test()
166 __set_bit(pfn + k, bm); in pageattr_test()
/linux-4.1.27/arch/frv/include/asm/
Dpage.h43 #define devmem_is_allowed(pfn) 1 argument
48 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
55 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
58 #define pfn_valid(pfn) ((pfn) >= min_low_pfn && (pfn) < max_low_pfn) argument
/linux-4.1.27/arch/x86/kernel/
Dcrash_dump_32.c17 static inline bool is_crashed_pfn_valid(unsigned long pfn) in is_crashed_pfn_valid() argument
27 return pte_pfn(pfn_pte(pfn, __pgprot(0))) == pfn; in is_crashed_pfn_valid()
50 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument
58 if (!is_crashed_pfn_valid(pfn)) in copy_oldmem_page()
61 vaddr = kmap_atomic_pfn(pfn); in copy_oldmem_page()
Dcrash_dump_64.c26 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument
34 vaddr = ioremap_cache(pfn << PAGE_SHIFT, PAGE_SIZE); in copy_oldmem_page()
/linux-4.1.27/arch/powerpc/kvm/
De500_mmu_host.c166 pfn_t pfn; in kvmppc_map_magic() local
168 pfn = (pfn_t)virt_to_phys((void *)shared_page) >> PAGE_SHIFT; in kvmppc_map_magic()
169 get_page(pfn_to_page(pfn)); in kvmppc_map_magic()
177 magic.mas7_3 = ((u64)pfn << PAGE_SHIFT) | in kvmppc_map_magic()
249 pfn_t pfn, unsigned int wimg) in kvmppc_e500_ref_setup() argument
251 ref->pfn = pfn; in kvmppc_e500_ref_setup()
258 kvm_set_pfn_accessed(pfn); in kvmppc_e500_ref_setup()
261 kvm_set_pfn_dirty(pfn); in kvmppc_e500_ref_setup()
268 trace_kvm_booke206_ref_release(ref->pfn, ref->flags); in kvmppc_e500_ref_release()
312 pfn_t pfn = ref->pfn; in kvmppc_e500_setup_stlbe() local
[all …]
Dtrace_booke.h137 TP_PROTO(__u64 pfn, __u32 flags),
138 TP_ARGS(pfn, flags),
141 __field( __u64, pfn )
146 __entry->pfn = pfn;
151 __entry->pfn, __entry->flags)
Dbook3s_64_mmu_host.c100 unsigned long pfn; in kvmppc_mmu_map_page() local
107 pfn = kvmppc_gpa_to_pfn(vcpu, orig_pte->raddr, iswrite, &writable); in kvmppc_mmu_map_page()
108 if (is_error_noslot_pfn(pfn)) { in kvmppc_mmu_map_page()
114 hpaddr = pfn << PAGE_SHIFT; in kvmppc_mmu_map_page()
134 kvm_set_pfn_accessed(pfn); in kvmppc_mmu_map_page()
139 kvm_set_pfn_dirty(pfn); in kvmppc_mmu_map_page()
145 kvmppc_mmu_flush_icache(pfn); in kvmppc_mmu_map_page()
199 cpte->pfn = pfn; in kvmppc_mmu_map_page()
208 kvm_release_pfn_clean(pfn); in kvmppc_mmu_map_page()
Dtrace_pr.h70 __field( u64, pfn )
79 __entry->pfn = pte->pfn;
89 __entry->host_vpn, __entry->pfn, __entry->eaddr,
99 __field( u64, pfn )
108 __entry->pfn = pte->pfn;
118 __entry->host_vpn, __entry->pfn, __entry->eaddr,
/linux-4.1.27/arch/s390/pci/
Dpci_mmio.c15 unsigned long *pfn) in get_pfn() argument
28 ret = follow_pfn(vma, user_addr, pfn); in get_pfn()
40 unsigned long pfn; in SYSCALL_DEFINE3() local
55 ret = get_pfn(mmio_addr, VM_WRITE, &pfn); in SYSCALL_DEFINE3()
58 io_addr = (void __iomem *)((pfn << PAGE_SHIFT) | (mmio_addr & ~PAGE_MASK)); in SYSCALL_DEFINE3()
80 unsigned long pfn; in SYSCALL_DEFINE3() local
95 ret = get_pfn(mmio_addr, VM_READ, &pfn); in SYSCALL_DEFINE3()
98 io_addr = (void __iomem *)((pfn << PAGE_SHIFT) | (mmio_addr & ~PAGE_MASK)); in SYSCALL_DEFINE3()
/linux-4.1.27/arch/powerpc/mm/
Dmem.c79 int page_is_ram(unsigned long pfn) in page_is_ram() argument
82 return pfn < max_pfn; in page_is_ram()
84 unsigned long paddr = (pfn << PAGE_SHIFT); in page_is_ram()
94 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument
98 return ppc_md.phys_mem_access_prot(file, pfn, size, vma_prot); in phys_mem_access_prot()
100 if (!page_is_ram(pfn)) in phys_mem_access_prot()
341 unsigned long pfn, highmem_mapnr; in mem_init() local
344 for (pfn = highmem_mapnr; pfn < max_mapnr; ++pfn) { in mem_init()
345 phys_addr_t paddr = (phys_addr_t)pfn << PAGE_SHIFT; in mem_init()
346 struct page *page = pfn_to_page(pfn); in mem_init()
[all …]
/linux-4.1.27/arch/tile/include/asm/
Dpage.h153 #define __pfn_to_highbits(pfn) ((pfn) >> (NR_PA_HIGHBIT_SHIFT - PAGE_SHIFT)) argument
253 static inline void *pfn_to_kaddr(unsigned long pfn) in pfn_to_kaddr() argument
255 return vbase_map[__pfn_to_highbits(pfn)] + (pfn << PAGE_SHIFT); in pfn_to_kaddr()
260 unsigned long pfn = kaddr_to_pfn(kaddr); in virt_to_phys() local
261 return ((phys_addr_t)pfn << PAGE_SHIFT) + in virt_to_phys()
284 static inline void *pfn_to_kaddr(unsigned long pfn) in pfn_to_kaddr() argument
286 return (void *)((pfn << PAGE_SHIFT) + PAGE_OFFSET); in pfn_to_kaddr()
311 static inline int pfn_valid(unsigned long pfn) in pfn_valid() argument
313 return pfn < max_mapnr; in pfn_valid()
Dmmzone.h38 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument
40 return highbits_to_node[__pfn_to_highbits(pfn)]; in pfn_to_nid()
45 static inline int pfn_valid(unsigned long pfn) in pfn_valid() argument
47 int nid = pfn_to_nid(pfn); in pfn_valid()
50 return (pfn < node_end_pfn(nid)); in pfn_valid()
Dhighmem.h64 void *kmap_atomic_pfn(unsigned long pfn);
65 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot);
/linux-4.1.27/arch/microblaze/include/asm/
Dpage.h142 extern int page_is_ram(unsigned long pfn);
145 # define pfn_to_phys(pfn) (PFN_PHYS(pfn)) argument
148 # define pfn_to_virt(pfn) __va(pfn_to_phys((pfn))) argument
165 # define pfn_valid(pfn) (((pfn) >= min_low_pfn) && \ argument
166 ((pfn) <= (min_low_pfn + max_mapnr)))
170 # define pfn_valid(pfn) ((pfn) < (max_mapnr + ARCH_PFN_OFFSET)) argument
Dcacheflush.h92 #define flush_cache_page(vma, vmaddr, pfn) \ argument
93 flush_dcache_range(pfn << PAGE_SHIFT, (pfn << PAGE_SHIFT) + PAGE_SIZE);
/linux-4.1.27/drivers/xen/
Dballoon.c243 unsigned long pfn, i; in reserve_additional_memory() local
245 pfn = PFN_DOWN(hotplug_start_paddr); in reserve_additional_memory()
247 if (!set_phys_to_machine(pfn + i, INVALID_P2M_ENTRY)) { in reserve_additional_memory()
329 unsigned long pfn, i; in increase_reservation() local
369 pfn = page_to_pfn(page); in increase_reservation()
373 set_phys_to_machine(pfn, frame_list[i]); in increase_reservation()
379 (unsigned long)__va(pfn << PAGE_SHIFT), in increase_reservation()
399 unsigned long pfn, i; in decrease_reservation() local
443 pfn = frame_list[i]; in decrease_reservation()
444 frame_list[i] = pfn_to_mfn(pfn); in decrease_reservation()
[all …]
Dtmem.c132 u32 index, unsigned long pfn) in xen_tmem_put_page() argument
134 unsigned long gmfn = xen_pv_domain() ? pfn_to_mfn(pfn) : pfn; in xen_tmem_put_page()
141 u32 index, unsigned long pfn) in xen_tmem_get_page() argument
143 unsigned long gmfn = xen_pv_domain() ? pfn_to_mfn(pfn) : pfn; in xen_tmem_get_page()
176 unsigned long pfn = page_to_pfn(page); in tmem_cleancache_put_page() local
183 (void)xen_tmem_put_page((u32)pool, oid, ind, pfn); in tmem_cleancache_put_page()
191 unsigned long pfn = page_to_pfn(page); in tmem_cleancache_get_page() local
199 ret = xen_tmem_get_page((u32)pool, oid, ind, pfn); in tmem_cleancache_get_page()
290 unsigned long pfn = page_to_pfn(page); in tmem_frontswap_store() local
299 ret = xen_tmem_put_page(pool, oswiz(type, ind), iswiz(ind), pfn); in tmem_frontswap_store()
[all …]
Dxlate_mmu.c81 unsigned long pfn = page_to_pfn(page); in remap_pte_fn() local
82 pte_t pte = pte_mkspecial(pfn_pte(pfn, info->prot)); in remap_pte_fn()
85 rc = map_foreign_page(pfn, *info->fgmfn, info->domid); in remap_pte_fn()
133 unsigned long pfn; in xen_xlate_unmap_gfn_range() local
135 pfn = page_to_pfn(pages[i]); in xen_xlate_unmap_gfn_range()
138 xrp.gpfn = pfn; in xen_xlate_unmap_gfn_range()
Dswiotlb-xen.c95 unsigned long pfn = mfn_to_pfn(PFN_DOWN(baddr)); in xen_bus_to_phys() local
96 dma_addr_t dma = (dma_addr_t)pfn << PAGE_SHIFT; in xen_bus_to_phys()
109 static int check_pages_physically_contiguous(unsigned long pfn, in check_pages_physically_contiguous() argument
117 next_mfn = pfn_to_mfn(pfn); in check_pages_physically_contiguous()
121 if (pfn_to_mfn(++pfn) != ++next_mfn) in check_pages_physically_contiguous()
129 unsigned long pfn = PFN_DOWN(p); in range_straddles_page_boundary() local
134 if (check_pages_physically_contiguous(pfn, offset, size)) in range_straddles_page_boundary()
142 unsigned long pfn = mfn_to_local_pfn(mfn); in is_xen_swiotlb_buffer() local
149 if (pfn_valid(pfn)) { in is_xen_swiotlb_buffer()
150 paddr = PFN_PHYS(pfn); in is_xen_swiotlb_buffer()
Dgrant-table.c395 int gnttab_grant_foreign_transfer(domid_t domid, unsigned long pfn) in gnttab_grant_foreign_transfer() argument
402 gnttab_grant_foreign_transfer_ref(ref, domid, pfn); in gnttab_grant_foreign_transfer()
409 unsigned long pfn) in gnttab_grant_foreign_transfer_ref() argument
411 gnttab_interface->update_entry(ref, domid, pfn, GTF_accept_transfer); in gnttab_grant_foreign_transfer_ref()
638 xen_pfn_t *pfn; in gnttab_setup_auto_xlat_frames() local
652 pfn = kcalloc(max_nr_gframes, sizeof(pfn[0]), GFP_KERNEL); in gnttab_setup_auto_xlat_frames()
653 if (!pfn) { in gnttab_setup_auto_xlat_frames()
658 pfn[i] = PFN_DOWN(addr) + i; in gnttab_setup_auto_xlat_frames()
661 xen_auto_xlat_grant_frames.pfn = pfn; in gnttab_setup_auto_xlat_frames()
672 kfree(xen_auto_xlat_grant_frames.pfn); in gnttab_free_auto_xlat_frames()
[all …]
/linux-4.1.27/arch/metag/include/asm/
Dmmzone.h10 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument
15 if (pfn >= node_start_pfn(nid) && pfn <= node_end_pfn(nid)) in pfn_to_nid()
21 static inline struct pglist_data *pfn_to_pgdat(unsigned long pfn) in pfn_to_pgdat() argument
23 return NODE_DATA(pfn_to_nid(pfn)); in pfn_to_pgdat()
Dpage.h115 #define pfn_valid(pfn) ((pfn) >= min_low_pfn && (pfn) < max_pfn) argument
118 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
/linux-4.1.27/arch/m68k/include/asm/
Dpage_no.h26 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument
31 #define pfn_to_page(pfn) virt_to_page(pfn_to_virt(pfn)) argument
33 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
Dpage_mm.h127 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument
158 #define pfn_to_page(pfn) ({ \ argument
159 unsigned long __pfn = (pfn); \
172 #define pfn_valid(pfn) virt_addr_valid(pfn_to_virt(pfn)) argument
/linux-4.1.27/arch/sh/include/asm/
Dmmzone.h12 static inline int pfn_to_nid(unsigned long pfn) in pfn_to_nid() argument
17 if (pfn >= node_start_pfn(nid) && pfn <= node_end_pfn(nid)) in pfn_to_nid()
23 static inline struct pglist_data *pfn_to_pgdat(unsigned long pfn) in pfn_to_pgdat() argument
25 return NODE_DATA(pfn_to_nid(pfn)); in pfn_to_pgdat()
Dpage.h166 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
179 #define pfn_valid(pfn) ((pfn) >= min_low_pfn && (pfn) < max_low_pfn) argument
/linux-4.1.27/arch/arm/mach-omap1/
Dio.c34 .pfn = __phys_to_pfn(OMAP1_IO_PHYS),
44 .pfn = __phys_to_pfn(OMAP7XX_DSP_START),
49 .pfn = __phys_to_pfn(OMAP7XX_DSPREG_START),
60 .pfn = __phys_to_pfn(OMAP1510_DSP_START),
65 .pfn = __phys_to_pfn(OMAP1510_DSPREG_START),
76 .pfn = __phys_to_pfn(OMAP16XX_DSP_START),
81 .pfn = __phys_to_pfn(OMAP16XX_DSPREG_START),
/linux-4.1.27/arch/score/include/asm/
Dpage.h65 #define pfn_to_phys(pfn) (PFN_PHYS(pfn)) argument
68 #define pfn_to_virt(pfn) __va(pfn_to_phys((pfn))) argument
77 #define pfn_valid(pfn) (((pfn) >= min_low_pfn) && ((pfn) < max_low_pfn)) argument
/linux-4.1.27/drivers/base/
Dnode.c362 static int get_nid_for_pfn(unsigned long pfn) in get_nid_for_pfn() argument
366 if (!pfn_valid_within(pfn)) in get_nid_for_pfn()
368 page = pfn_to_page(pfn); in get_nid_for_pfn()
371 return pfn_to_nid(pfn); in get_nid_for_pfn()
378 unsigned long pfn, sect_start_pfn, sect_end_pfn; in register_mem_sect_under_node() local
388 for (pfn = sect_start_pfn; pfn <= sect_end_pfn; pfn++) { in register_mem_sect_under_node()
395 if (!pfn_present(pfn)) { in register_mem_sect_under_node()
396 pfn = round_down(pfn + PAGES_PER_SECTION, in register_mem_sect_under_node()
401 page_nid = get_nid_for_pfn(pfn); in register_mem_sect_under_node()
425 unsigned long pfn, sect_start_pfn, sect_end_pfn; in unregister_mem_sect_under_nodes() local
[all …]
Dmemory.c127 unsigned long i, pfn; in show_mem_removable() local
134 pfn = section_nr_to_pfn(mem->start_section_nr + i); in show_mem_removable()
135 ret &= is_mem_section_removable(pfn, PAGES_PER_SECTION); in show_mem_removable()
192 unsigned long pfn = start_pfn; in pages_correctly_reserved() local
199 for (i = 0; i < sections_per_block; i++, pfn += PAGES_PER_SECTION) { in pages_correctly_reserved()
200 if (WARN_ON_ONCE(!pfn_valid(pfn))) in pages_correctly_reserved()
202 page = pfn_to_page(pfn); in pages_correctly_reserved()
210 pfn_to_section_nr(pfn), j); in pages_correctly_reserved()
490 u64 pfn; in store_soft_offline_page() local
493 if (kstrtoull(buf, 0, &pfn) < 0) in store_soft_offline_page()
[all …]
Ddma-mapping.c253 unsigned long pfn = page_to_pfn(virt_to_page(cpu_addr)); in dma_common_mmap() local
263 pfn + off, in dma_common_mmap()
310 unsigned long pfn; in dma_common_contiguous_remap() local
316 for (i = 0, pfn = page_to_pfn(page); i < (size >> PAGE_SHIFT); i++) in dma_common_contiguous_remap()
317 pages[i] = pfn_to_page(pfn + i); in dma_common_contiguous_remap()
/linux-4.1.27/fs/proc/
Dpage.c30 unsigned long pfn; in kpagecount_read() local
34 pfn = src / KPMSIZE; in kpagecount_read()
40 if (pfn_valid(pfn)) in kpagecount_read()
41 ppage = pfn_to_page(pfn); in kpagecount_read()
54 pfn++; in kpagecount_read()
193 unsigned long pfn; in kpageflags_read() local
196 pfn = src / KPMSIZE; in kpageflags_read()
202 if (pfn_valid(pfn)) in kpageflags_read()
203 ppage = pfn_to_page(pfn); in kpageflags_read()
212 pfn++; in kpageflags_read()
Dvmcore.c51 static int (*oldmem_pfn_is_ram)(unsigned long pfn);
53 int register_oldmem_pfn_is_ram(int (*fn)(unsigned long pfn)) in register_oldmem_pfn_is_ram() argument
69 static int pfn_is_ram(unsigned long pfn) in pfn_is_ram() argument
71 int (*fn)(unsigned long pfn); in pfn_is_ram()
82 ret = fn(pfn); in pfn_is_ram()
91 unsigned long pfn, offset; in read_from_oldmem() local
99 pfn = (unsigned long)(*ppos / PAGE_SIZE); in read_from_oldmem()
108 if (pfn_is_ram(pfn) == 0) in read_from_oldmem()
111 tmp = copy_oldmem_page(pfn, buf, nr_bytes, in read_from_oldmem()
120 ++pfn; in read_from_oldmem()
[all …]
/linux-4.1.27/arch/parisc/kernel/
Dcache.c75 #define pfn_va(pfn) __va(PFN_PHYS(pfn)) argument
80 unsigned long pfn = pte_pfn(*ptep); in update_mmu_cache() local
86 if (!pfn_valid(pfn)) in update_mmu_cache()
89 page = pfn_to_page(pfn); in update_mmu_cache()
91 flush_kernel_dcache_page_addr(pfn_va(pfn)); in update_mmu_cache()
94 flush_kernel_dcache_page_addr(pfn_va(pfn)); in update_mmu_cache()
542 unsigned long pfn; in flush_cache_mm() local
546 pfn = pte_pfn(*ptep); in flush_cache_mm()
547 if (!pfn_valid(pfn)) in flush_cache_mm()
549 __flush_cache_page(vma, addr, PFN_PHYS(pfn)); in flush_cache_mm()
[all …]
/linux-4.1.27/arch/powerpc/kernel/
Ddma.c31 u64 pfn = (dev->coherent_dma_mask >> PAGE_SHIFT) + 1; in get_pfn_limit() local
36 pfn = min_t(u64, pfn, sd->max_direct_dma_addr >> PAGE_SHIFT); in get_pfn_limit()
39 return pfn; in get_pfn_limit()
57 u64 pfn = get_pfn_limit(dev); in dma_direct_alloc_coherent() local
66 zone = dma_pfn_limit_to_zone(pfn); in dma_direct_alloc_coherent()
69 __func__, pfn); in dma_direct_alloc_coherent()
114 unsigned long pfn; in dma_direct_mmap_coherent() local
118 pfn = __dma_get_coherent_pfn((unsigned long)cpu_addr); in dma_direct_mmap_coherent()
120 pfn = page_to_pfn(virt_to_page(cpu_addr)); in dma_direct_mmap_coherent()
123 pfn + vma->vm_pgoff, in dma_direct_mmap_coherent()
Dsuspend.c18 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
22 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
/linux-4.1.27/arch/arm64/mm/
Dmmap.c114 int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) in valid_mmap_phys_addr_range() argument
116 return !(((pfn << PAGE_SHIFT) + size) & ~PHYS_MASK); in valid_mmap_phys_addr_range()
129 int devmem_is_allowed(unsigned long pfn) in devmem_is_allowed() argument
131 if (iomem_is_exclusive(pfn << PAGE_SHIFT)) in devmem_is_allowed()
133 if (!page_is_ram(pfn)) in devmem_is_allowed()
/linux-4.1.27/arch/arm/mach-exynos/
Dexynos.c41 .pfn = __phys_to_pfn(EXYNOS4_PA_SROMC),
46 .pfn = __phys_to_pfn(EXYNOS4_PA_CMU),
51 .pfn = __phys_to_pfn(EXYNOS4_PA_COREPERI),
56 .pfn = __phys_to_pfn(EXYNOS4_PA_DMC0),
61 .pfn = __phys_to_pfn(EXYNOS4_PA_DMC1),
70 .pfn = __phys_to_pfn(EXYNOS5_PA_SROMC),
75 .pfn = __phys_to_pfn(EXYNOS5_PA_CMU),
135 iodesc.pfn = __phys_to_pfn(be32_to_cpu(reg[0])); in exynos_fdt_map_chipid()
/linux-4.1.27/drivers/vfio/
Dvfio_iommu_type1.c191 static bool is_invalid_reserved_pfn(unsigned long pfn) in is_invalid_reserved_pfn() argument
193 if (pfn_valid(pfn)) { in is_invalid_reserved_pfn()
195 struct page *tail = pfn_to_page(pfn); in is_invalid_reserved_pfn()
219 static int put_pfn(unsigned long pfn, int prot) in put_pfn() argument
221 if (!is_invalid_reserved_pfn(pfn)) { in put_pfn()
222 struct page *page = pfn_to_page(pfn); in put_pfn()
231 static int vaddr_get_pfn(unsigned long vaddr, int prot, unsigned long *pfn) in vaddr_get_pfn() argument
238 *pfn = page_to_pfn(page[0]); in vaddr_get_pfn()
247 *pfn = ((vaddr - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; in vaddr_get_pfn()
248 if (is_invalid_reserved_pfn(*pfn)) in vaddr_get_pfn()
[all …]
/linux-4.1.27/arch/mn10300/include/asm/
Dpage.h107 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
108 #define pfn_to_page(pfn) (mem_map + ((pfn) - __pfn_disp)) argument
111 #define pfn_valid(pfn) \ argument
113 unsigned long __pfn = (pfn) - __pfn_disp; \
Dpgtable.h317 #define __pfn_addr(pfn) ((pfn) << PAGE_SHIFT) argument
318 #define pfn_pte(pfn, prot) __pte(__pfn_addr(pfn) | pgprot_val(prot)) argument
319 #define pfn_pmd(pfn, prot) __pmd(__pfn_addr(pfn) | pgprot_val(prot)) argument
480 #define MK_IOSPACE_PFN(space, pfn) (pfn) argument
481 #define GET_IOSPACE(pfn) 0 argument
482 #define GET_PFN(pfn) (pfn) argument
/linux-4.1.27/arch/arm/mach-footbridge/
Dcommon.c143 .pfn = __phys_to_pfn(DC21285_ARMCSR_BASE),
157 .pfn = __phys_to_pfn(DC21285_PCI_MEM),
162 .pfn = __phys_to_pfn(DC21285_PCI_TYPE_0_CONFIG),
167 .pfn = __phys_to_pfn(DC21285_PCI_TYPE_1_CONFIG),
172 .pfn = __phys_to_pfn(DC21285_PCI_IACK),
256 unsigned long __pfn_to_bus(unsigned long pfn) in __pfn_to_bus() argument
258 return __pfn_to_phys(pfn) + (fb_bus_sdram_offset() - PHYS_OFFSET); in __pfn_to_bus()
/linux-4.1.27/arch/frv/mm/
Dinit.c123 unsigned long pfn; in mem_init() local
125 for (pfn = get_num_physpages() - 1; in mem_init()
126 pfn >= num_mappedpages; pfn--) in mem_init()
127 free_highmem_page(&mem_map[pfn]); in mem_init()
/linux-4.1.27/arch/arm/include/asm/
Ddma-mapping.h59 static inline dma_addr_t pfn_to_dma(struct device *dev, unsigned long pfn) in pfn_to_dma() argument
62 pfn -= dev->dma_pfn_offset; in pfn_to_dma()
63 return (dma_addr_t)__pfn_to_bus(pfn); in pfn_to_dma()
68 unsigned long pfn = __bus_to_pfn(addr); in dma_to_pfn() local
71 pfn += dev->dma_pfn_offset; in dma_to_pfn()
73 return pfn; in dma_to_pfn()
79 unsigned long pfn = dma_to_pfn(dev, addr); in dma_to_virt() local
81 return phys_to_virt(__pfn_to_phys(pfn)); in dma_to_virt()
96 static inline dma_addr_t pfn_to_dma(struct device *dev, unsigned long pfn) in pfn_to_dma() argument
98 return __arch_pfn_to_dma(dev, pfn); in pfn_to_dma()
Dkvm_mmu.h185 static inline void __coherent_cache_guest_page(struct kvm_vcpu *vcpu, pfn_t pfn, in __coherent_cache_guest_page() argument
215 void *va = kmap_atomic_pfn(pfn); in __coherent_cache_guest_page()
225 pfn++; in __coherent_cache_guest_page()
249 pfn_t pfn = pmd_pfn(pmd); in __kvm_flush_dcache_pmd() local
252 void *va = kmap_atomic_pfn(pfn); in __kvm_flush_dcache_pmd()
256 pfn++; in __kvm_flush_dcache_pmd()
Dmemory.h127 #define __pfn_to_phys(pfn) ((phys_addr_t)(pfn) << PAGE_SHIFT) argument
294 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
/linux-4.1.27/arch/arm/mach-tegra/
Dio.c36 .pfn = __phys_to_pfn(IO_PPSB_PHYS),
42 .pfn = __phys_to_pfn(IO_APB_PHYS),
48 .pfn = __phys_to_pfn(IO_CPU_PHYS),
54 .pfn = __phys_to_pfn(IO_IRAM_PHYS),
/linux-4.1.27/arch/mips/include/asm/
Dpage.h195 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
199 static inline int pfn_valid(unsigned long pfn) in pfn_valid() argument
204 return pfn >= ARCH_PFN_OFFSET && pfn < max_mapnr; in pfn_valid()
213 #define pfn_valid(pfn) \ argument
215 unsigned long __pfn = (pfn); \
Dpgtable-32.h110 pfn_pte(unsigned long pfn, pgprot_t prot) in pfn_pte() argument
114 pte.pte_low = (pfn >> _PAGE_PRESENT_SHIFT) | in pfn_pte()
116 pte.pte_high = (pfn << _PFN_SHIFT) | in pfn_pte()
127 #define pfn_pte(pfn, prot) __pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot)) argument
130 #define pfn_pte(pfn, prot) __pte(((unsigned long long)(pfn) << _PFN_SHIFT) | pgprot_val(prot)) argument
Dpgtable-64.h229 #define pfn_pte(pfn, prot) __pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot)) argument
232 #define pfn_pte(pfn, prot) __pte(((pfn) << _PFN_SHIFT) | pgprot_val(prot)) argument
233 #define pfn_pmd(pfn, prot) __pmd(((pfn) << _PFN_SHIFT) | pgprot_val(prot)) argument
Dmmzone.h13 #define pfn_to_nid(pfn) pa_to_nid((pfn) << PAGE_SHIFT) argument
/linux-4.1.27/arch/arm/kvm/
Dmmu.c101 static bool kvm_is_device_pfn(unsigned long pfn) in kvm_is_device_pfn() argument
103 return !pfn_valid(pfn); in kvm_is_device_pfn()
452 unsigned long end, unsigned long pfn, in create_hyp_pte_mappings() argument
461 kvm_set_pte(pte, pfn_pte(pfn, prot)); in create_hyp_pte_mappings()
464 pfn++; in create_hyp_pte_mappings()
469 unsigned long end, unsigned long pfn, in create_hyp_pmd_mappings() argument
495 create_hyp_pte_mappings(pmd, addr, next, pfn, prot); in create_hyp_pmd_mappings()
496 pfn += (next - addr) >> PAGE_SHIFT; in create_hyp_pmd_mappings()
503 unsigned long end, unsigned long pfn, in create_hyp_pud_mappings() argument
527 ret = create_hyp_pmd_mappings(pud, addr, next, pfn, prot); in create_hyp_pud_mappings()
[all …]
/linux-4.1.27/arch/avr32/include/asm/
Dpgtable-2level.h42 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument
43 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) argument
Dpage.h88 #define pfn_to_page(pfn) (mem_map + ((pfn) - PHYS_PFN_OFFSET)) argument
90 #define pfn_valid(pfn) ((pfn) >= PHYS_PFN_OFFSET && (pfn) < (PHYS_PFN_OFFSET + max_mapnr)) argument
/linux-4.1.27/arch/arm/mach-cns3xxx/
Dcore.c32 .pfn = __phys_to_pfn(CNS3XXX_TC11MP_SCU_BASE),
37 .pfn = __phys_to_pfn(CNS3XXX_TIMER1_2_3_BASE),
42 .pfn = __phys_to_pfn(CNS3XXX_MISC_BASE),
47 .pfn = __phys_to_pfn(CNS3XXX_PM_BASE),
53 .pfn = __phys_to_pfn(CNS3XXX_PCIE0_HOST_BASE),
58 .pfn = __phys_to_pfn(CNS3XXX_PCIE0_CFG0_BASE),
63 .pfn = __phys_to_pfn(CNS3XXX_PCIE0_CFG1_BASE),
68 .pfn = __phys_to_pfn(CNS3XXX_PCIE1_HOST_BASE),
73 .pfn = __phys_to_pfn(CNS3XXX_PCIE1_CFG0_BASE),
78 .pfn = __phys_to_pfn(CNS3XXX_PCIE1_CFG1_BASE),
/linux-4.1.27/arch/microblaze/mm/
Dinit.c76 unsigned long pfn; in highmem_setup() local
78 for (pfn = max_low_pfn; pfn < max_pfn; ++pfn) { in highmem_setup()
79 struct page *page = pfn_to_page(pfn); in highmem_setup()
82 if (!memblock_is_reserved(pfn << PAGE_SHIFT)) in highmem_setup()
267 int page_is_ram(unsigned long pfn) in page_is_ram() argument
269 return __range_ok(pfn, 0); in page_is_ram()
272 int page_is_ram(unsigned long pfn) in page_is_ram() argument
274 return pfn < max_low_pfn; in page_is_ram()
Dconsistent.c204 unsigned long pfn; in consistent_free() local
207 pfn = pte_pfn(*ptep); in consistent_free()
209 if (pfn_valid(pfn)) { in consistent_free()
210 page = pfn_to_page(pfn); in consistent_free()
/linux-4.1.27/arch/arm/mach-realview/
Drealview_pbx.c53 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE),
58 .pfn = __phys_to_pfn(REALVIEW_PBX_GIC_CPU_BASE),
63 .pfn = __phys_to_pfn(REALVIEW_PBX_GIC_DIST_BASE),
68 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE),
73 .pfn = __phys_to_pfn(REALVIEW_PBX_TIMER0_1_BASE),
78 .pfn = __phys_to_pfn(REALVIEW_PBX_TIMER2_3_BASE),
85 .pfn = __phys_to_pfn(REALVIEW_PBX_PCI_BASE),
93 .pfn = __phys_to_pfn(REALVIEW_PBX_UART0_BASE),
103 .pfn = __phys_to_pfn(REALVIEW_PBX_TILE_SCU_BASE),
108 .pfn = __phys_to_pfn(REALVIEW_PBX_TILE_GIC_DIST_BASE),
[all …]
Drealview_pb1176.c56 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE),
61 .pfn = __phys_to_pfn(REALVIEW_PB1176_GIC_CPU_BASE),
66 .pfn = __phys_to_pfn(REALVIEW_PB1176_GIC_DIST_BASE),
71 .pfn = __phys_to_pfn(REALVIEW_DC1176_GIC_CPU_BASE),
76 .pfn = __phys_to_pfn(REALVIEW_DC1176_GIC_DIST_BASE),
81 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE),
86 .pfn = __phys_to_pfn(REALVIEW_PB1176_TIMER0_1_BASE),
91 .pfn = __phys_to_pfn(REALVIEW_PB1176_TIMER2_3_BASE),
96 .pfn = __phys_to_pfn(REALVIEW_PB1176_L220_BASE),
103 .pfn = __phys_to_pfn(REALVIEW_PB1176_UART0_BASE),
Drealview_pba8.c51 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE),
56 .pfn = __phys_to_pfn(REALVIEW_PBA8_GIC_CPU_BASE),
61 .pfn = __phys_to_pfn(REALVIEW_PBA8_GIC_DIST_BASE),
66 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE),
71 .pfn = __phys_to_pfn(REALVIEW_PBA8_TIMER0_1_BASE),
76 .pfn = __phys_to_pfn(REALVIEW_PBA8_TIMER2_3_BASE),
83 .pfn = __phys_to_pfn(REALVIEW_PBA8_PCI_BASE),
91 .pfn = __phys_to_pfn(REALVIEW_PBA8_UART0_BASE),
Drealview_pb11mp.c54 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE),
59 .pfn = __phys_to_pfn(REALVIEW_PB11MP_GIC_CPU_BASE),
64 .pfn = __phys_to_pfn(REALVIEW_PB11MP_GIC_DIST_BASE),
69 .pfn = __phys_to_pfn(REALVIEW_TC11MP_PRIV_MEM_BASE),
74 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE),
79 .pfn = __phys_to_pfn(REALVIEW_PB11MP_TIMER0_1_BASE),
84 .pfn = __phys_to_pfn(REALVIEW_PB11MP_TIMER2_3_BASE),
89 .pfn = __phys_to_pfn(REALVIEW_TC11MP_L220_BASE),
96 .pfn = __phys_to_pfn(REALVIEW_PB11MP_UART0_BASE),
Drealview_eb.c54 .pfn = __phys_to_pfn(REALVIEW_SYS_BASE),
59 .pfn = __phys_to_pfn(REALVIEW_EB_GIC_CPU_BASE),
64 .pfn = __phys_to_pfn(REALVIEW_EB_GIC_DIST_BASE),
69 .pfn = __phys_to_pfn(REALVIEW_SCTL_BASE),
74 .pfn = __phys_to_pfn(REALVIEW_EB_TIMER0_1_BASE),
79 .pfn = __phys_to_pfn(REALVIEW_EB_TIMER2_3_BASE),
86 .pfn = __phys_to_pfn(REALVIEW_EB_UART0_BASE),
96 .pfn = __phys_to_pfn(REALVIEW_EB11MP_PRIV_MEM_BASE),
101 .pfn = __phys_to_pfn(REALVIEW_EB11MP_L220_BASE),
/linux-4.1.27/arch/um/include/asm/
Dpgtable-2level.h41 #define pfn_pte(pfn, prot) __pte(pfn_to_phys(pfn) | pgprot_val(prot)) argument
42 #define pfn_pmd(pfn, prot) __pmd(pfn_to_phys(pfn) | pgprot_val(prot)) argument
Dpage.h113 #define pfn_to_phys(pfn) ((phys_t) ((pfn) << PAGE_SHIFT)) argument
115 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
/linux-4.1.27/arch/nios2/include/asm/
Dpage.h89 # define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
90 # define pfn_valid(pfn) ((pfn) >= ARCH_PFN_OFFSET && \ argument
91 (pfn) < max_mapnr)
/linux-4.1.27/arch/sparc/include/asm/
Dpgtable_32.h262 #define pfn_pte(pfn, prot) mk_pte(pfn_to_page(pfn), prot) argument
405 #define MK_IOSPACE_PFN(space, pfn) (pfn | (space << (BITS_PER_LONG - 4))) argument
406 #define GET_IOSPACE(pfn) (pfn >> (BITS_PER_LONG - 4)) argument
407 #define GET_PFN(pfn) (pfn & 0x0fffffffUL) argument
413 unsigned long from, unsigned long pfn, in io_remap_pfn_range() argument
418 offset = ((unsigned long long) GET_PFN(pfn)) << PAGE_SHIFT; in io_remap_pfn_range()
419 space = GET_IOSPACE(pfn); in io_remap_pfn_range()
Dpgtable_64.h237 static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot) in pfn_pte() argument
239 unsigned long paddr = pfn << PAGE_SHIFT; in pfn_pte()
801 unsigned long pfn; in __pmd_page() local
803 pfn = pte_pfn(pte); in __pmd_page()
805 return ((unsigned long) __va(pfn << PAGE_SHIFT)); in __pmd_page()
966 #define MK_IOSPACE_PFN(space, pfn) (pfn | (space << (BITS_PER_LONG - 4))) argument
967 #define GET_IOSPACE(pfn) (pfn >> (BITS_PER_LONG - 4)) argument
968 #define GET_PFN(pfn) (pfn & 0x0fffffffffffffffUL) argument
974 unsigned long from, unsigned long pfn, in io_remap_pfn_range() argument
977 unsigned long offset = GET_PFN(pfn) << PAGE_SHIFT; in io_remap_pfn_range()
[all …]
/linux-4.1.27/arch/arm/mach-s3c64xx/
Dcommon.c109 .pfn = __phys_to_pfn(S3C64XX_PA_SYSCON),
114 .pfn = __phys_to_pfn(S3C64XX_PA_SROM),
119 .pfn = __phys_to_pfn(S3C_PA_UART),
124 .pfn = __phys_to_pfn(S3C64XX_PA_VIC0),
129 .pfn = __phys_to_pfn(S3C64XX_PA_VIC1),
134 .pfn = __phys_to_pfn(S3C_PA_TIMER),
139 .pfn = __phys_to_pfn(S3C64XX_PA_GPIO),
144 .pfn = __phys_to_pfn(S3C64XX_PA_MODEM),
149 .pfn = __phys_to_pfn(S3C64XX_PA_WATCHDOG),
154 .pfn = __phys_to_pfn(S3C64XX_PA_USB_HSPHY),
/linux-4.1.27/drivers/media/v4l2-core/
Dvideobuf2-dma-contig.c440 unsigned long pfn, start_pfn, prev_pfn; in vb2_dc_get_user_pfn() local
447 ret = follow_pfn(vma, start, &pfn); in vb2_dc_get_user_pfn()
451 start_pfn = pfn; in vb2_dc_get_user_pfn()
455 prev_pfn = pfn; in vb2_dc_get_user_pfn()
456 ret = follow_pfn(vma, start, &pfn); in vb2_dc_get_user_pfn()
462 if (pfn != prev_pfn + 1) in vb2_dc_get_user_pfn()
478 unsigned long pfn; in vb2_dc_get_user_pages() local
479 int ret = follow_pfn(vma, start, &pfn); in vb2_dc_get_user_pages()
481 if (!pfn_valid(pfn)) in vb2_dc_get_user_pages()
488 pages[i] = pfn_to_page(pfn); in vb2_dc_get_user_pages()
[all …]
/linux-4.1.27/arch/arm/mach-integrator/
Dintegrator_cp.c69 .pfn = __phys_to_pfn(INTEGRATOR_HDR_BASE),
74 .pfn = __phys_to_pfn(INTEGRATOR_CT_BASE),
79 .pfn = __phys_to_pfn(INTEGRATOR_IC_BASE),
84 .pfn = __phys_to_pfn(INTEGRATOR_UART0_BASE),
89 .pfn = __phys_to_pfn(INTEGRATOR_DBG_BASE),
94 .pfn = __phys_to_pfn(INTEGRATOR_CP_GPIO_BASE),
99 .pfn = __phys_to_pfn(INTEGRATOR_CP_SIC_BASE),
/linux-4.1.27/arch/sh/kernel/
Dswsusp.c22 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
27 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
Dcrash_dump.c25 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument
33 vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); in copy_oldmem_page()
/linux-4.1.27/arch/unicore32/include/asm/
Dmemory.h67 #define __pfn_to_phys(pfn) ((pfn) << PAGE_SHIFT) argument
96 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
Dio.h63 static inline int devmem_is_allowed(unsigned long pfn) in devmem_is_allowed() argument
65 if (iomem_is_exclusive(pfn << PAGE_SHIFT)) in devmem_is_allowed()
67 if (!page_is_ram(pfn)) in devmem_is_allowed()
/linux-4.1.27/arch/sparc/power/
Dhibernate.c21 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
26 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
/linux-4.1.27/arch/arm/mach-spear/
Dspear13xx.c63 .pfn = __phys_to_pfn(PERIP_GRP2_BASE),
68 .pfn = __phys_to_pfn(PERIP_GRP1_BASE),
73 .pfn = __phys_to_pfn(A9SM_AND_MPMC_BASE),
78 .pfn = __phys_to_pfn(L2CC_BASE),
/linux-4.1.27/arch/powerpc/include/asm/
Dpage.h127 #define pfn_valid(pfn) ((pfn) >= ARCH_PFN_OFFSET && (pfn) < max_mapnr) argument
131 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
395 extern int page_is_ram(unsigned long pfn);
396 extern int devmem_is_allowed(unsigned long pfn);
Dpte-hash64-64k.h97 #define remap_4k_pfn(vma, addr, pfn, prot) \ argument
98 (WARN_ON(((pfn) >= (1UL << (64 - PTE_RPN_SHIFT)))) ? -EINVAL : \
99 remap_pfn_range((vma), (addr), (pfn), PAGE_SIZE, \
/linux-4.1.27/arch/mips/power/
Dcpu.c37 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
42 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn); in pfn_is_nosave()
/linux-4.1.27/arch/arm/mach-s3c24xx/
Dmach-smdk2443.c53 .pfn = __phys_to_pfn(S3C2410_CS2),
58 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
63 .pfn = __phys_to_pfn(S3C2410_CS2),
68 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
Dmach-smdk2440.c53 .pfn = __phys_to_pfn(S3C2410_CS2),
58 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
63 .pfn = __phys_to_pfn(S3C2410_CS2),
68 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
Dmach-bast.c83 .pfn = PA_CS2(BAST_PA_ISAIO),
88 .pfn = PA_CS3(BAST_PA_ISAIO),
95 .pfn = __phys_to_pfn(BAST_PA_CTRL1),
100 .pfn = __phys_to_pfn(BAST_PA_CTRL2),
105 .pfn = __phys_to_pfn(BAST_PA_CTRL3),
110 .pfn = __phys_to_pfn(BAST_PA_CTRL4),
117 .pfn = __phys_to_pfn(BAST_PA_PC104_IRQREQ),
122 .pfn = __phys_to_pfn(BAST_PA_PC104_IRQRAW),
127 .pfn = __phys_to_pfn(BAST_PA_PC104_IRQMASK),
Dmach-vr1000.c72 .pfn = PA_CS2(BAST_PA_ISAIO),
77 .pfn = PA_CS3(BAST_PA_ISAIO),
85 .pfn = __phys_to_pfn(VR1000_PA_CTRL1),
90 .pfn = __phys_to_pfn(VR1000_PA_CTRL2),
95 .pfn = __phys_to_pfn(VR1000_PA_CTRL3),
100 .pfn = __phys_to_pfn(VR1000_PA_CTRL4),
Dmach-osiris.c65 .pfn = __phys_to_pfn(S3C2410_CS5),
70 .pfn = __phys_to_pfn(S3C2410_CS5),
79 .pfn = __phys_to_pfn(OSIRIS_PA_CTRL0),
84 .pfn = __phys_to_pfn(OSIRIS_PA_CTRL1),
89 .pfn = __phys_to_pfn(OSIRIS_PA_CTRL2),
94 .pfn = __phys_to_pfn(OSIRIS_PA_IDREG),
Dmach-smdk2416.c65 .pfn = __phys_to_pfn(S3C2410_CS2),
70 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
75 .pfn = __phys_to_pfn(S3C2410_CS2),
80 .pfn = __phys_to_pfn(S3C2410_CS2 + (1<<24)),
/linux-4.1.27/arch/arm/mach-ebsa110/
Dcore.c78 .pfn = __phys_to_pfn(TRICK4_PHYS),
83 .pfn = __phys_to_pfn(TRICK3_PHYS),
88 .pfn = __phys_to_pfn(TRICK1_PHYS),
93 .pfn = __phys_to_pfn(TRICK0_PHYS),
103 .pfn = __phys_to_pfn(ISAIO_PHYS),
108 .pfn = __phys_to_pfn(ISAMEM_PHYS),
/linux-4.1.27/arch/score/mm/
Dcache.c82 unsigned long pfn, addr; in __update_cache() local
85 pfn = pte_pfn(pte); in __update_cache()
86 if (unlikely(!pfn_valid(pfn))) in __update_cache()
88 page = pfn_to_page(pfn); in __update_cache()
212 unsigned long addr, unsigned long pfn) in flush_cache_page() argument
215 unsigned long kaddr = 0xa0000000 | (pfn << PAGE_SHIFT); in flush_cache_page()
/linux-4.1.27/arch/openrisc/include/asm/
Dpage.h83 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument
92 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
/linux-4.1.27/arch/tile/gxio/
Dkiorpc.c34 unsigned long pfn; in iorpc_ioremap() local
56 pfn = pte_pfn(mmio_base) + PFN_DOWN(offset); in iorpc_ioremap()
58 return ioremap_prot(PFN_PHYS(pfn), size, prot); in iorpc_ioremap()
/linux-4.1.27/drivers/infiniband/hw/mlx5/
Dmem.c56 u64 pfn; in mlx5_ib_cont_pages() local
80 pfn = sg_dma_address(sg) >> page_shift; in mlx5_ib_cont_pages()
83 tmp = (unsigned long)pfn; in mlx5_ib_cont_pages()
87 base = pfn; in mlx5_ib_cont_pages()
90 if (base + p != pfn) { in mlx5_ib_cont_pages()
95 base = pfn; in mlx5_ib_cont_pages()
/linux-4.1.27/arch/hexagon/include/asm/
Dpage.h112 #define pfn_valid(pfn) ((pfn) < max_mapnr) argument
144 #define pfn_to_virt(pfn) __va((pfn) << PAGE_SHIFT) argument
/linux-4.1.27/arch/arm64/include/asm/
Dmemory.h87 #define __pfn_to_phys(pfn) ((phys_addr_t)(pfn) << PAGE_SHIFT) argument
156 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) argument
/linux-4.1.27/arch/arm/kernel/
Dhibernate.c27 int pfn_is_nosave(unsigned long pfn) in pfn_is_nosave() argument
32 return (pfn >= nosave_begin_pfn) && (pfn <= nosave_end_pfn); in pfn_is_nosave()
Dcrash_dump.c33 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument
42 vaddr = ioremap(__pfn_to_phys(pfn), PAGE_SIZE); in copy_oldmem_page()
/linux-4.1.27/arch/arm/mach-ep93xx/
Dts72xx.c35 .pfn = __phys_to_pfn(TS72XX_MODEL_PHYS_BASE),
40 .pfn = __phys_to_pfn(TS72XX_OPTIONS_PHYS_BASE),
45 .pfn = __phys_to_pfn(TS72XX_OPTIONS2_PHYS_BASE),
50 .pfn = __phys_to_pfn(TS72XX_RTC_INDEX_PHYS_BASE),
55 .pfn = __phys_to_pfn(TS72XX_RTC_DATA_PHYS_BASE),
/linux-4.1.27/drivers/char/
Dmem.c55 static inline int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) in valid_mmap_phys_addr_range() argument
62 static inline int range_is_allowed(unsigned long pfn, unsigned long size) in range_is_allowed() argument
64 u64 from = ((u64)pfn) << PAGE_SHIFT; in range_is_allowed()
69 if (!devmem_is_allowed(pfn)) { in range_is_allowed()
76 pfn++; in range_is_allowed()
81 static inline int range_is_allowed(unsigned long pfn, unsigned long size) in range_is_allowed() argument
224 unsigned long pfn, unsigned long size, pgprot_t *vma_prot) in phys_mem_access_prot_allowed() argument
265 static pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, in phys_mem_access_prot() argument
269 phys_addr_t offset = pfn << PAGE_SHIFT; in phys_mem_access_prot()
357 unsigned long pfn; in mmap_kmem() local
[all …]
/linux-4.1.27/drivers/net/ethernet/ibm/ehea/
Dehea_qmr.c628 static int ehea_update_busmap(unsigned long pfn, unsigned long nr_pages, int add) in ehea_update_busmap() argument
641 start_section = (pfn * PAGE_SIZE) / EHEA_SECTSIZE; in ehea_update_busmap()
671 int ehea_add_sect_bmap(unsigned long pfn, unsigned long nr_pages) in ehea_add_sect_bmap() argument
676 ret = ehea_update_busmap(pfn, nr_pages, EHEA_BUSMAP_ADD_SECT); in ehea_add_sect_bmap()
681 int ehea_rem_sect_bmap(unsigned long pfn, unsigned long nr_pages) in ehea_rem_sect_bmap() argument
686 ret = ehea_update_busmap(pfn, nr_pages, EHEA_BUSMAP_REM_SECT); in ehea_rem_sect_bmap()
691 static int ehea_is_hugepage(unsigned long pfn) in ehea_is_hugepage() argument
695 if (pfn & EHEA_HUGEPAGE_PFN_MASK) in ehea_is_hugepage()
698 page_order = compound_order(pfn_to_page(pfn)); in ehea_is_hugepage()
709 unsigned long pfn, start_pfn, end_pfn, nr_pages; in ehea_create_busmap_callback() local
[all …]
/linux-4.1.27/arch/arm/mach-sa1100/
Dnanoengine.c70 .pfn = __phys_to_pfn(0x10000000),
76 .pfn = __phys_to_pfn(NANO_PCI_MEM_RW_PHYS),
82 .pfn = __phys_to_pfn(NANO_PCI_CONFIG_SPACE_PHYS),
/linux-4.1.27/drivers/misc/
Dvmw_balloon.c318 static int vmballoon_send_lock_page(struct vmballoon *b, unsigned long pfn, in vmballoon_send_lock_page() argument
324 pfn32 = (u32)pfn; in vmballoon_send_lock_page()
325 if (pfn32 != pfn) in vmballoon_send_lock_page()
330 *hv_status = status = VMWARE_BALLOON_CMD(LOCK, pfn, dummy); in vmballoon_send_lock_page()
334 pr_debug("%s - ppn %lx, hv returns %ld\n", __func__, pfn, status); in vmballoon_send_lock_page()
343 static bool vmballoon_send_unlock_page(struct vmballoon *b, unsigned long pfn) in vmballoon_send_unlock_page() argument
348 pfn32 = (u32)pfn; in vmballoon_send_unlock_page()
349 if (pfn32 != pfn) in vmballoon_send_unlock_page()
354 status = VMWARE_BALLOON_CMD(UNLOCK, pfn, dummy); in vmballoon_send_unlock_page()
358 pr_debug("%s - ppn %lx, hv returns %ld\n", __func__, pfn, status); in vmballoon_send_unlock_page()
/linux-4.1.27/arch/mips/include/asm/mach-ip27/
Dkernel-entry-init.h51 dsrl t1, 12 # 4K pfn
52 dsrl t2, 12 # 4K pfn
53 dsll t1, 6 # Get pfn into place
54 dsll t2, 6 # Get pfn into place
/linux-4.1.27/lib/
Dioremap.c57 u64 pfn; in ioremap_pte_range() local
59 pfn = phys_addr >> PAGE_SHIFT; in ioremap_pte_range()
65 set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot)); in ioremap_pte_range()
66 pfn++; in ioremap_pte_range()
/linux-4.1.27/arch/mips/mm/
Dcache.c31 unsigned long pfn);
131 unsigned long pfn, addr; in __update_cache() local
134 pfn = pte_pfn(pte); in __update_cache()
135 if (unlikely(!pfn_valid(pfn))) in __update_cache()
137 page = pfn_to_page(pfn); in __update_cache()
Dioremap.c23 unsigned long pfn; in remap_area_pte() local
32 pfn = phys_addr >> PAGE_SHIFT; in remap_area_pte()
38 set_pte(pte, pfn_pte(pfn, pgprot)); in remap_area_pte()
40 pfn++; in remap_area_pte()
Dinit.c396 unsigned long pfn; in free_init_pages() local
398 for (pfn = PFN_UP(begin); pfn < PFN_DOWN(end); pfn++) { in free_init_pages()
399 struct page *page = pfn_to_page(pfn); in free_init_pages()
400 void *addr = phys_to_virt(PFN_PHYS(pfn)); in free_init_pages()
/linux-4.1.27/arch/arm/mach-socfpga/
Dsocfpga.c37 .pfn = 0, /* run-time */
44 .pfn = __phys_to_pfn(0xffc02000),
56 scu_io_desc.pfn = __phys_to_pfn(base); in socfpga_scu_map_io()
/linux-4.1.27/tools/testing/selftests/vm/
Dtranshuge-stress.c110 int64_t pfn; in main() local
112 pfn = allocate_transhuge(p); in main()
114 if (pfn < 0) { in main()
117 size_t idx = pfn >> (HPAGE_SHIFT - PAGE_SHIFT); in main()
/linux-4.1.27/arch/xtensa/mm/
Dinit.c238 unsigned long pfn; in bootmem_init() local
258 pfn = PAGE_ALIGN(sysmem.bank[i].start) >> PAGE_SHIFT; in bootmem_init()
259 if (pfn < min_low_pfn) in bootmem_init()
260 min_low_pfn = pfn; in bootmem_init()
261 pfn = PAGE_ALIGN(sysmem.bank[i].end - 1) >> PAGE_SHIFT; in bootmem_init()
262 if (pfn > max_pfn) in bootmem_init()
263 max_pfn = pfn; in bootmem_init()
Dcache.c200 unsigned long pfn) in local_flush_cache_page() argument
204 unsigned long phys = page_to_phys(pfn_to_page(pfn)); in local_flush_cache_page()
216 unsigned long pfn = pte_pfn(*ptep); in update_mmu_cache() local
219 if (!pfn_valid(pfn)) in update_mmu_cache()
222 page = pfn_to_page(pfn); in update_mmu_cache()
/linux-4.1.27/arch/arm/mach-lpc32xx/
Dcommon.c168 .pfn = __phys_to_pfn(LPC32XX_AHB0_START),
174 .pfn = __phys_to_pfn(LPC32XX_AHB1_START),
180 .pfn = __phys_to_pfn(LPC32XX_FABAPB_START),
186 .pfn = __phys_to_pfn(LPC32XX_IRAM_BASE),
/linux-4.1.27/Documentation/virtual/kvm/
Dlocking.txt31 1): The mapping from gfn to pfn
32 The mapping from gfn to pfn may be changed since we can only ensure the pfn
63 to pin gfn to pfn, because after gfn_to_pfn_atomic():
64 - We have held the refcount of pfn that means the pfn can not be freed and
66 - The pfn is writable that means it can not be shared between different gfns
106 kvm_set_pfn_accessed(spte.pfn);
108 kvm_set_pfn_dirty(spte.pfn);
/linux-4.1.27/drivers/gpu/drm/gma500/
Dgem.c179 unsigned long pfn; in psb_gem_fault() local
212 pfn = (dev_priv->stolen_base + r->offset) >> PAGE_SHIFT; in psb_gem_fault()
214 pfn = page_to_pfn(r->pages[page_offset]); in psb_gem_fault()
215 ret = vm_insert_pfn(vma, (unsigned long)vmf->virtual_address, pfn); in psb_gem_fault()
/linux-4.1.27/arch/nios2/mm/
Dioremap.c27 unsigned long pfn; in remap_area_pte() local
37 pfn = PFN_DOWN(phys_addr); in remap_area_pte()
43 set_pte(pte, pfn_pte(pfn, pgprot)); in remap_area_pte()
45 pfn++; in remap_area_pte()
/linux-4.1.27/drivers/edac/
Dcell_edac.c38 unsigned long address, pfn, offset, syndrome; in cell_edac_count_ce() local
47 pfn = address >> PAGE_SHIFT; in cell_edac_count_ce()
53 csrow->first_page + pfn, offset, syndrome, in cell_edac_count_ce()
61 unsigned long address, pfn, offset; in cell_edac_count_ue() local
70 pfn = address >> PAGE_SHIFT; in cell_edac_count_ue()
75 csrow->first_page + pfn, offset, 0, in cell_edac_count_ue()
/linux-4.1.27/arch/ia64/kernel/
Dcrash_dump.c34 copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument
41 vaddr = __va(pfn<<PAGE_SHIFT); in copy_oldmem_page()
/linux-4.1.27/drivers/scsi/csiostor/
Dcsio_mb.c281 (FW_LDST_CMD_LC_F | FW_LDST_CMD_FN_V(hw->pfn)); in csio_mb_ldst()
449 FW_IQ_CMD_PFN_V(iq_params->pfn) | in csio_mb_iq_alloc()
505 FW_IQ_CMD_PFN_V(iq_params->pfn) | in csio_mb_iq_write()
628 FW_IQ_CMD_PFN_V(iq_params->pfn) | in csio_mb_iq_free()
663 FW_EQ_OFLD_CMD_PFN_V(eq_ofld_params->pfn) | in csio_mb_eq_ofld_alloc()
711 FW_EQ_OFLD_CMD_PFN_V(eq_ofld_params->pfn) | in csio_mb_eq_ofld_write()
814 FW_EQ_OFLD_CMD_PFN_V(eq_ofld_params->pfn) | in csio_mb_eq_ofld_free()
1158 uint32_t ctl_reg = PF_REG(hw->pfn, CIM_PF_MAILBOX_CTRL_A); in csio_mb_debug_cmd_handler()
1159 uint32_t data_reg = PF_REG(hw->pfn, CIM_PF_MAILBOX_DATA_A); in csio_mb_debug_cmd_handler()
1192 uint32_t ctl_reg = PF_REG(hw->pfn, CIM_PF_MAILBOX_CTRL_A); in csio_mb_issue()
[all …]
/linux-4.1.27/virt/kvm/
Dkvm_main.c105 static void kvm_release_pfn_dirty(pfn_t pfn);
114 bool kvm_is_reserved_pfn(pfn_t pfn) in kvm_is_reserved_pfn() argument
116 if (pfn_valid(pfn)) in kvm_is_reserved_pfn()
117 return PageReserved(pfn_to_page(pfn)); in kvm_is_reserved_pfn()
1217 bool write_fault, bool *writable, pfn_t *pfn) in hva_to_pfn_fast() argument
1235 *pfn = page_to_pfn(page[0]); in hva_to_pfn_fast()
1250 bool *writable, pfn_t *pfn) in hva_to_pfn_slow() argument
1285 *pfn = page_to_pfn(page[0]); in hva_to_pfn_slow()
1318 pfn_t pfn = 0; in hva_to_pfn() local
1324 if (hva_to_pfn_fast(addr, atomic, async, write_fault, writable, &pfn)) in hva_to_pfn()
[all …]
/linux-4.1.27/arch/arm/mach-mmp/
Dcommon.c30 .pfn = __phys_to_pfn(APB_PHYS_BASE),
35 .pfn = __phys_to_pfn(AXI_PHYS_BASE),
/linux-4.1.27/arch/microblaze/kernel/
Ddma.c165 unsigned long pfn; in dma_direct_mmap_coherent() local
172 pfn = consistent_virt_to_pfn(cpu_addr); in dma_direct_mmap_coherent()
174 pfn = virt_to_pfn(cpu_addr); in dma_direct_mmap_coherent()
176 return remap_pfn_range(vma, vma->vm_start, pfn + off, in dma_direct_mmap_coherent()
/linux-4.1.27/arch/arm/mach-ux500/
Dsetup.h31 .pfn = __phys_to_pfn(x), \
38 .pfn = __phys_to_pfn(x), \
/linux-4.1.27/arch/mips/kernel/
Dcrash_dump.c26 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, in copy_oldmem_page() argument
34 vaddr = kmap_atomic_pfn(pfn); in copy_oldmem_page()
/linux-4.1.27/tools/vm/
Dpage-types.c292 unsigned long pfn; in pagemap_pfn() local
295 pfn = PM_PFRAME(val); in pagemap_pfn()
297 pfn = 0; in pagemap_pfn()
299 return pfn; in pagemap_pfn()
620 unsigned long pfn; in walk_vma() local
630 pfn = pagemap_pfn(buf[i]); in walk_vma()
631 if (pfn) in walk_vma()
632 walk_pfn(index + i, pfn, 1, buf[i]); in walk_vma()
852 unsigned long nr_pages, pfn, i; in walk_file() local
904 pfn = pagemap_pfn(buf[i]); in walk_file()
[all …]
/linux-4.1.27/arch/mn10300/mm/
Dpgtable.c35 void set_pmd_pfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags) in set_pmd_pfn() argument
45 if (pfn & (PTRS_PER_PTE-1)) { /* pfn is misaligned */ in set_pmd_pfn()
56 set_pmd(pmd, pfn_pmd(pfn, flags)); in set_pmd_pfn()
/linux-4.1.27/drivers/gpu/drm/ttm/
Dttm_bo_vm.c93 unsigned long pfn; in ttm_bo_vm_fault() local
216 pfn = ((bo->mem.bus.base + bo->mem.bus.offset) >> PAGE_SHIFT) + page_offset; in ttm_bo_vm_fault()
228 pfn = page_to_pfn(page); in ttm_bo_vm_fault()
232 ret = vm_insert_mixed(&cvma, address, pfn); in ttm_bo_vm_fault()
234 ret = vm_insert_pfn(&cvma, address, pfn); in ttm_bo_vm_fault()
/linux-4.1.27/arch/arm/mach-versatile/
Dcore.c134 .pfn = __phys_to_pfn(VERSATILE_SYS_BASE),
139 .pfn = __phys_to_pfn(VERSATILE_SIC_BASE),
144 .pfn = __phys_to_pfn(VERSATILE_VIC_BASE),
149 .pfn = __phys_to_pfn(VERSATILE_SCTL_BASE),
156 .pfn = __phys_to_pfn(VERSATILE_IB2_BASE),
164 .pfn = __phys_to_pfn(VERSATILE_UART0_BASE),
172 .pfn = __phys_to_pfn(VERSATILE_PCI_CORE_BASE),
177 .pfn = __phys_to_pfn(VERSATILE_PCI_BASE),
182 .pfn = __phys_to_pfn(VERSATILE_PCI_CFG_BASE),
/linux-4.1.27/drivers/iommu/
Diova.c291 struct iova *find_iova(struct iova_domain *iovad, unsigned long pfn) in find_iova() argument
303 if ((pfn >= iova->pfn_lo) && (pfn <= iova->pfn_hi)) { in find_iova()
314 if (pfn < iova->pfn_lo) in find_iova()
316 else if (pfn > iova->pfn_lo) in find_iova()
350 free_iova(struct iova_domain *iovad, unsigned long pfn) in free_iova() argument
352 struct iova *iova = find_iova(iovad, pfn); in free_iova()

123