Home
last modified time | relevance | path

Searched refs:nr_pages (Results 1 – 200 of 319) sorted by relevance

12

/linux-4.4.14/mm/
Dpage_counter.c20 void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages) in page_counter_cancel() argument
24 new = atomic_long_sub_return(nr_pages, &counter->count); in page_counter_cancel()
36 void page_counter_charge(struct page_counter *counter, unsigned long nr_pages) in page_counter_charge() argument
43 new = atomic_long_add_return(nr_pages, &c->count); in page_counter_charge()
63 unsigned long nr_pages, in page_counter_try_charge() argument
84 new = atomic_long_add_return(nr_pages, &c->count); in page_counter_try_charge()
86 atomic_long_sub(nr_pages, &c->count); in page_counter_try_charge()
106 page_counter_cancel(c, nr_pages); in page_counter_try_charge()
116 void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages) in page_counter_uncharge() argument
121 page_counter_cancel(c, nr_pages); in page_counter_uncharge()
[all …]
Dpercpu-km.c50 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; in pcpu_create_chunk() local
59 pages = alloc_pages(GFP_KERNEL, order_base_2(nr_pages)); in pcpu_create_chunk()
65 for (i = 0; i < nr_pages; i++) in pcpu_create_chunk()
72 pcpu_chunk_populated(chunk, 0, nr_pages); in pcpu_create_chunk()
80 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; in pcpu_destroy_chunk() local
83 __free_pages(chunk->data, order_base_2(nr_pages)); in pcpu_destroy_chunk()
94 size_t nr_pages, alloc_pages; in pcpu_verify_alloc_info() local
102 nr_pages = (ai->groups[0].nr_units * ai->unit_size) >> PAGE_SHIFT; in pcpu_verify_alloc_info()
103 alloc_pages = roundup_pow_of_two(nr_pages); in pcpu_verify_alloc_info()
105 if (alloc_pages > nr_pages) in pcpu_verify_alloc_info()
[all …]
Dmemory_hotplug.c247 unsigned long i, pfn, end_pfn, nr_pages; in register_page_bootmem_info_node() local
252 nr_pages = PAGE_ALIGN(sizeof(struct pglist_data)) >> PAGE_SHIFT; in register_page_bootmem_info_node()
255 for (i = 0; i < nr_pages; i++, page++) in register_page_bootmem_info_node()
261 nr_pages = zone->wait_table_hash_nr_entries in register_page_bootmem_info_node()
263 nr_pages = PAGE_ALIGN(nr_pages) >> PAGE_SHIFT; in register_page_bootmem_info_node()
266 for (i = 0; i < nr_pages; i++, page++) in register_page_bootmem_info_node()
446 int nr_pages = PAGES_PER_SECTION; in __add_zone() local
453 ret = ensure_zone_is_initialized(zone, phys_start_pfn, nr_pages); in __add_zone()
458 grow_zone_span(zone, phys_start_pfn, phys_start_pfn + nr_pages); in __add_zone()
460 phys_start_pfn + nr_pages); in __add_zone()
[all …]
Dpage_ext.c128 unsigned long nr_pages; in alloc_node_page_ext() local
130 nr_pages = NODE_DATA(nid)->node_spanned_pages; in alloc_node_page_ext()
131 if (!nr_pages) in alloc_node_page_ext()
141 nr_pages += MAX_ORDER_NR_PAGES; in alloc_node_page_ext()
143 table_size = sizeof(struct page_ext) * nr_pages; in alloc_node_page_ext()
280 unsigned long nr_pages, in online_page_ext() argument
287 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in online_page_ext()
315 unsigned long nr_pages, int nid) in offline_page_ext() argument
320 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in offline_page_ext()
337 mn->nr_pages, mn->status_change_nid); in page_ext_callback()
[all …]
Dhugetlb_cgroup.c112 unsigned int nr_pages; in hugetlb_cgroup_move_parent() local
126 nr_pages = 1 << compound_order(page); in hugetlb_cgroup_move_parent()
130 page_counter_charge(&parent->hugepage[idx], nr_pages); in hugetlb_cgroup_move_parent()
134 page_counter_cancel(counter, nr_pages); in hugetlb_cgroup_move_parent()
165 int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, in hugetlb_cgroup_charge_cgroup() argument
189 if (!page_counter_try_charge(&h_cg->hugepage[idx], nr_pages, &counter)) in hugetlb_cgroup_charge_cgroup()
198 void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, in hugetlb_cgroup_commit_charge() argument
212 void hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, in hugetlb_cgroup_uncharge_page() argument
224 page_counter_uncharge(&h_cg->hugepage[idx], nr_pages); in hugetlb_cgroup_uncharge_page()
228 void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, in hugetlb_cgroup_uncharge_cgroup() argument
[all …]
Dpage_isolation.c27 arg.nr_pages = pageblock_nr_pages; in set_migratetype_isolate()
60 unsigned long nr_pages; in set_migratetype_isolate() local
65 nr_pages = move_freepages_block(zone, page, MIGRATE_ISOLATE); in set_migratetype_isolate()
67 __mod_zone_freepage_state(zone, -nr_pages, migratetype); in set_migratetype_isolate()
79 unsigned long flags, nr_pages; in unset_migratetype_isolate() local
121 nr_pages = move_freepages_block(zone, page, migratetype); in unset_migratetype_isolate()
122 __mod_zone_freepage_state(zone, nr_pages, migratetype); in unset_migratetype_isolate()
133 __first_valid_page(unsigned long pfn, unsigned long nr_pages) in __first_valid_page() argument
136 for (i = 0; i < nr_pages; i++) in __first_valid_page()
139 if (unlikely(i == nr_pages)) in __first_valid_page()
Dgup.c454 unsigned long start, unsigned long nr_pages, in __get_user_pages() argument
462 if (!nr_pages) in __get_user_pages()
498 &start, &nr_pages, i, in __get_user_pages()
550 if (page_increm > nr_pages) in __get_user_pages()
551 page_increm = nr_pages; in __get_user_pages()
554 nr_pages -= page_increm; in __get_user_pages()
555 } while (nr_pages); in __get_user_pages()
624 unsigned long nr_pages, in __get_user_pages_locked() argument
651 ret = __get_user_pages(tsk, mm, start, nr_pages, flags, pages, in __get_user_pages_locked()
660 BUG_ON(ret >= nr_pages); in __get_user_pages_locked()
[all …]
Dprocess_vm_access.c87 unsigned long nr_pages; in process_vm_rw_single_vec() local
95 nr_pages = (addr + len - 1) / PAGE_SIZE - addr / PAGE_SIZE + 1; in process_vm_rw_single_vec()
97 while (!rc && nr_pages && iov_iter_count(iter)) { in process_vm_rw_single_vec()
98 int pages = min(nr_pages, max_pages_per_loop); in process_vm_rw_single_vec()
116 nr_pages -= pages; in process_vm_rw_single_vec()
152 unsigned long nr_pages = 0; in process_vm_rw_core() local
168 nr_pages = max(nr_pages, nr_pages_iov); in process_vm_rw_core()
172 if (nr_pages == 0) in process_vm_rw_core()
175 if (nr_pages > PVM_MAX_PP_ARRAY_COUNT) { in process_vm_rw_core()
179 sizeof(struct pages *)*nr_pages), in process_vm_rw_core()
Dmemcontrol.c551 unsigned long nr_pages = page_counter_read(&memcg->memory); in soft_limit_excess() local
555 if (nr_pages > soft_limit) in soft_limit_excess()
556 excess = nr_pages - soft_limit; in soft_limit_excess()
700 int nr_pages) in mem_cgroup_charge_statistics() argument
708 nr_pages); in mem_cgroup_charge_statistics()
711 nr_pages); in mem_cgroup_charge_statistics()
715 nr_pages); in mem_cgroup_charge_statistics()
718 if (nr_pages > 0) in mem_cgroup_charge_statistics()
722 nr_pages = -nr_pages; /* for event */ in mem_cgroup_charge_statistics()
725 __this_cpu_add(memcg->stat->nr_page_events, nr_pages); in mem_cgroup_charge_statistics()
[all …]
Dmigrate.c479 int nr_pages) in __copy_gigantic_page() argument
485 for (i = 0; i < nr_pages; ) { in __copy_gigantic_page()
498 int nr_pages; in copy_huge_page() local
503 nr_pages = pages_per_huge_page(h); in copy_huge_page()
505 if (unlikely(nr_pages > MAX_ORDER_NR_PAGES)) { in copy_huge_page()
506 __copy_gigantic_page(dst, src, nr_pages); in copy_huge_page()
512 nr_pages = hpage_nr_pages(src); in copy_huge_page()
515 for (i = 0; i < nr_pages; i++) { in copy_huge_page()
1321 unsigned long nr_pages, in do_pages_move() argument
1345 chunk_start < nr_pages; in do_pages_move()
[all …]
Dquicklist.c55 pages_to_free = q->nr_pages - max_pages(min_pages); in min_pages_to_free()
70 if (q->nr_pages > min_pages) { in quicklist_trim()
98 count += q->nr_pages; in quicklist_total_size()
Dmlock.c175 int nr_pages; in munlock_vma_page() local
188 nr_pages = hpage_nr_pages(page); in munlock_vma_page()
192 __mod_zone_page_state(zone, NR_MLOCK, -nr_pages); in munlock_vma_page()
205 return nr_pages - 1; in munlock_vma_page()
503 int nr_pages; in mlock_fixup() local
537 nr_pages = (end - start) >> PAGE_SHIFT; in mlock_fixup()
539 nr_pages = -nr_pages; in mlock_fixup()
540 mm->locked_vm += nr_pages; in mlock_fixup()
Dmemory-failure.c1037 int nr_pages = 1 << compound_order(hpage); in set_page_hwpoison_huge_page() local
1038 for (i = 0; i < nr_pages; i++) in set_page_hwpoison_huge_page()
1045 int nr_pages = 1 << compound_order(hpage); in clear_page_hwpoison_huge_page() local
1046 for (i = 0; i < nr_pages; i++) in clear_page_hwpoison_huge_page()
1075 unsigned int nr_pages; in memory_failure() local
1103 nr_pages = 1 << compound_order(hpage); in memory_failure()
1105 nr_pages = 1; in memory_failure()
1106 num_poisoned_pages_add(nr_pages); in memory_failure()
1134 num_poisoned_pages_sub(nr_pages); in memory_failure()
1158 num_poisoned_pages_sub(nr_pages); in memory_failure()
[all …]
Dpercpu-vm.c133 static void __pcpu_unmap_pages(unsigned long addr, int nr_pages) in __pcpu_unmap_pages() argument
135 unmap_kernel_range_noflush(addr, nr_pages << PAGE_SHIFT); in __pcpu_unmap_pages()
192 int nr_pages) in __pcpu_map_pages() argument
194 return map_kernel_range_noflush(addr, nr_pages << PAGE_SHIFT, in __pcpu_map_pages()
Dsparse.c198 unsigned long nr_pages = 0; in node_memmap_size_bytes() local
206 nr_pages += PAGES_PER_SECTION; in node_memmap_size_bytes()
209 return nr_pages * sizeof(struct page); in node_memmap_size_bytes()
659 unsigned long magic, nr_pages; in free_map_bootmem() local
662 nr_pages = PAGE_ALIGN(PAGES_PER_SECTION * sizeof(struct page)) in free_map_bootmem()
665 for (i = 0; i < nr_pages; i++, page++) { in free_map_bootmem()
744 static void clear_hwpoisoned_pages(struct page *memmap, int nr_pages) in clear_hwpoisoned_pages() argument
759 static inline void clear_hwpoisoned_pages(struct page *memmap, int nr_pages) in clear_hwpoisoned_pages() argument
Dinternal.h301 int nr_pages = hpage_nr_pages(page); in mlock_migrate_page() local
304 __mod_zone_page_state(page_zone(page), NR_MLOCK, -nr_pages); in mlock_migrate_page()
306 __mod_zone_page_state(page_zone(newpage), NR_MLOCK, nr_pages); in mlock_migrate_page()
Dswapfile.c134 nr_blocks = ((sector_t)se->nr_pages - 1) << (PAGE_SHIFT - 9); in discard_swap()
145 nr_blocks = (sector_t)se->nr_pages << (PAGE_SHIFT - 9); in discard_swap()
162 pgoff_t start_page, pgoff_t nr_pages) in discard_swap_cluster() argument
167 while (nr_pages) { in discard_swap_cluster()
171 start_page < se->start_page + se->nr_pages) { in discard_swap_cluster()
174 sector_t nr_blocks = se->nr_pages - offset; in discard_swap_cluster()
176 if (nr_blocks > nr_pages) in discard_swap_cluster()
177 nr_blocks = nr_pages; in discard_swap_cluster()
179 nr_pages -= nr_blocks; in discard_swap_cluster()
1639 offset < (se->start_page + se->nr_pages)) { in map_swap_entry()
[all …]
Dswap.c1080 pgoff_t start, unsigned nr_pages, in pagevec_lookup_entries() argument
1083 pvec->nr = find_get_entries(mapping, start, nr_pages, in pagevec_lookup_entries()
1126 pgoff_t start, unsigned nr_pages) in pagevec_lookup() argument
1128 pvec->nr = find_get_pages(mapping, start, nr_pages, pvec->pages); in pagevec_lookup()
1134 pgoff_t *index, int tag, unsigned nr_pages) in pagevec_lookup_tag() argument
1137 nr_pages, pvec->pages); in pagevec_lookup_tag()
Dpage-writeback.c278 unsigned long nr_pages; in zone_dirtyable_memory() local
280 nr_pages = zone_page_state(zone, NR_FREE_PAGES); in zone_dirtyable_memory()
281 nr_pages -= min(nr_pages, zone->dirty_balance_reserve); in zone_dirtyable_memory()
283 nr_pages += zone_page_state(zone, NR_INACTIVE_FILE); in zone_dirtyable_memory()
284 nr_pages += zone_page_state(zone, NR_ACTIVE_FILE); in zone_dirtyable_memory()
286 return nr_pages; in zone_dirtyable_memory()
1969 int nr_pages = global_page_state(NR_FILE_DIRTY) + in laptop_mode_timer_fn() local
1983 wb_start_writeback(wb, nr_pages, true, in laptop_mode_timer_fn()
2151 int nr_pages; in write_cache_pages() local
2187 nr_pages = pagevec_lookup_tag(&pvec, mapping, &index, tag, in write_cache_pages()
[all …]
Dutil.c247 int nr_pages, int write, struct page **pages) in __get_user_pages_fast() argument
278 int nr_pages, int write, struct page **pages) in get_user_pages_fast() argument
281 return get_user_pages_unlocked(current, mm, start, nr_pages, in get_user_pages_fast()
Dvmscan.c1368 int nr_pages; in isolate_lru_pages() local
1377 nr_pages = hpage_nr_pages(page); in isolate_lru_pages()
1378 mem_cgroup_update_lru_size(lruvec, lru, -nr_pages); in isolate_lru_pages()
1380 nr_taken += nr_pages; in isolate_lru_pages()
1728 int nr_pages; in move_active_pages_to_lru() local
1737 nr_pages = hpage_nr_pages(page); in move_active_pages_to_lru()
1738 mem_cgroup_update_lru_size(lruvec, lru, nr_pages); in move_active_pages_to_lru()
1740 pgmoved += nr_pages; in move_active_pages_to_lru()
2901 unsigned long nr_pages, in try_to_free_mem_cgroup_pages() argument
2909 .nr_to_reclaim = max(nr_pages, SWAP_CLUSTER_MAX), in try_to_free_mem_cgroup_pages()
[all …]
Dhugetlb.c1010 int nr_pages = 1 << order; in destroy_compound_gigantic_page() local
1013 for (i = 1; i < nr_pages; i++, p = mem_map_next(p, page, i)) { in destroy_compound_gigantic_page()
1028 unsigned long nr_pages) in __alloc_gigantic_page() argument
1030 unsigned long end_pfn = start_pfn + nr_pages; in __alloc_gigantic_page()
1035 unsigned long nr_pages) in pfn_range_valid_gigantic() argument
1037 unsigned long i, end_pfn = start_pfn + nr_pages; in pfn_range_valid_gigantic()
1060 unsigned long start_pfn, unsigned long nr_pages) in zone_spans_last_pfn() argument
1062 unsigned long last_pfn = start_pfn + nr_pages - 1; in zone_spans_last_pfn()
1068 unsigned long nr_pages = 1 << order; in alloc_gigantic_page() local
1076 pfn = ALIGN(z->zone_start_pfn, nr_pages); in alloc_gigantic_page()
[all …]
Dreadahead.c112 struct list_head *pages, unsigned nr_pages) in read_pages() argument
121 ret = mapping->a_ops->readpages(filp, mapping, pages, nr_pages); in read_pages()
127 for (page_idx = 0; page_idx < nr_pages; page_idx++) { in read_pages()
Dfilemap.c340 int nr_pages; in __filemap_fdatawait_range() local
348 (nr_pages = pagevec_lookup_tag(&pvec, mapping, &index, in __filemap_fdatawait_range()
353 for (i = 0; i < nr_pages; i++) { in __filemap_fdatawait_range()
1286 unsigned int nr_pages, struct page **pages) in find_get_pages() argument
1292 if (unlikely(!nr_pages)) in find_get_pages()
1332 if (++ret == nr_pages) in find_get_pages()
1353 unsigned int nr_pages, struct page **pages) in find_get_pages_contig() argument
1359 if (unlikely(!nr_pages)) in find_get_pages_contig()
1409 if (++ret == nr_pages) in find_get_pages_contig()
1429 int tag, unsigned int nr_pages, struct page **pages) in find_get_pages_tag() argument
[all …]
Dnommu.c136 unsigned long start, unsigned long nr_pages, in __get_user_pages() argument
152 for (i = 0; i < nr_pages; i++) { in __get_user_pages()
186 unsigned long start, unsigned long nr_pages, in get_user_pages() argument
197 return __get_user_pages(tsk, mm, start, nr_pages, flags, pages, vmas, in get_user_pages()
203 unsigned long start, unsigned long nr_pages, in get_user_pages_locked() argument
207 return get_user_pages(tsk, mm, start, nr_pages, write, force, in get_user_pages_locked()
213 unsigned long start, unsigned long nr_pages, in __get_user_pages_unlocked() argument
219 ret = get_user_pages(tsk, mm, start, nr_pages, write, force, in __get_user_pages_unlocked()
227 unsigned long start, unsigned long nr_pages, in get_user_pages_unlocked() argument
230 return __get_user_pages_unlocked(tsk, mm, start, nr_pages, write, in get_user_pages_unlocked()
Dvmalloc.c1476 for (i = 0; i < area->nr_pages; i++) { in __vunmap()
1584 unsigned int nr_pages, array_size, i; in __vmalloc_area_node() local
1588 nr_pages = get_vm_area_size(area) >> PAGE_SHIFT; in __vmalloc_area_node()
1589 array_size = (nr_pages * sizeof(struct page *)); in __vmalloc_area_node()
1591 area->nr_pages = nr_pages; in __vmalloc_area_node()
1607 for (i = 0; i < area->nr_pages; i++) { in __vmalloc_area_node()
1617 area->nr_pages = i; in __vmalloc_area_node()
1632 (area->nr_pages*PAGE_SIZE), area->size); in __vmalloc_area_node()
2607 for (nr = 0; nr < v->nr_pages; nr++) in show_numa_info()
2636 if (v->nr_pages) in s_show()
[all …]
Dshmem.c1648 unsigned int loff, nr_pages, req_pages; in shmem_file_splice_read() local
1678 nr_pages = min(req_pages, spd.nr_pages_max); in shmem_file_splice_read()
1680 spd.nr_pages = find_get_pages_contig(mapping, index, in shmem_file_splice_read()
1681 nr_pages, spd.pages); in shmem_file_splice_read()
1682 index += spd.nr_pages; in shmem_file_splice_read()
1685 while (spd.nr_pages < nr_pages) { in shmem_file_splice_read()
1690 spd.pages[spd.nr_pages++] = page; in shmem_file_splice_read()
1695 nr_pages = spd.nr_pages; in shmem_file_splice_read()
1696 spd.nr_pages = 0; in shmem_file_splice_read()
1698 for (page_nr = 0; page_nr < nr_pages; page_nr++) { in shmem_file_splice_read()
[all …]
/linux-4.4.14/drivers/media/v4l2-core/
Dvideobuf-dma-sg.c66 int nr_pages) in videobuf_vmalloc_to_sg() argument
72 sglist = vzalloc(nr_pages * sizeof(*sglist)); in videobuf_vmalloc_to_sg()
75 sg_init_table(sglist, nr_pages); in videobuf_vmalloc_to_sg()
76 for (i = 0; i < nr_pages; i++, virt += PAGE_SIZE) { in videobuf_vmalloc_to_sg()
96 int nr_pages, int offset, size_t size) in videobuf_pages_to_sg() argument
103 sglist = vmalloc(nr_pages * sizeof(*sglist)); in videobuf_pages_to_sg()
106 sg_init_table(sglist, nr_pages); in videobuf_pages_to_sg()
114 for (i = 1; i < nr_pages; i++) { in videobuf_pages_to_sg()
176 dma->nr_pages = last-first+1; in videobuf_dma_init_user_locked()
177 dma->pages = kmalloc(dma->nr_pages * sizeof(struct page *), GFP_KERNEL); in videobuf_dma_init_user_locked()
[all …]
/linux-4.4.14/include/linux/
Dhugetlb_cgroup.h54 extern int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages,
56 extern void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages,
59 extern void hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages,
61 extern void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages,
85 hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, in hugetlb_cgroup_charge_cgroup() argument
92 hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, in hugetlb_cgroup_commit_charge() argument
100 hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, struct page *page) in hugetlb_cgroup_uncharge_page() argument
106 hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, in hugetlb_cgroup_uncharge_cgroup() argument
Dmm_inline.h28 int nr_pages = hpage_nr_pages(page); in add_page_to_lru_list() local
29 mem_cgroup_update_lru_size(lruvec, lru, nr_pages); in add_page_to_lru_list()
31 __mod_zone_page_state(lruvec_zone(lruvec), NR_LRU_BASE + lru, nr_pages); in add_page_to_lru_list()
37 int nr_pages = hpage_nr_pages(page); in del_page_from_lru_list() local
38 mem_cgroup_update_lru_size(lruvec, lru, -nr_pages); in del_page_from_lru_list()
40 __mod_zone_page_state(lruvec_zone(lruvec), NR_LRU_BASE + lru, -nr_pages); in del_page_from_lru_list()
Dpage_counter.h37 void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages);
38 void page_counter_charge(struct page_counter *counter, unsigned long nr_pages);
40 unsigned long nr_pages,
42 void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages);
45 unsigned long *nr_pages);
Dmemory_hotplug.h84 extern int zone_grow_waitqueues(struct zone *zone, unsigned long nr_pages);
106 unsigned long nr_pages);
111 unsigned long nr_pages);
245 extern int is_mem_section_removable(unsigned long pfn, unsigned long nr_pages);
247 extern int offline_pages(unsigned long start_pfn, unsigned long nr_pages);
252 unsigned long nr_pages) in is_mem_section_removable() argument
259 static inline int offline_pages(unsigned long start_pfn, unsigned long nr_pages) in offline_pages() argument
274 extern int offline_pages(unsigned long start_pfn, unsigned long nr_pages);
Dquicklist.h18 int nr_pages; member
43 q->nr_pages--; in quicklist_alloc()
63 q->nr_pages++; in __quicklist_free()
Dpagevec.h31 pgoff_t start, unsigned nr_pages);
34 unsigned nr_pages);
Dswap.h131 pgoff_t nr_pages; member
323 unsigned long nr_pages,
330 extern unsigned long shrink_all_memory(unsigned long nr_pages);
349 extern void check_move_unevictable_pages(struct page **, int nr_pages);
390 unsigned long nr_pages, sector_t start_block);
Dmemory.h50 unsigned long nr_pages; member
65 unsigned int nr_pages; /* # pages in range to check */ member
Dvmstat.h257 static inline void __mod_zone_freepage_state(struct zone *zone, int nr_pages, in __mod_zone_freepage_state() argument
260 __mod_zone_page_state(zone, NR_FREE_PAGES, nr_pages); in __mod_zone_freepage_state()
262 __mod_zone_page_state(zone, NR_FREE_CMA_PAGES, nr_pages); in __mod_zone_freepage_state()
Dmpage.h17 unsigned nr_pages, get_block_t get_block);
Dsplice.h56 int nr_pages; /* number of populated pages in map */ member
Dpagemap.h359 unsigned int nr_pages, struct page **pages);
361 unsigned int nr_pages, struct page **pages);
363 int tag, unsigned int nr_pages, struct page **pages);
Dfscache.h594 unsigned *nr_pages, in fscache_read_or_alloc_pages() argument
601 nr_pages, end_io_func, in fscache_read_or_alloc_pages()
Dmm.h1193 unsigned long start, unsigned long nr_pages,
1197 unsigned long start, unsigned long nr_pages,
1201 unsigned long start, unsigned long nr_pages,
1205 unsigned long start, unsigned long nr_pages,
1209 unsigned long start, unsigned long nr_pages,
1211 int get_user_pages_fast(unsigned long start, int nr_pages, int write,
1257 int get_kernel_pages(const struct kvec *iov, int nr_pages, int write,
1331 int __get_user_pages_fast(unsigned long start, int nr_pages, int write,
/linux-4.4.14/drivers/xen/
Dballoon.c412 static enum bp_state increase_reservation(unsigned long nr_pages) in increase_reservation() argument
423 if (nr_pages > ARRAY_SIZE(frame_list)) in increase_reservation()
424 nr_pages = ARRAY_SIZE(frame_list); in increase_reservation()
427 for (i = 0; i < nr_pages; i++) { in increase_reservation()
429 nr_pages = i; in increase_reservation()
441 reservation.nr_extents = nr_pages; in increase_reservation()
483 static enum bp_state decrease_reservation(unsigned long nr_pages, gfp_t gfp) in decrease_reservation() argument
496 if (nr_pages > ARRAY_SIZE(frame_list)) in decrease_reservation()
497 nr_pages = ARRAY_SIZE(frame_list); in decrease_reservation()
499 for (i = 0; i < nr_pages; i++) { in decrease_reservation()
[all …]
Dprivcmd.c49 unsigned long nr_pages);
425 unsigned long nr_pages; in privcmd_ioctl_mmap_batch() local
449 nr_pages = DIV_ROUND_UP(m.num, XEN_PFN_PER_PAGE); in privcmd_ioctl_mmap_batch()
450 if ((m.num <= 0) || (nr_pages > (LONG_MAX >> PAGE_SHIFT))) in privcmd_ioctl_mmap_batch()
492 m.addr + (nr_pages << PAGE_SHIFT) != vma->vm_end) { in privcmd_ioctl_mmap_batch()
497 ret = alloc_empty_pages(vma, nr_pages); in privcmd_ioctl_mmap_batch()
504 m.addr + (nr_pages << PAGE_SHIFT) > vma->vm_end) { in privcmd_ioctl_mmap_batch()
508 if (privcmd_vma_range_is_mapped(vma, m.addr, nr_pages)) { in privcmd_ioctl_mmap_batch()
641 unsigned long nr_pages) in privcmd_vma_range_is_mapped() argument
643 return apply_to_page_range(vma->vm_mm, addr, nr_pages << PAGE_SHIFT, in privcmd_vma_range_is_mapped()
Dgrant-table.c685 int gnttab_alloc_pages(int nr_pages, struct page **pages) in gnttab_alloc_pages() argument
690 ret = alloc_xenballooned_pages(nr_pages, pages); in gnttab_alloc_pages()
694 for (i = 0; i < nr_pages; i++) { in gnttab_alloc_pages()
700 gnttab_free_pages(nr_pages, pages); in gnttab_alloc_pages()
717 void gnttab_free_pages(int nr_pages, struct page **pages) in gnttab_free_pages() argument
721 for (i = 0; i < nr_pages; i++) { in gnttab_free_pages()
729 free_xenballooned_pages(nr_pages, pages); in gnttab_free_pages()
Dswiotlb-xen.c115 int nr_pages; in check_pages_physically_contiguous() local
118 nr_pages = (offset + length + XEN_PAGE_SIZE-1) >> XEN_PAGE_SHIFT; in check_pages_physically_contiguous()
120 for (i = 1; i < nr_pages; i++) { in check_pages_physically_contiguous()
/linux-4.4.14/drivers/gpu/drm/radeon/
Ddrm_buffer.c46 int nr_pages = size / PAGE_SIZE + 1; in drm_buffer_alloc() local
51 *buf = kzalloc(sizeof(struct drm_buffer) + nr_pages*sizeof(char *), in drm_buffer_alloc()
57 size, nr_pages); in drm_buffer_alloc()
63 for (idx = 0; idx < nr_pages; ++idx) { in drm_buffer_alloc()
73 idx + 1, size, nr_pages); in drm_buffer_alloc()
100 int nr_pages = size / PAGE_SIZE + 1; in drm_buffer_copy_from_user() local
110 for (idx = 0; idx < nr_pages; ++idx) { in drm_buffer_copy_from_user()
134 int nr_pages = buf->size / PAGE_SIZE + 1; in drm_buffer_free() local
136 for (idx = 0; idx < nr_pages; ++idx) in drm_buffer_free()
/linux-4.4.14/drivers/misc/mic/scif/
Dscif_rma.c88 scif_create_pinned_pages(int nr_pages, int prot) in scif_create_pinned_pages() argument
97 pin->pages = scif_zalloc(nr_pages * sizeof(*pin->pages)); in scif_create_pinned_pages()
123 for (j = 0; j < pin->nr_pages; j++) { in scif_destroy_pinned_pages()
132 pin->nr_pages * sizeof(*pin->pages)); in scif_destroy_pinned_pages()
146 struct scif_window *scif_create_window(struct scif_endpt *ep, int nr_pages, in scif_create_window() argument
156 window->dma_addr = scif_zalloc(nr_pages * sizeof(*window->dma_addr)); in scif_create_window()
160 window->num_pages = scif_zalloc(nr_pages * sizeof(*window->num_pages)); in scif_create_window()
178 nr_pages * sizeof(*window->dma_addr)); in scif_create_window()
195 int nr_pages = window->nr_pages; in scif_destroy_incomplete_window() local
220 scif_free(window->dma_addr, nr_pages * sizeof(*window->dma_addr)); in scif_destroy_incomplete_window()
[all …]
Dscif_mmap.c43 req.nr_bytes = recv_window->nr_pages << PAGE_SHIFT; in scif_recv_munmap()
57 scif_put_window(window, window->nr_pages); in scif_recv_munmap()
139 scif_put_window(window, window->nr_pages); in __scif_cleanup_rma_for_zombies()
227 int nr_pages, err, i; in scif_get_pages() local
242 nr_pages = len >> PAGE_SHIFT; in scif_get_pages()
268 (*pages)->phys_addr = scif_zalloc(nr_pages * sizeof(dma_addr_t)); in scif_get_pages()
276 ((*pages)->va = scif_zalloc(nr_pages * sizeof(void *))); in scif_get_pages()
284 (*pages)->nr_pages = nr_pages; in scif_get_pages()
287 for (i = 0; i < nr_pages; i++) { in scif_get_pages()
300 scif_get_window(window, nr_pages); in scif_get_pages()
[all …]
Dscif_rma.h187 s64 nr_pages; member
243 s64 nr_pages; member
310 s64 offset, int nr_pages, s64 *out_offset);
315 struct scif_window *scif_create_window(struct scif_endpt *ep, int nr_pages,
426 static inline void scif_get_window(struct scif_window *window, int nr_pages) in scif_get_window() argument
428 window->ref_count += nr_pages; in scif_get_window()
431 static inline void scif_put_window(struct scif_window *window, int nr_pages) in scif_put_window() argument
433 window->ref_count -= nr_pages; in scif_put_window()
436 static inline void scif_set_window_ref(struct scif_window *window, int nr_pages) in scif_set_window_ref() argument
438 window->ref_count = nr_pages; in scif_set_window_ref()
Dscif_rma_list.c74 scif_set_window_ref(window, window->nr_pages); in scif_insert_window()
101 (window->nr_pages << PAGE_SHIFT); in scif_query_tcw()
109 (window->nr_pages << PAGE_SHIFT); in scif_query_tcw()
157 (window->nr_pages << PAGE_SHIFT); in scif_query_window()
204 s64 offset, int nr_pages) in scif_rma_list_unregister() argument
214 end_offset = window->offset + (window->nr_pages << PAGE_SHIFT); in scif_rma_list_unregister()
216 nr_pages); in scif_rma_list_unregister()
220 nr_pages -= loop_nr_pages; in scif_rma_list_unregister()
222 if (!nr_pages) in scif_rma_list_unregister()
Dscif_dma.c140 end_va = start_va + (window->nr_pages << PAGE_SHIFT); in __scif_rma_destroy_tcw()
409 (*out_window)->nr_pages = pinned_pages->nr_pages; in scif_register_temp()
664 atomic_sub(window->nr_pages, in scif_rma_destroy_tcw_invalid()
787 if (window->nr_pages == window->nr_contig_chunks) { in scif_off_to_dma_addr()
851 (window->nr_pages << PAGE_SHIFT); in scif_rma_local_cpu_copy()
856 (window->nr_pages << PAGE_SHIFT); in scif_rma_local_cpu_copy()
954 (window->nr_pages << PAGE_SHIFT); in scif_rma_list_dma_copy_unaligned()
962 (window->nr_pages << PAGE_SHIFT); in scif_rma_list_dma_copy_unaligned()
1069 (window->nr_pages << PAGE_SHIFT); in scif_rma_list_dma_copy_unaligned()
1145 (src_window->nr_pages << PAGE_SHIFT); in _scif_rma_list_dma_copy_aligned()
[all …]
Dscif_debugfs.c74 window->nr_pages, window->nr_contig_chunks, window->prot); in scif_display_window()
85 for (j = 0; j < window->nr_pages; j++) in scif_display_window()
Dscif_rma_list.h56 int nr_pages);
/linux-4.4.14/net/rds/
Dinfo.c163 unsigned long nr_pages = 0; in rds_info_getsockopt() local
188 nr_pages = (PAGE_ALIGN(start + len) - (start & PAGE_MASK)) in rds_info_getsockopt()
191 pages = kmalloc(nr_pages * sizeof(struct page *), GFP_KERNEL); in rds_info_getsockopt()
196 ret = get_user_pages_fast(start, nr_pages, 1, pages); in rds_info_getsockopt()
197 if (ret != nr_pages) { in rds_info_getsockopt()
199 nr_pages = ret; in rds_info_getsockopt()
201 nr_pages = 0; in rds_info_getsockopt()
206 rdsdebug("len %d nr_pages %lu\n", len, nr_pages); in rds_info_getsockopt()
238 for (i = 0; pages && i < nr_pages; i++) in rds_info_getsockopt()
Drdma.c157 static int rds_pin_pages(unsigned long user_addr, unsigned int nr_pages, in rds_pin_pages() argument
162 ret = get_user_pages_fast(user_addr, nr_pages, write, pages); in rds_pin_pages()
164 if (ret >= 0 && ret < nr_pages) { in rds_pin_pages()
177 unsigned int nr_pages; in __rds_rdma_map() local
197 nr_pages = rds_pages_in_vec(&args->vec); in __rds_rdma_map()
198 if (nr_pages == 0) { in __rds_rdma_map()
204 args->vec.addr, args->vec.bytes, nr_pages); in __rds_rdma_map()
207 pages = kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL); in __rds_rdma_map()
241 ret = rds_pin_pages(args->vec.addr, nr_pages, pages, 1); in __rds_rdma_map()
488 unsigned int nr_pages; in rds_rdma_pages() local
[all …]
/linux-4.4.14/kernel/events/
Dring_buffer.c128 if (unlikely(!rb->nr_pages)) in perf_output_begin()
175 handle->page = (offset >> page_shift) & (rb->nr_pages - 1); in perf_output_begin()
471 pgoff_t pgoff, int nr_pages, long watermark, int flags) in rb_alloc_aux() argument
485 max_order = ilog2(nr_pages); in rb_alloc_aux()
500 rb->aux_pages = kzalloc_node(nr_pages * sizeof(void *), GFP_KERNEL, node); in rb_alloc_aux()
505 for (rb->aux_nr_pages = 0; rb->aux_nr_pages < nr_pages;) { in rb_alloc_aux()
509 order = min(max_order, ilog2(nr_pages - rb->aux_nr_pages)); in rb_alloc_aux()
533 rb->aux_priv = event->pmu->setup_aux(event->cpu, rb->aux_pages, nr_pages, in rb_alloc_aux()
552 rb->aux_watermark = nr_pages << (PAGE_SHIFT - 1); in rb_alloc_aux()
608 if (pgoff > rb->nr_pages) in __perf_mmap_to_page()
[all …]
Dinternal.h19 int nr_pages; /* nr of data pages */ member
69 rb_alloc(int nr_pages, long watermark, int cpu, int flags);
72 pgoff_t pgoff, int nr_pages, long watermark, int flags);
110 return rb->nr_pages << (PAGE_SHIFT + page_order(rb)); in perf_data_size()
138 handle->page &= rb->nr_pages - 1; \
/linux-4.4.14/include/trace/events/
Dmigrate.h75 TP_PROTO(struct task_struct *p, int dst_nid, unsigned long nr_pages),
77 TP_ARGS(p, dst_nid, nr_pages),
83 __field( unsigned long, nr_pages)
90 __entry->nr_pages = nr_pages;
97 __entry->nr_pages)
Dwriteback.h242 __field(long, nr_pages)
254 __entry->nr_pages = work->nr_pages;
267 __entry->nr_pages,
/linux-4.4.14/fs/
Dfs-writeback.c46 long nr_pages; member
757 static long wb_split_bdi_pages(struct bdi_writeback *wb, long nr_pages) in wb_split_bdi_pages() argument
762 if (nr_pages == LONG_MAX) in wb_split_bdi_pages()
771 return nr_pages; in wb_split_bdi_pages()
773 return DIV_ROUND_UP_ULL((u64)nr_pages * this_bw, tot_bw); in wb_split_bdi_pages()
802 long nr_pages; in bdi_split_work_to_wbs() local
817 nr_pages = wb_split_bdi_pages(wb, base_work->nr_pages); in bdi_split_work_to_wbs()
822 work->nr_pages = nr_pages; in bdi_split_work_to_wbs()
831 work->nr_pages = nr_pages; in bdi_split_work_to_wbs()
905 static long wb_split_bdi_pages(struct bdi_writeback *wb, long nr_pages) in wb_split_bdi_pages() argument
[all …]
Dsplice.c185 unsigned int spd_pages = spd->nr_pages; in splice_to_pipe()
224 if (!--spd->nr_pages) in splice_to_pipe()
312 unsigned int loff, nr_pages, req_pages; in __generic_file_splice_read() local
334 nr_pages = min(req_pages, spd.nr_pages_max); in __generic_file_splice_read()
339 spd.nr_pages = find_get_pages_contig(mapping, index, nr_pages, spd.pages); in __generic_file_splice_read()
340 index += spd.nr_pages; in __generic_file_splice_read()
346 if (spd.nr_pages < nr_pages) in __generic_file_splice_read()
348 index, req_pages - spd.nr_pages); in __generic_file_splice_read()
351 while (spd.nr_pages < nr_pages) { in __generic_file_splice_read()
380 spd.pages[spd.nr_pages++] = page; in __generic_file_splice_read()
[all …]
Dpipe.c1000 static long pipe_set_size(struct pipe_inode_info *pipe, unsigned long nr_pages) in pipe_set_size() argument
1010 if (nr_pages < pipe->nrbufs) in pipe_set_size()
1013 bufs = kcalloc(nr_pages, sizeof(*bufs), GFP_KERNEL | __GFP_NOWARN); in pipe_set_size()
1038 account_pipe_buffers(pipe, pipe->buffers, nr_pages); in pipe_set_size()
1042 pipe->buffers = nr_pages; in pipe_set_size()
1043 return nr_pages * PAGE_SIZE; in pipe_set_size()
1052 unsigned long nr_pages; in round_pipe_size() local
1054 nr_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; in round_pipe_size()
1055 return roundup_pow_of_two(nr_pages) << PAGE_SHIFT; in round_pipe_size()
1098 unsigned int size, nr_pages; in pipe_fcntl() local
[all …]
Daio.c116 long nr_pages; member
205 static struct file *aio_private_file(struct kioctx *ctx, loff_t nr_pages) in aio_private_file() argument
216 inode->i_size = PAGE_SIZE * nr_pages; in aio_private_file()
294 for (i = 0; i < ctx->nr_pages; i++) { in aio_free_ring()
388 if (idx < (pgoff_t)ctx->nr_pages) { in aio_migratepage()
442 int nr_pages; in aio_setup_ring() local
452 nr_pages = PFN_UP(size); in aio_setup_ring()
453 if (nr_pages < 0) in aio_setup_ring()
456 file = aio_private_file(ctx, nr_pages); in aio_setup_ring()
463 nr_events = (PAGE_SIZE * nr_pages - sizeof(struct aio_ring)) in aio_setup_ring()
[all …]
Dmpage.c140 do_mpage_readpage(struct bio *bio, struct page *page, unsigned nr_pages, in do_mpage_readpage() argument
165 last_block = block_in_file + nr_pages * blocks_per_page; in do_mpage_readpage()
281 min_t(int, nr_pages, BIO_MAX_PAGES), gfp); in do_mpage_readpage()
357 unsigned nr_pages, get_block_t get_block) in mpage_readpages() argument
368 for (page_idx = 0; page_idx < nr_pages; page_idx++) { in mpage_readpages()
377 nr_pages - page_idx, in mpage_readpages()
/linux-4.4.14/fs/logfs/
Ddev_bdev.c75 size_t nr_pages) in __bdev_writeseg() argument
84 max_pages = min_t(size_t, nr_pages, BIO_MAX_PAGES); in __bdev_writeseg()
89 for (i = 0; i < nr_pages; i++) { in __bdev_writeseg()
103 nr_pages -= i; in __bdev_writeseg()
119 bio->bi_vcnt = nr_pages; in __bdev_writeseg()
120 bio->bi_iter.bi_size = nr_pages * PAGE_SIZE; in __bdev_writeseg()
167 size_t nr_pages) in do_erase() argument
174 max_pages = min_t(size_t, nr_pages, BIO_MAX_PAGES); in do_erase()
179 for (i = 0; i < nr_pages; i++) { in do_erase()
193 nr_pages -= i; in do_erase()
[all …]
Ddev_mtd.c184 size_t nr_pages) in __logfs_mtd_writeseg() argument
191 for (i = 0; i < nr_pages; i++) { in __logfs_mtd_writeseg()
/linux-4.4.14/net/ipv4/
Dtcp_memcontrol.c57 static int tcp_update_limit(struct mem_cgroup *memcg, unsigned long nr_pages) in tcp_update_limit() argument
67 ret = page_counter_limit(&cg_proto->memory_allocated, nr_pages); in tcp_update_limit()
72 cg_proto->sysctl_mem[i] = min_t(long, nr_pages, in tcp_update_limit()
75 if (nr_pages == PAGE_COUNTER_MAX) in tcp_update_limit()
119 unsigned long nr_pages; in tcp_cgroup_write() local
127 ret = page_counter_memparse(buf, "-1", &nr_pages); in tcp_cgroup_write()
131 ret = tcp_update_limit(memcg, nr_pages); in tcp_cgroup_write()
/linux-4.4.14/arch/arm/mach-rpc/include/mach/
Duncompress.h23 unsigned long nr_pages; member
119 unsigned int nr_pages = 0, page_size = PAGE_SIZE; in arch_decomp_setup() local
138 nr_pages += (t->u.mem.size / PAGE_SIZE); in arch_decomp_setup()
144 nr_pages = params->nr_pages; in arch_decomp_setup()
189 if (nr_pages * page_size < 4096*1024) error("<4M of mem\n"); in arch_decomp_setup()
/linux-4.4.14/drivers/net/ethernet/ibm/ehea/
Dehea_qmr.c107 int i, nr_pages; in hw_queue_dtor() local
114 nr_pages = queue->queue_length / queue->pagesize; in hw_queue_dtor()
116 for (i = 0; i < nr_pages; i += pages_per_kpage) in hw_queue_dtor()
153 ret = hw_queue_ctor(&cq->hw_queue, cq->attr.nr_pages, in ehea_create_cq()
158 for (counter = 0; counter < cq->attr.nr_pages; counter++) { in ehea_create_cq()
171 cq, hret, counter, cq->attr.nr_pages); in ehea_create_cq()
175 if (counter == (cq->attr.nr_pages - 1)) { in ehea_create_cq()
275 ret = hw_queue_ctor(&eq->hw_queue, eq->attr.nr_pages, in ehea_create_eq()
282 for (i = 0; i < eq->attr.nr_pages; i++) { in ehea_create_eq()
296 if (i == (eq->attr.nr_pages - 1)) { in ehea_create_eq()
[all …]
Dehea.h260 u32 nr_pages; member
305 u32 nr_pages; member
Dehea_qmr.h398 int ehea_add_sect_bmap(unsigned long pfn, unsigned long nr_pages);
399 int ehea_rem_sect_bmap(unsigned long pfn, unsigned long nr_pages);
/linux-4.4.14/drivers/firmware/efi/libstub/
Defi-stub-helper.c148 unsigned long nr_pages; in efi_high_alloc() local
165 nr_pages = round_up(size, EFI_ALLOC_ALIGN) / EFI_PAGE_SIZE; in efi_high_alloc()
176 if (desc->num_pages < nr_pages) in efi_high_alloc()
209 nr_pages, &max_addr); in efi_high_alloc()
234 unsigned long nr_pages; in efi_low_alloc() local
250 nr_pages = round_up(size, EFI_ALLOC_ALIGN) / EFI_PAGE_SIZE; in efi_low_alloc()
261 if (desc->num_pages < nr_pages) in efi_low_alloc()
281 nr_pages, &start); in efi_low_alloc()
299 unsigned long nr_pages; in efi_free() local
304 nr_pages = round_up(size, EFI_ALLOC_ALIGN) / EFI_PAGE_SIZE; in efi_free()
[all …]
Darm64-stub.c26 unsigned long nr_pages; in handle_kernel_image() local
56 nr_pages = round_up(kernel_memsize, EFI_ALLOC_ALIGN) / in handle_kernel_image()
59 EFI_LOADER_DATA, nr_pages, in handle_kernel_image()
/linux-4.4.14/arch/s390/mm/
Dinit.c155 unsigned long nr_pages; in arch_add_memory() local
164 nr_pages = (start_pfn + size_pages > dma_end_pfn) ? in arch_add_memory()
168 nr_pages = (start_pfn + size_pages > normal_end_pfn) ? in arch_add_memory()
172 nr_pages = size_pages; in arch_add_memory()
179 start_pfn += nr_pages; in arch_add_memory()
180 size_pages -= nr_pages; in arch_add_memory()
Dgup.c176 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, in __get_user_pages_fast() argument
187 len = (unsigned long) nr_pages << PAGE_SHIFT; in __get_user_pages_fast()
230 int get_user_pages_fast(unsigned long start, int nr_pages, int write, in get_user_pages_fast() argument
237 nr = __get_user_pages_fast(start, nr_pages, write, pages); in get_user_pages_fast()
238 if (nr == nr_pages) in get_user_pages_fast()
245 nr_pages - nr, write, 0, pages); in get_user_pages_fast()
/linux-4.4.14/fs/btrfs/
Dcompression.c68 unsigned long nr_pages; member
117 for (i = 0; i < cb->nr_pages; i++) { in check_compressed_csum()
190 for (index = 0; index < cb->nr_pages; index++) { in end_compressed_bio_read()
230 unsigned long nr_pages = end_index - index + 1; in end_compressed_writeback() local
237 while (nr_pages > 0) { in end_compressed_writeback()
240 nr_pages, ARRAY_SIZE(pages)), pages); in end_compressed_writeback()
242 nr_pages -= 1; in end_compressed_writeback()
252 nr_pages -= ret; in end_compressed_writeback()
304 for (index = 0; index < cb->nr_pages; index++) { in end_compressed_bio_write()
330 unsigned long nr_pages) in btrfs_submit_compressed_write() argument
[all …]
Dzlib.c88 int nr_pages = 0; in zlib_compress_pages() local
116 nr_pages = 1; in zlib_compress_pages()
146 if (nr_pages == nr_dest_pages) { in zlib_compress_pages()
157 pages[nr_pages] = out_page; in zlib_compress_pages()
158 nr_pages++; in zlib_compress_pages()
202 *out_pages = nr_pages; in zlib_compress_pages()
Dlzo.c101 int nr_pages = 0; in lzo_compress_pages() local
135 nr_pages = 1; in lzo_compress_pages()
190 if (nr_pages == nr_dest_pages) { in lzo_compress_pages()
202 pages[nr_pages++] = out_page; in lzo_compress_pages()
245 *out_pages = nr_pages; in lzo_compress_pages()
Draid56.c141 int nr_pages; member
266 for (i = 0; i < rbio->nr_pages; i++) { in cache_rbio_pages()
313 for (i = 0; i < dest->nr_pages; i++) { in steal_rbio()
845 for (i = 0; i < rbio->nr_pages; i++) { in __free_raid_bio()
984 rbio->nr_pages = num_pages; in alloc_rbio()
1019 for (i = 0; i < rbio->nr_pages; i++) { in alloc_rbio_pages()
1039 for (; i < rbio->nr_pages; i++) { in alloc_rbio_parity_pages()
1511 int nr_pages = DIV_ROUND_UP(rbio->stripe_len, PAGE_CACHE_SIZE); in raid56_rmw_stripe() local
1530 for (pagenr = 0; pagenr < nr_pages; pagenr++) { in raid56_rmw_stripe()
1806 int nr_pages = DIV_ROUND_UP(rbio->stripe_len, PAGE_CACHE_SIZE); in __raid_recover_end_io() local
[all …]
Dcompression.h45 unsigned long nr_pages);
/linux-4.4.14/fs/9p/
Dcache.c159 int loop, nr_pages; in v9fs_cache_inode_now_uncached() local
165 nr_pages = pagevec_lookup(&pvec, v9inode->vfs_inode.i_mapping, in v9fs_cache_inode_now_uncached()
168 if (!nr_pages) in v9fs_cache_inode_now_uncached()
171 for (loop = 0; loop < nr_pages; loop++) in v9fs_cache_inode_now_uncached()
174 first = pvec.pages[nr_pages - 1]->index + 1; in v9fs_cache_inode_now_uncached()
176 pvec.nr = nr_pages; in v9fs_cache_inode_now_uncached()
358 unsigned *nr_pages) in __v9fs_readpages_from_fscache() argument
363 p9_debug(P9_DEBUG_FSC, "inode %p pages %u\n", inode, *nr_pages); in __v9fs_readpages_from_fscache()
368 mapping, pages, nr_pages, in __v9fs_readpages_from_fscache()
379 BUG_ON(*nr_pages != 0); in __v9fs_readpages_from_fscache()
Dcache.h52 unsigned *nr_pages);
77 unsigned *nr_pages) in v9fs_readpages_from_fscache() argument
80 nr_pages); in v9fs_readpages_from_fscache()
133 unsigned *nr_pages) in v9fs_readpages_from_fscache() argument
Dvfs_addr.c112 struct list_head *pages, unsigned nr_pages) in v9fs_vfs_readpages() argument
120 ret = v9fs_readpages_from_fscache(inode, mapping, pages, &nr_pages); in v9fs_vfs_readpages()
/linux-4.4.14/tools/testing/selftests/vm/
Duserfaultfd.c71 static unsigned long nr_cpus, nr_pages, nr_pages_per_cpu, page_size; variable
147 page_nr %= nr_pages; in locking_thread()
224 if (offset >= nr_pages * page_size) in copy_page()
387 if (madvise(area_src, nr_pages * page_size, MADV_DONTNEED)) { in stress()
428 if (posix_memalign(&area, page_size, nr_pages * page_size)) { in userfaultfd_stress()
433 if (posix_memalign(&area, page_size, nr_pages * page_size)) { in userfaultfd_stress()
458 count_verify = malloc(nr_pages * sizeof(unsigned long long)); in userfaultfd_stress()
464 for (nr = 0; nr < nr_pages; nr++) { in userfaultfd_stress()
525 uffdio_register.range.len = nr_pages * page_size; in userfaultfd_stress()
565 if (madvise(area_dst, nr_pages * page_size, MADV_DONTNEED)) { in userfaultfd_stress()
[all …]
Dtranshuge-stress.c104 int nr_succeed = 0, nr_failed = 0, nr_pages = 0; in main() local
128 nr_pages++; in main()
142 nr_succeed, nr_failed, nr_pages); in main()
/linux-4.4.14/arch/arm/xen/
Dp2m.c23 unsigned long nr_pages; member
72 entry->pfn + entry->nr_pages > pfn) { in __pfn_to_mfn()
120 unsigned long mfn, unsigned long nr_pages) in __set_phys_to_machine_multi() argument
132 p2m_entry->pfn + p2m_entry->nr_pages > pfn) { in __set_phys_to_machine_multi()
153 p2m_entry->nr_pages = nr_pages; in __set_phys_to_machine_multi()
/linux-4.4.14/drivers/misc/genwqe/
Dcard_utils.c306 sgl->nr_pages = DIV_ROUND_UP(sgl->fpage_offs + user_size, PAGE_SIZE); in genwqe_alloc_sync_sgl()
310 __func__, user_addr, user_size, sgl->nr_pages, in genwqe_alloc_sync_sgl()
315 sgl->sgl_size = genwqe_sgl_size(sgl->nr_pages); in genwqe_alloc_sync_sgl()
383 while (p < sgl->nr_pages) { in genwqe_setup_sgl()
401 } else if ((p == sgl->nr_pages - 1) && in genwqe_setup_sgl()
422 if (p == sgl->nr_pages) in genwqe_setup_sgl()
438 if (p == sgl->nr_pages) in genwqe_setup_sgl()
520 static int free_user_pages(struct page **page_list, unsigned int nr_pages, in free_user_pages() argument
525 for (i = 0; i < nr_pages; i++) { in free_user_pages()
575 m->nr_pages = DIV_ROUND_UP(offs + size, PAGE_SIZE); in genwqe_user_vmap()
[all …]
Dcard_base.h181 unsigned int nr_pages; /* number of pages */ member
360 unsigned long nr_pages; member
/linux-4.4.14/kernel/power/
Dsnapshot.c1378 static unsigned long preallocate_image_pages(unsigned long nr_pages, gfp_t mask) in preallocate_image_pages() argument
1382 while (nr_pages > 0) { in preallocate_image_pages()
1393 nr_pages--; in preallocate_image_pages()
1400 static unsigned long preallocate_image_memory(unsigned long nr_pages, in preallocate_image_memory() argument
1409 if (nr_pages < alloc) in preallocate_image_memory()
1410 alloc = nr_pages; in preallocate_image_memory()
1416 static unsigned long preallocate_image_highmem(unsigned long nr_pages) in preallocate_image_highmem() argument
1418 return preallocate_image_pages(nr_pages, GFP_IMAGE | __GFP_HIGHMEM); in preallocate_image_highmem()
1431 static unsigned long preallocate_highmem_fraction(unsigned long nr_pages, in preallocate_highmem_fraction() argument
1435 unsigned long alloc = __fraction(nr_pages, highmem, total); in preallocate_highmem_fraction()
[all …]
Dswap.c519 int nr_pages; in save_image() local
532 nr_pages = 0; in save_image()
541 if (!(nr_pages % m)) in save_image()
543 nr_pages / m * 10); in save_image()
544 nr_pages++; in save_image()
655 int nr_pages; in save_image_lzo() local
752 nr_pages = 0; in save_image_lzo()
767 if (!(nr_pages % m)) in save_image_lzo()
771 nr_pages / m * 10); in save_image_lzo()
772 nr_pages++; in save_image_lzo()
[all …]
/linux-4.4.14/drivers/edac/
Die31200_edac.c406 unsigned long nr_pages; in ie31200_probe1() local
408 nr_pages = IE31200_PAGES(dimm_info[j][i].size); in ie31200_probe1()
409 if (nr_pages == 0) in ie31200_probe1()
413 nr_pages = nr_pages / 2; in ie31200_probe1()
417 dimm->nr_pages = nr_pages; in ie31200_probe1()
418 edac_dbg(0, "set nr pages: 0x%lx\n", nr_pages); in ie31200_probe1()
426 dimm->nr_pages = nr_pages; in ie31200_probe1()
427 edac_dbg(0, "set nr pages: 0x%lx\n", nr_pages); in ie31200_probe1()
Dpasemi_edac.c157 dimm->nr_pages = 128 << (20 - PAGE_SHIFT); in pasemi_edac_init_csrows()
160 dimm->nr_pages = 256 << (20 - PAGE_SHIFT); in pasemi_edac_init_csrows()
164 dimm->nr_pages = 512 << (20 - PAGE_SHIFT); in pasemi_edac_init_csrows()
167 dimm->nr_pages = 1024 << (20 - PAGE_SHIFT); in pasemi_edac_init_csrows()
170 dimm->nr_pages = 2048 << (20 - PAGE_SHIFT); in pasemi_edac_init_csrows()
180 csrow->last_page = csrow->first_page + dimm->nr_pages - 1; in pasemi_edac_init_csrows()
181 last_page_in_mmc += dimm->nr_pages; in pasemi_edac_init_csrows()
Dcell_edac.c135 u32 nr_pages; in cell_edac_init_csrows() local
149 nr_pages = resource_size(&r) >> PAGE_SHIFT; in cell_edac_init_csrows()
150 csrow->last_page = csrow->first_page + nr_pages - 1; in cell_edac_init_csrows()
156 dimm->nr_pages = nr_pages / csrow->nr_channels; in cell_edac_init_csrows()
162 csrow->first_page, nr_pages); in cell_edac_init_csrows()
Di3200_edac.c395 unsigned long nr_pages; in i3200_probe1() local
401 nr_pages = drb_to_nr_pages(drbs, stacked, j, i); in i3200_probe1()
402 if (nr_pages == 0) in i3200_probe1()
406 stacked ? " (stacked)" : "", PAGES_TO_MiB(nr_pages)); in i3200_probe1()
408 dimm->nr_pages = nr_pages; in i3200_probe1()
409 dimm->grain = nr_pages << PAGE_SHIFT; in i3200_probe1()
Dedac_mc_sysfs.c187 u32 nr_pages = 0; in csrow_size_show() local
190 nr_pages += csrow->channels[i]->dimm->nr_pages; in csrow_size_show()
191 return sprintf(data, "%u\n", PAGES_TO_MiB(nr_pages)); in csrow_size_show()
375 if (!csrow->channels[idx]->dimm->nr_pages) in csrow_dev_is_visible()
399 int chan, nr_pages = 0; in nr_pages_per_csrow() local
402 nr_pages += csrow->channels[chan]->dimm->nr_pages; in nr_pages_per_csrow()
404 return nr_pages; in nr_pages_per_csrow()
526 return sprintf(data, "%u\n", PAGES_TO_MiB(dimm->nr_pages)); in dimmdev_size_show()
778 total_pages += dimm->nr_pages; in mci_size_mb_show()
928 if (!dimm->nr_pages) in edac_create_sysfs_mci_device()
[all …]
Dx38_edac.c376 unsigned long nr_pages; in x38_probe1() local
379 nr_pages = drb_to_nr_pages(drbs, stacked, in x38_probe1()
383 if (nr_pages == 0) in x38_probe1()
389 dimm->nr_pages = nr_pages / x38_channel_num; in x38_probe1()
390 dimm->grain = nr_pages << PAGE_SHIFT; in x38_probe1()
Dghes_edac.c91 dimm->nr_pages = MiB_TO_PAGES(32);/* Unknown */ in ghes_edac_dmidecode()
93 dimm->nr_pages = MiB_TO_PAGES(entry->extended_size); in ghes_edac_dmidecode()
96 dimm->nr_pages = MiB_TO_PAGES((entry->size & in ghes_edac_dmidecode()
99 dimm->nr_pages = MiB_TO_PAGES(entry->size); in ghes_edac_dmidecode()
154 if (dimm->nr_pages) { in ghes_edac_dmidecode()
157 PAGES_TO_MiB(dimm->nr_pages), in ghes_edac_dmidecode()
492 dimm->nr_pages = 1; in ghes_edac_register()
Di82975x_edac.c376 u32 cumul_size, nr_pages; in i82975x_init_csrows() local
407 nr_pages = cumul_size - last_cumul_size; in i82975x_init_csrows()
408 if (!nr_pages) in i82975x_init_csrows()
421 dimm->nr_pages = nr_pages / csrow->nr_channels; in i82975x_init_csrows()
Dedac_mc.c97 edac_dbg(4, " dimm->nr_pages = 0x%x\n", dimm->nr_pages); in edac_mc_dump_dimm()
99 edac_dbg(4, " dimm->nr_pages = 0x%x\n", dimm->nr_pages); in edac_mc_dump_dimm()
741 u32 nr_pages = 0; in edac_mc_add_mc_with_groups() local
745 nr_pages += csrow->channels[j]->dimm->nr_pages; in edac_mc_add_mc_with_groups()
746 if (!nr_pages) in edac_mc_add_mc_with_groups()
750 if (csrow->channels[j]->dimm->nr_pages) in edac_mc_add_mc_with_groups()
754 if (mci->dimms[i]->nr_pages) in edac_mc_add_mc_with_groups()
897 n += dimm->nr_pages; in edac_mc_find_csrow_by_page()
1229 if (e->enable_per_layer_report && dimm->nr_pages) { in edac_mc_handle_error()
Di3000_edac.c317 unsigned long last_cumul_size, nr_pages; in i3000_probe1() local
408 nr_pages = cumul_size - last_cumul_size; in i3000_probe1()
414 dimm->nr_pages = nr_pages / nr_channels; in i3000_probe1()
Di82875p_edac.c351 u32 cumul_size, nr_pages; in i82875p_init_csrows() local
374 nr_pages = cumul_size - last_cumul_size; in i82875p_init_csrows()
380 dimm->nr_pages = nr_pages / nr_chans; in i82875p_init_csrows()
Damd76x_edac.c211 dimm->nr_pages = (mba_mask + 1) >> PAGE_SHIFT; in amd76x_init_csrows()
212 csrow->last_page = csrow->first_page + dimm->nr_pages - 1; in amd76x_init_csrows()
214 dimm->grain = dimm->nr_pages << PAGE_SHIFT; in amd76x_init_csrows()
De7xxx_edac.c361 u32 dra, cumul_size, nr_pages; in e7xxx_init_csrows() local
392 nr_pages = cumul_size - last_cumul_size; in e7xxx_init_csrows()
413 dimm->nr_pages = nr_pages / (drc_chan + 1); in e7xxx_init_csrows()
Dcpc925_edac.c335 unsigned long row_size, nr_pages, last_nr_pages = 0; in cpc925_init_csrows() local
354 nr_pages = row_size >> PAGE_SHIFT; in cpc925_init_csrows()
355 csrow->last_page = csrow->first_page + nr_pages - 1; in cpc925_init_csrows()
383 dimm->nr_pages = nr_pages / csrow->nr_channels; in cpc925_init_csrows()
Dppc4xx_edac.c902 u32 mbxcf, size, nr_pages; in ppc4xx_edac_init_csrows() local
953 nr_pages = SDRAM_MBCF_SZ_TO_PAGES(size); in ppc4xx_edac_init_csrows()
979 dimm->nr_pages = nr_pages / csi->nr_channels; in ppc4xx_edac_init_csrows()
Damd64_edac.c2389 u32 cs_mode, nr_pages; in get_csrow_nr_pages() local
2402 nr_pages = pvt->ops->dbam_to_cs(pvt, dct, cs_mode, (csrow_nr / 2)) in get_csrow_nr_pages()
2407 edac_dbg(0, "nr_pages/channel: %u\n", nr_pages); in get_csrow_nr_pages()
2409 return nr_pages; in get_csrow_nr_pages()
2423 int nr_pages = 0; in init_csrows() local
2454 nr_pages = get_csrow_nr_pages(pvt, 0, i); in init_csrows()
2455 csrow->channels[0]->dimm->nr_pages = nr_pages; in init_csrows()
2462 csrow->channels[1]->dimm->nr_pages = row_dct1_pages; in init_csrows()
2463 nr_pages += row_dct1_pages; in init_csrows()
2466 edac_dbg(1, "Total csrow%d pages: %u\n", i, nr_pages); in init_csrows()
/linux-4.4.14/drivers/staging/rdma/ehca/
Dehca_eq.c59 u32 nr_pages; in ehca_create_eq() local
83 &nr_pages, &eq->ist); in ehca_create_eq()
90 ret = ipz_queue_ctor(NULL, &eq->ipz_queue, nr_pages, in ehca_create_eq()
97 for (i = 0; i < nr_pages; i++) { in ehca_create_eq()
110 if (i == (nr_pages - 1)) { in ehca_create_eq()
Dipz_pt_fn.c255 int i, nr_pages; in ipz_queue_dtor() local
265 nr_pages = queue->queue_length / queue->pagesize; in ipz_queue_dtor()
266 for (i = 0; i < nr_pages; i += PAGES_PER_KPAGE) in ipz_queue_dtor()
/linux-4.4.14/fs/exofs/
Dinode.c57 unsigned nr_pages; member
78 pcol->nr_pages = 0; in _pcol_init()
87 pcol->expected_pages -= min(pcol->nr_pages, pcol->expected_pages); in _pcol_reset()
91 pcol->nr_pages = 0; in _pcol_reset()
140 if (unlikely(pcol->nr_pages >= pcol->alloc_pages)) in pcol_add_page()
143 pcol->pages[pcol->nr_pages++] = page; in pcol_add_page()
212 pcol->nr_pages); in __readpages_done()
214 for (i = 0; i < pcol->nr_pages; i++) { in __readpages_done()
256 for (i = 0; i < pcol->nr_pages; i++) { in _unlock_pcol_pages()
272 BUG_ON(pcol_src->nr_pages < ios->nr_pages); in _maybe_not_all_in_one_io()
[all …]
/linux-4.4.14/drivers/media/pci/cx23885/
Dcx23885-alsa.c83 static int cx23885_alsa_dma_init(struct cx23885_audio_dev *chip, int nr_pages) in cx23885_alsa_dma_init() argument
89 buf->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT); in cx23885_alsa_dma_init()
91 dprintk(1, "vmalloc_32(%d pages) failed\n", nr_pages); in cx23885_alsa_dma_init()
97 nr_pages << PAGE_SHIFT); in cx23885_alsa_dma_init()
99 memset(buf->vaddr, 0, nr_pages << PAGE_SHIFT); in cx23885_alsa_dma_init()
100 buf->nr_pages = nr_pages; in cx23885_alsa_dma_init()
102 buf->sglist = vzalloc(buf->nr_pages * sizeof(*buf->sglist)); in cx23885_alsa_dma_init()
106 sg_init_table(buf->sglist, buf->nr_pages); in cx23885_alsa_dma_init()
107 for (i = 0; i < buf->nr_pages; i++) { in cx23885_alsa_dma_init()
129 buf->nr_pages, PCI_DMA_FROMDEVICE); in cx23885_alsa_dma_map()
/linux-4.4.14/fs/nfs/
Dfscache-index.c266 int loop, nr_pages; in nfs_fscache_inode_now_uncached() local
275 nr_pages = pagevec_lookup(&pvec, in nfs_fscache_inode_now_uncached()
279 if (!nr_pages) in nfs_fscache_inode_now_uncached()
282 for (loop = 0; loop < nr_pages; loop++) in nfs_fscache_inode_now_uncached()
285 first = pvec.pages[nr_pages - 1]->index + 1; in nfs_fscache_inode_now_uncached()
287 pvec.nr = nr_pages; in nfs_fscache_inode_now_uncached()
Dfscache.c370 unsigned *nr_pages) in __nfs_readpages_from_fscache() argument
372 unsigned npages = *nr_pages; in __nfs_readpages_from_fscache()
379 mapping, pages, nr_pages, in __nfs_readpages_from_fscache()
383 if (*nr_pages < npages) in __nfs_readpages_from_fscache()
386 if (*nr_pages > 0) in __nfs_readpages_from_fscache()
388 *nr_pages); in __nfs_readpages_from_fscache()
393 BUG_ON(*nr_pages != 0); in __nfs_readpages_from_fscache()
Dfscache.h133 unsigned *nr_pages) in nfs_readpages_from_fscache() argument
137 nr_pages); in nfs_readpages_from_fscache()
212 unsigned *nr_pages) in nfs_readpages_from_fscache() argument
Dread.c376 struct list_head *pages, unsigned nr_pages) in nfs_readpages() argument
390 nr_pages); in nfs_readpages()
407 pages, &nr_pages); in nfs_readpages()
/linux-4.4.14/fs/ceph/
Dcache.c151 int loop, nr_pages; in ceph_fscache_inode_now_uncached() local
159 nr_pages = pagevec_lookup(&pvec, ci->vfs_inode.i_mapping, first, in ceph_fscache_inode_now_uncached()
162 if (!nr_pages) in ceph_fscache_inode_now_uncached()
165 for (loop = 0; loop < nr_pages; loop++) in ceph_fscache_inode_now_uncached()
168 first = pvec.pages[nr_pages - 1]->index + 1; in ceph_fscache_inode_now_uncached()
170 pvec.nr = nr_pages; in ceph_fscache_inode_now_uncached()
282 unsigned *nr_pages) in ceph_readpages_from_fscache() argument
290 ret = fscache_read_or_alloc_pages(ci->fscache, mapping, pages, nr_pages, in ceph_readpages_from_fscache()
Dcache.h46 unsigned *nr_pages);
134 unsigned *nr_pages) in ceph_readpages_from_fscache() argument
Daddr.c324 int nr_pages = 0; in start_read() local
334 nr_pages++; in start_read()
336 if (max && nr_pages == max) in start_read()
339 len = nr_pages << PAGE_CACHE_SHIFT; in start_read()
340 dout("start_read %p nr_pages %d is %lld~%lld\n", inode, nr_pages, in start_read()
352 nr_pages = calc_pages_for(0, len); in start_read()
353 pages = kmalloc(sizeof(*pages) * nr_pages, GFP_KERNEL); in start_read()
357 for (i = 0; i < nr_pages; ++i) { in start_read()
370 nr_pages = i; in start_read()
386 return nr_pages; in start_read()
[all …]
/linux-4.4.14/arch/x86/xen/
Dsetup.c254 unsigned long end_pfn, unsigned long nr_pages) in xen_set_identity_and_release_chunk() argument
262 end = min(end_pfn, nr_pages); in xen_set_identity_and_release_chunk()
389 unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, in xen_set_identity_and_remap_chunk() argument
397 remap_pfn = nr_pages; in xen_set_identity_and_remap_chunk()
406 if (cur_pfn >= nr_pages) { in xen_set_identity_and_remap_chunk()
411 if (cur_pfn + size > nr_pages) in xen_set_identity_and_remap_chunk()
412 size = nr_pages - cur_pfn; in xen_set_identity_and_remap_chunk()
418 cur_pfn + left, nr_pages); in xen_set_identity_and_remap_chunk()
445 unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, in xen_count_remap_pages() argument
448 if (start_pfn >= nr_pages) in xen_count_remap_pages()
[all …]
/linux-4.4.14/arch/powerpc/kvm/
Dbook3s_hv_builtin.c52 struct page *kvm_alloc_hpt(unsigned long nr_pages) in kvm_alloc_hpt() argument
54 VM_BUG_ON(order_base_2(nr_pages) < KVM_CMA_CHUNK_ORDER - PAGE_SHIFT); in kvm_alloc_hpt()
56 return cma_alloc(kvm_cma, nr_pages, order_base_2(HPT_ALIGN_PAGES)); in kvm_alloc_hpt()
60 void kvm_release_hpt(struct page *page, unsigned long nr_pages) in kvm_release_hpt() argument
62 cma_release(kvm_cma, page, nr_pages); in kvm_release_hpt()
/linux-4.4.14/arch/arm/kernel/
Datags_compat.c46 unsigned long nr_pages; /* 4 */ member
107 if (params->u1.s.nr_pages != 0x02000 && in build_tag_list()
108 params->u1.s.nr_pages != 0x04000 && in build_tag_list()
109 params->u1.s.nr_pages != 0x08000 && in build_tag_list()
110 params->u1.s.nr_pages != 0x10000) { in build_tag_list()
113 params->u1.s.nr_pages = 0x1000; /* 16MB */ in build_tag_list()
161 tag = memtag(tag, PHYS_OFFSET, params->u1.s.nr_pages * PAGE_SIZE); in build_tag_list()
/linux-4.4.14/arch/x86/kernel/cpu/
Dperf_event_intel_pt.c440 p = virt_to_page(buf->data_pages[buf->nr_pages]); in topa_insert_pages()
462 buf->nr_pages += 1ul << order; in topa_insert_pages()
540 ((buf->nr_pages << PAGE_SHIFT) - 1)); in pt_update_head()
542 base += buf->nr_pages << PAGE_SHIFT; in pt_update_head()
659 pg &= buf->nr_pages - 1; in pt_topa_next_entry()
708 idx &= buf->nr_pages - 1; in pt_buffer_reset_markers()
718 idx &= buf->nr_pages - 1; in pt_buffer_reset_markers()
742 while (pg < buf->nr_pages) { in pt_buffer_setup_topa_index()
783 head &= (buf->nr_pages << PAGE_SHIFT) - 1; in pt_buffer_reset_offsets()
785 pg = (head >> PAGE_SHIFT) & (buf->nr_pages - 1); in pt_buffer_reset_offsets()
[all …]
Dperf_event_intel_bts.c51 unsigned int nr_pages; member
71 bts_buffer_setup_aux(int cpu, void **pages, int nr_pages, bool overwrite) in bts_buffer_setup_aux() argument
77 size_t size = nr_pages << PAGE_SHIFT; in bts_buffer_setup_aux()
81 for (pg = 0, nbuf = 0; pg < nr_pages;) { in bts_buffer_setup_aux()
83 if (WARN_ON_ONCE(!PagePrivate(page) && nr_pages > 1)) in bts_buffer_setup_aux()
99 buf->nr_pages = nr_pages; in bts_buffer_setup_aux()
310 head = handle->head & ((buf->nr_pages << PAGE_SHIFT) - 1); in bts_buffer_reset()
430 buf->nr_pages << PAGE_SHIFT); in bts_event_del()
Dintel_pt.h96 unsigned long nr_pages; member
/linux-4.4.14/fs/cifs/
Dfile.c1892 unsigned int nr_pages; in wdata_alloc_and_fillpages() local
1910 nr_pages = find_get_pages_tag(mapping, index, in wdata_alloc_and_fillpages()
1913 *found_pages += nr_pages; in wdata_alloc_and_fillpages()
1914 tofind -= nr_pages; in wdata_alloc_and_fillpages()
1915 pages += nr_pages; in wdata_alloc_and_fillpages()
1916 } while (nr_pages && tofind && *index <= end); in wdata_alloc_and_fillpages()
1927 unsigned int nr_pages = 0, i; in wdata_prepare_pages() local
1940 if (nr_pages == 0) in wdata_prepare_pages()
1985 ++nr_pages; in wdata_prepare_pages()
1989 if (nr_pages == 0) in wdata_prepare_pages()
[all …]
Dcache.c300 int loop, nr_pages; in cifs_fscache_inode_now_uncached() local
308 nr_pages = pagevec_lookup(&pvec, in cifs_fscache_inode_now_uncached()
311 if (!nr_pages) in cifs_fscache_inode_now_uncached()
314 for (loop = 0; loop < nr_pages; loop++) in cifs_fscache_inode_now_uncached()
317 first = pvec.pages[nr_pages - 1]->index + 1; in cifs_fscache_inode_now_uncached()
319 pvec.nr = nr_pages; in cifs_fscache_inode_now_uncached()
Dfscache.h80 unsigned *nr_pages) in cifs_readpages_from_fscache() argument
84 nr_pages); in cifs_readpages_from_fscache()
134 unsigned *nr_pages) in cifs_readpages_from_fscache() argument
Dfscache.c187 unsigned *nr_pages) in __cifs_readpages_from_fscache() argument
192 __func__, CIFS_I(inode)->fscache, *nr_pages, inode); in __cifs_readpages_from_fscache()
194 pages, nr_pages, in __cifs_readpages_from_fscache()
/linux-4.4.14/fs/ntfs/
Dfile.c509 pgoff_t index, const unsigned nr_pages, struct page **pages, in __ntfs_grab_cache_pages() argument
514 BUG_ON(!nr_pages); in __ntfs_grab_cache_pages()
540 } while (nr < nr_pages); in __ntfs_grab_cache_pages()
585 unsigned nr_pages, s64 pos, size_t bytes) in ntfs_prepare_pages_for_non_resident_write() argument
613 BUG_ON(!nr_pages); in ntfs_prepare_pages_for_non_resident_write()
621 vi->i_ino, ni->type, pages[0]->index, nr_pages, in ntfs_prepare_pages_for_non_resident_write()
638 } while (++u < nr_pages); in ntfs_prepare_pages_for_non_resident_write()
1191 if (likely(!err && ++u < nr_pages)) in ntfs_prepare_pages_for_non_resident_write()
1237 } while (++u < nr_pages); in ntfs_prepare_pages_for_non_resident_write()
1345 nr_pages = u; in ntfs_prepare_pages_for_non_resident_write()
[all …]
Dcompress.c517 unsigned int nr_pages = (end_vcn - start_vcn) << in ntfs_read_compressed_block() local
526 "%i.", index, cb_size, nr_pages); in ntfs_read_compressed_block()
534 pages = kmalloc(nr_pages * sizeof(struct page *), GFP_NOFS); in ntfs_read_compressed_block()
575 if (nr_pages < max_page) in ntfs_read_compressed_block()
576 max_page = nr_pages; in ntfs_read_compressed_block()
/linux-4.4.14/arch/x86/mm/
Dnuma_32.c67 unsigned long nr_pages = end_pfn - start_pfn; in node_memmap_size_bytes() local
69 if (!nr_pages) in node_memmap_size_bytes()
72 return (nr_pages + 1) * sizeof(struct page); in node_memmap_size_bytes()
Dgup.c255 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, in __get_user_pages_fast() argument
267 len = (unsigned long) nr_pages << PAGE_SHIFT; in __get_user_pages_fast()
323 int get_user_pages_fast(unsigned long start, int nr_pages, int write, in get_user_pages_fast() argument
334 len = (unsigned long) nr_pages << PAGE_SHIFT; in get_user_pages_fast()
Dinit_64.c696 unsigned long nr_pages = size >> PAGE_SHIFT; in arch_add_memory() local
701 ret = __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory()
716 unsigned int nr_pages = 1 << order; in free_pagetable() local
724 while (nr_pages--) in free_pagetable()
727 while (nr_pages--) in free_pagetable()
1020 unsigned long nr_pages = size >> PAGE_SHIFT; in arch_remove_memory() local
1026 ret = __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
1318 unsigned int nr_pages; in register_page_bootmem_memmap() local
1358 nr_pages = 1 << (get_order(PMD_SIZE)); in register_page_bootmem_memmap()
1360 while (nr_pages--) in register_page_bootmem_memmap()
Dtlb.c117 unsigned long nr_pages = in flush_tlb_func() local
124 trace_tlb_flush(TLB_REMOTE_SHOOTDOWN, nr_pages); in flush_tlb_func()
Dinit_32.c832 unsigned long nr_pages = size >> PAGE_SHIFT; in arch_add_memory() local
834 return __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory()
841 unsigned long nr_pages = size >> PAGE_SHIFT; in arch_remove_memory() local
845 return __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
Dioremap.c56 static int __ioremap_check_ram(unsigned long start_pfn, unsigned long nr_pages, in __ioremap_check_ram() argument
61 for (i = 0; i < nr_pages; ++i) in __ioremap_check_ram()
/linux-4.4.14/drivers/media/pci/cx25821/
Dcx25821-alsa.c70 int nr_pages; member
147 static int cx25821_alsa_dma_init(struct cx25821_audio_dev *chip, int nr_pages) in cx25821_alsa_dma_init() argument
153 buf->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT); in cx25821_alsa_dma_init()
155 dprintk(1, "vmalloc_32(%d pages) failed\n", nr_pages); in cx25821_alsa_dma_init()
161 nr_pages << PAGE_SHIFT); in cx25821_alsa_dma_init()
163 memset(buf->vaddr, 0, nr_pages << PAGE_SHIFT); in cx25821_alsa_dma_init()
164 buf->nr_pages = nr_pages; in cx25821_alsa_dma_init()
166 buf->sglist = vzalloc(buf->nr_pages * sizeof(*buf->sglist)); in cx25821_alsa_dma_init()
170 sg_init_table(buf->sglist, buf->nr_pages); in cx25821_alsa_dma_init()
171 for (i = 0; i < buf->nr_pages; i++) { in cx25821_alsa_dma_init()
[all …]
/linux-4.4.14/arch/s390/pci/
Dpci_dma.c135 unsigned int nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; in dma_update_trans() local
142 if (!nr_pages) in dma_update_trans()
151 for (i = 0; i < nr_pages; i++) { in dma_update_trans()
175 nr_pages * PAGE_SIZE); in dma_update_trans()
288 unsigned long nr_pages, iommu_page_index; in s390_dma_map_pages() local
295 nr_pages = iommu_num_pages(pa, size, PAGE_SIZE); in s390_dma_map_pages()
296 iommu_page_index = dma_alloc_iommu(zdev, nr_pages); in s390_dma_map_pages()
303 size = nr_pages * PAGE_SIZE; in s390_dma_map_pages()
318 atomic64_add(nr_pages, &zdev->mapped_pages); in s390_dma_map_pages()
322 dma_free_iommu(zdev, iommu_page_index, nr_pages); in s390_dma_map_pages()
/linux-4.4.14/drivers/hwtracing/intel_th/
Dmsu.c113 unsigned long nr_pages; member
493 reg = msc->nr_pages; in msc_configure()
549 msc->single_sz = reg & ((msc->nr_pages << PAGE_SHIFT) - 1); in msc_disable()
624 msc->nr_pages = size >> PAGE_SHIFT; in msc_buffer_contig_alloc()
639 for (off = 0; off < msc->nr_pages << PAGE_SHIFT; off += PAGE_SIZE) { in msc_buffer_contig_free()
646 msc->nr_pages = 0; in msc_buffer_contig_free()
659 if (pgoff >= msc->nr_pages) in msc_buffer_contig_get_page()
719 msc->nr_pages += nr_blocks; in msc_buffer_win_alloc()
749 msc->nr_pages -= win->nr_blocks; in msc_buffer_win_free()
841 static int msc_buffer_multi_alloc(struct msc *msc, unsigned long *nr_pages, in msc_buffer_multi_alloc() argument
[all …]
/linux-4.4.14/net/9p/
Dtrans_common.c21 void p9_release_pages(struct page **pages, int nr_pages) in p9_release_pages() argument
25 for (i = 0; i < nr_pages; i++) in p9_release_pages()
Dtrans_virtio.c225 struct page **pdata, int nr_pages, size_t offs, int count) in pack_sg_list_p() argument
231 BUG_ON(nr_pages > (limit - start)); in pack_sg_list_p()
236 while (nr_pages) { in pack_sg_list_p()
245 nr_pages--; in pack_sg_list_p()
321 int nr_pages; in p9_get_mapped_pages() local
343 nr_pages = DIV_ROUND_UP(n + *offs, PAGE_SIZE); in p9_get_mapped_pages()
344 atomic_add(nr_pages, &vp_pinned); in p9_get_mapped_pages()
364 nr_pages = DIV_ROUND_UP((unsigned long)p + len, PAGE_SIZE) - in p9_get_mapped_pages()
367 *pages = kmalloc(sizeof(struct page *) * nr_pages, GFP_NOFS); in p9_get_mapped_pages()
373 for (index = 0; index < nr_pages; index++) { in p9_get_mapped_pages()
/linux-4.4.14/fs/btrfs/tests/
Dextent-io-tests.c35 unsigned long nr_pages = end_index - index + 1; in process_page_range() local
40 while (nr_pages > 0) { in process_page_range()
42 min_t(unsigned long, nr_pages, in process_page_range()
54 nr_pages -= ret; in process_page_range()
59 …KERN_ERR "stuck in a loop, start %Lu, end %Lu, nr_pages %lu, ret %d\n", start, end, nr_pages, ret); in process_page_range()
/linux-4.4.14/tools/vm/
Dpage-types.c202 static unsigned long nr_pages[HASH_SIZE]; variable
399 for (i = 0; i < ARRAY_SIZE(nr_pages); i++) { in show_summary()
400 if (nr_pages[i]) in show_summary()
403 nr_pages[i], in show_summary()
404 pages2mb(nr_pages[i]), in show_summary()
587 nr_pages[hash_slot(flags)]++; in add_page()
854 unsigned long nr_pages, pfn, i; in walk_file() local
864 nr_pages = (end - off + page_size - 1) / page_size; in walk_file()
865 if (nr_pages > PAGEMAP_BATCH) in walk_file()
866 nr_pages = PAGEMAP_BATCH; in walk_file()
[all …]
/linux-4.4.14/drivers/xen/xenbus/
Dxenbus_client.c381 unsigned int nr_pages, grant_ref_t *grefs) in xenbus_grant_ring() argument
386 for (i = 0; i < nr_pages; i++) { in xenbus_grant_ring()
634 unsigned int nr_pages = XENBUS_PAGES(nr_grefs); in xenbus_map_ring_valloc_hvm() local
645 err = alloc_xenballooned_pages(nr_pages, node->hvm.pages); in xenbus_map_ring_valloc_hvm()
660 addr = vmap(node->hvm.pages, nr_pages, VM_MAP | VM_IOREMAP, in xenbus_map_ring_valloc_hvm()
681 addr, nr_pages); in xenbus_map_ring_valloc_hvm()
684 free_xenballooned_pages(nr_pages, node->hvm.pages); in xenbus_map_ring_valloc_hvm()
836 unsigned int nr_pages; in xenbus_unmap_ring_vfree_hvm() local
856 nr_pages = XENBUS_PAGES(node->nr_handles); in xenbus_unmap_ring_vfree_hvm()
866 free_xenballooned_pages(nr_pages, node->hvm.pages); in xenbus_unmap_ring_vfree_hvm()
[all …]
/linux-4.4.14/include/xen/
Dballoon.h26 int alloc_xenballooned_pages(int nr_pages, struct page **pages);
27 void free_xenballooned_pages(int nr_pages, struct page **pages);
Dgrant_table.h195 int gnttab_alloc_pages(int nr_pages, struct page **pages);
196 void gnttab_free_pages(int nr_pages, struct page **pages);
/linux-4.4.14/drivers/lightnvm/
Drrpc.h96 unsigned long long nr_pages; member
197 BUG_ON((laddr + pages) > rrpc->nr_pages); in rrpc_lock_laddr()
232 uint8_t pages = rqd->nr_pages; in rrpc_unlock_rq()
234 BUG_ON((r->l_start + pages) > rrpc->nr_pages); in rrpc_unlock_rq()
Dgennvm.c323 if (rqd->nr_pages > 1) { in gennvm_addr_to_generic_mode()
324 for (i = 0; i < rqd->nr_pages; i++) in gennvm_addr_to_generic_mode()
336 if (rqd->nr_pages > 1) { in gennvm_generic_to_addr_mode()
337 for (i = 0; i < rqd->nr_pages; i++) in gennvm_generic_to_addr_mode()
396 if (rqd->nr_pages > 1) in gennvm_mark_blk_bad()
397 for (i = 0; i < rqd->nr_pages; i++) in gennvm_mark_blk_bad()
437 rqd.nr_pages = 1; in gennvm_erase_blk()
441 rqd.nr_pages = plane_cnt; in gennvm_erase_blk()
Drrpc.c515 BUG_ON(laddr >= rrpc->nr_pages); in rrpc_update_map()
647 uint8_t npages = rqd->nr_pages; in rrpc_end_io()
685 BUG_ON(!(laddr + i >= 0 && laddr + i < rrpc->nr_pages)); in rrpc_read_ppalist_rq()
716 BUG_ON(!(laddr >= 0 && laddr < rrpc->nr_pages)); in rrpc_read_rq()
823 uint8_t nr_pages = rrpc_get_pages(bio); in rrpc_submit_io() local
831 err = rrpc_setup_rq(rrpc, bio, rqd, flags, nr_pages); in rrpc_submit_io()
838 rqd->nr_pages = nr_pages; in rrpc_submit_io()
998 rrpc->trans_map = vzalloc(sizeof(struct rrpc_addr) * rrpc->nr_pages); in rrpc_map_init()
1003 * rrpc->nr_pages); in rrpc_map_init()
1007 for (i = 0; i < rrpc->nr_pages; i++) { in rrpc_map_init()
[all …]
/linux-4.4.14/fs/ext4/
Dreadpage.c135 unsigned nr_pages) in ext4_mpage_readpages() argument
160 for (page_idx = 0; nr_pages; page_idx++, nr_pages--) { in ext4_mpage_readpages()
177 last_block = block_in_file + nr_pages * blocks_per_page; in ext4_mpage_readpages()
287 min_t(int, nr_pages, BIO_MAX_PAGES)); in ext4_mpage_readpages()
Dfile.c461 unsigned long nr_pages; in ext4_find_unwritten_pgoff() local
464 nr_pages = pagevec_lookup(&pvec, inode->i_mapping, index, in ext4_find_unwritten_pgoff()
466 if (nr_pages == 0) { in ext4_find_unwritten_pgoff()
492 for (i = 0; i < nr_pages; i++) { in ext4_find_unwritten_pgoff()
550 if (nr_pages < num && whence == SEEK_HOLE) { in ext4_find_unwritten_pgoff()
/linux-4.4.14/block/
Dbio.c1129 int nr_pages = 0; in bio_copy_user_iov() local
1149 nr_pages += end - start; in bio_copy_user_iov()
1153 nr_pages++; in bio_copy_user_iov()
1170 bio = bio_kmalloc(gfp_mask, nr_pages); in bio_copy_user_iov()
1180 nr_pages = 1 << map_data->page_order; in bio_copy_user_iov()
1192 if (i == map_data->nr_entries * nr_pages) { in bio_copy_user_iov()
1197 page = map_data->pages[i / nr_pages]; in bio_copy_user_iov()
1198 page += (i % nr_pages); in bio_copy_user_iov()
1254 int nr_pages = 0; in bio_map_user_iov() local
1274 nr_pages += end - start; in bio_map_user_iov()
[all …]
Dbio-integrity.c276 unsigned int len, nr_pages; in bio_integrity_prep() local
297 nr_pages = end - start; in bio_integrity_prep()
300 bip = bio_integrity_alloc(bio, GFP_NOIO, nr_pages); in bio_integrity_prep()
316 for (i = 0 ; i < nr_pages ; i++) { in bio_integrity_prep()
/linux-4.4.14/kernel/trace/
Dring_buffer.c440 unsigned long nr_pages; member
1140 static int __rb_allocate_pages(long nr_pages, struct list_head *pages, int cpu) in __rb_allocate_pages() argument
1145 for (i = 0; i < nr_pages; i++) { in __rb_allocate_pages()
1180 unsigned long nr_pages) in rb_allocate_pages() argument
1184 WARN_ON(!nr_pages); in rb_allocate_pages()
1186 if (__rb_allocate_pages(nr_pages, &pages, cpu_buffer->cpu)) in rb_allocate_pages()
1197 cpu_buffer->nr_pages = nr_pages; in rb_allocate_pages()
1205 rb_allocate_cpu_buffer(struct ring_buffer *buffer, long nr_pages, int cpu) in rb_allocate_cpu_buffer() argument
1245 ret = rb_allocate_pages(cpu_buffer, nr_pages); in rb_allocate_cpu_buffer()
1305 long nr_pages; in __ring_buffer_alloc() local
[all …]
/linux-4.4.14/arch/sh/mm/
Dgup.c163 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, in __get_user_pages_fast() argument
175 len = (unsigned long) nr_pages << PAGE_SHIFT; in __get_user_pages_fast()
217 int get_user_pages_fast(unsigned long start, int nr_pages, int write, in get_user_pages_fast() argument
228 len = (unsigned long) nr_pages << PAGE_SHIFT; in get_user_pages_fast()
Dinit.c492 unsigned long nr_pages = size >> PAGE_SHIFT; in arch_add_memory() local
501 start_pfn, nr_pages); in arch_add_memory()
522 unsigned long nr_pages = size >> PAGE_SHIFT; in arch_remove_memory() local
527 ret = __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
/linux-4.4.14/arch/powerpc/mm/
Dmem.c121 unsigned long nr_pages = size >> PAGE_SHIFT; in arch_add_memory() local
133 return __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory()
140 unsigned long nr_pages = size >> PAGE_SHIFT; in arch_remove_memory() local
145 ret = __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
170 walk_system_ram_range(unsigned long start_pfn, unsigned long nr_pages, in walk_system_ram_range() argument
174 unsigned long end_pfn = start_pfn + nr_pages; in walk_system_ram_range()
Dinit_64.c394 unsigned int nr_pages = in vmemmap_free() local
396 while (nr_pages--) in vmemmap_free()
/linux-4.4.14/fs/afs/
Dcache.c375 int loop, nr_pages; in afs_vnode_cache_now_uncached() local
385 nr_pages = pagevec_lookup(&pvec, vnode->vfs_inode.i_mapping, in afs_vnode_cache_now_uncached()
388 if (!nr_pages) in afs_vnode_cache_now_uncached()
391 for (loop = 0; loop < nr_pages; loop++) in afs_vnode_cache_now_uncached()
394 first = pvec.pages[nr_pages - 1]->index + 1; in afs_vnode_cache_now_uncached()
396 pvec.nr = nr_pages; in afs_vnode_cache_now_uncached()
Dfile.c28 struct list_head *pages, unsigned nr_pages);
241 struct list_head *pages, unsigned nr_pages) in afs_readpages() argument
248 key_serial(key), mapping->host->i_ino, nr_pages); in afs_readpages()
263 &nr_pages, in afs_readpages()
275 BUG_ON(nr_pages != 0); in afs_readpages()
/linux-4.4.14/drivers/media/pci/cx88/
Dcx88-alsa.c68 int nr_pages; member
294 static int cx88_alsa_dma_init(struct cx88_audio_dev *chip, int nr_pages) in cx88_alsa_dma_init() argument
300 buf->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT); in cx88_alsa_dma_init()
302 dprintk(1, "vmalloc_32(%d pages) failed\n", nr_pages); in cx88_alsa_dma_init()
308 nr_pages << PAGE_SHIFT); in cx88_alsa_dma_init()
310 memset(buf->vaddr, 0, nr_pages << PAGE_SHIFT); in cx88_alsa_dma_init()
311 buf->nr_pages = nr_pages; in cx88_alsa_dma_init()
313 buf->sglist = vzalloc(buf->nr_pages * sizeof(*buf->sglist)); in cx88_alsa_dma_init()
317 sg_init_table(buf->sglist, buf->nr_pages); in cx88_alsa_dma_init()
318 for (i = 0; i < buf->nr_pages; i++) { in cx88_alsa_dma_init()
[all …]
/linux-4.4.14/arch/sparc/mm/
Dgup.c163 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, in __get_user_pages_fast() argument
174 len = (unsigned long) nr_pages << PAGE_SHIFT; in __get_user_pages_fast()
193 int get_user_pages_fast(unsigned long start, int nr_pages, int write, in get_user_pages_fast() argument
204 len = (unsigned long) nr_pages << PAGE_SHIFT; in get_user_pages_fast()
/linux-4.4.14/drivers/iommu/
Ds390-iommu.c218 unsigned long irq_flags, nr_pages, i; in s390_iommu_update_trans() local
226 nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; in s390_iommu_update_trans()
227 if (!nr_pages) in s390_iommu_update_trans()
231 for (i = 0; i < nr_pages; i++) { in s390_iommu_update_trans()
245 start_dma_addr, nr_pages * PAGE_SIZE); in s390_iommu_update_trans()
/linux-4.4.14/fs/cachefiles/
Drdwr.c677 unsigned *nr_pages, in cachefiles_read_or_alloc_pages() argument
696 *nr_pages); in cachefiles_read_or_alloc_pages()
702 if (cachefiles_has_space(cache, 0, *nr_pages) < 0) in cachefiles_read_or_alloc_pages()
745 (*nr_pages)--; in cachefiles_read_or_alloc_pages()
771 ret, *nr_pages, list_empty(pages) ? " empty" : ""); in cachefiles_read_or_alloc_pages()
775 fscache_retrieval_complete(op, *nr_pages); in cachefiles_read_or_alloc_pages()
829 unsigned *nr_pages, in cachefiles_allocate_pages() argument
843 _enter("%p,,,%d,", object, *nr_pages); in cachefiles_allocate_pages()
845 ret = cachefiles_has_space(cache, 0, *nr_pages); in cachefiles_allocate_pages()
861 fscache_retrieval_complete(op, *nr_pages); in cachefiles_allocate_pages()
/linux-4.4.14/drivers/media/common/saa7146/
Dsaa7146_core.c149 static struct scatterlist* vmalloc_to_sg(unsigned char *virt, int nr_pages) in vmalloc_to_sg() argument
155 sglist = kcalloc(nr_pages, sizeof(struct scatterlist), GFP_KERNEL); in vmalloc_to_sg()
158 sg_init_table(sglist, nr_pages); in vmalloc_to_sg()
159 for (i = 0; i < nr_pages; i++, virt += PAGE_SIZE) { in vmalloc_to_sg()
251 int nr_pages = 0; in saa7146_pgtable_build_single() local
270 nr_pages++; in saa7146_pgtable_build_single()
277 for(i=nr_pages;i<1024;i++) { in saa7146_pgtable_build_single()
/linux-4.4.14/drivers/gpu/drm/virtio/
Dvirtgpu_object.c124 int nr_pages = bo->tbo.num_pages; in virtio_gpu_object_get_sg_table() local
136 ret = sg_alloc_table_from_pages(bo->pages, pages, nr_pages, 0, in virtio_gpu_object_get_sg_table()
137 nr_pages << PAGE_SHIFT, GFP_KERNEL); in virtio_gpu_object_get_sg_table()
/linux-4.4.14/arch/mips/mm/
Dgup.c196 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, in __get_user_pages_fast() argument
208 len = (unsigned long) nr_pages << PAGE_SHIFT; in __get_user_pages_fast()
263 int get_user_pages_fast(unsigned long start, int nr_pages, int write, in get_user_pages_fast() argument
274 len = (unsigned long) nr_pages << PAGE_SHIFT; in get_user_pages_fast()
/linux-4.4.14/arch/s390/kvm/
Dgaccess.c714 unsigned long *pages, unsigned long nr_pages, in guest_page_range() argument
724 while (nr_pages) { in guest_page_range()
749 nr_pages--; in guest_page_range()
758 unsigned long _len, nr_pages, gpa, idx; in access_guest() local
770 nr_pages = (((ga & ~PAGE_MASK) + len - 1) >> PAGE_SHIFT) + 1; in access_guest()
772 if (nr_pages > ARRAY_SIZE(pages_array)) in access_guest()
773 pages = vmalloc(nr_pages * sizeof(unsigned long)); in access_guest()
779 rc = guest_page_range(vcpu, ga, pages, nr_pages, asce, write); in access_guest()
780 for (idx = 0; idx < nr_pages && !rc; idx++) { in access_guest()
793 if (nr_pages > ARRAY_SIZE(pages_array)) in access_guest()
/linux-4.4.14/kernel/
Drelay.c1215 unsigned int pidx, poff, total_len, subbuf_pages, nr_pages; in subbuf_splice_actor() local
1228 .nr_pages = 0, in subbuf_splice_actor()
1251 nr_pages = min_t(unsigned int, subbuf_pages, spd.nr_pages_max); in subbuf_splice_actor()
1253 for (total_len = 0; spd.nr_pages < nr_pages; spd.nr_pages++) { in subbuf_splice_actor()
1263 spd.pages[spd.nr_pages] = rbuf->page_array[pidx]; in subbuf_splice_actor()
1264 spd.partial[spd.nr_pages].offset = poff; in subbuf_splice_actor()
1271 spd.partial[spd.nr_pages].len = this_len; in subbuf_splice_actor()
1272 spd.partial[spd.nr_pages].private = private; in subbuf_splice_actor()
1280 spd.nr_pages++; in subbuf_splice_actor()
1286 if (!spd.nr_pages) in subbuf_splice_actor()
/linux-4.4.14/arch/alpha/mm/
Dinit.c192 unsigned long nr_pages = 0; in callback_init() local
198 nr_pages += crb->map[i].count; in callback_init()
202 console_remap_vm.size = nr_pages << PAGE_SHIFT; in callback_init()
/linux-4.4.14/drivers/media/pci/saa7134/
Dsaa7134-alsa.c268 static int saa7134_alsa_dma_init(struct saa7134_dev *dev, int nr_pages) in saa7134_alsa_dma_init() argument
274 dma->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT); in saa7134_alsa_dma_init()
276 pr_debug("vmalloc_32(%d pages) failed\n", nr_pages); in saa7134_alsa_dma_init()
282 nr_pages << PAGE_SHIFT); in saa7134_alsa_dma_init()
284 memset(dma->vaddr, 0, nr_pages << PAGE_SHIFT); in saa7134_alsa_dma_init()
285 dma->nr_pages = nr_pages; in saa7134_alsa_dma_init()
287 dma->sglist = vzalloc(dma->nr_pages * sizeof(*dma->sglist)); in saa7134_alsa_dma_init()
291 sg_init_table(dma->sglist, dma->nr_pages); in saa7134_alsa_dma_init()
292 for (i = 0; i < dma->nr_pages; i++) { in saa7134_alsa_dma_init()
314 dma->nr_pages, PCI_DMA_FROMDEVICE); in saa7134_alsa_dma_map()
/linux-4.4.14/drivers/video/fbdev/
Dpvr2fb.c679 unsigned int nr_pages; in pvr2fb_write() local
683 nr_pages = (count + PAGE_SIZE - 1) >> PAGE_SHIFT; in pvr2fb_write()
685 pages = kmalloc(nr_pages * sizeof(struct page *), GFP_KERNEL); in pvr2fb_write()
690 nr_pages, WRITE, 0, pages); in pvr2fb_write()
692 if (ret < nr_pages) { in pvr2fb_write()
693 nr_pages = ret; in pvr2fb_write()
702 end = (unsigned long)page_address(pages[nr_pages]); in pvr2fb_write()
703 len = nr_pages << PAGE_SHIFT; in pvr2fb_write()
721 for (i = 0; i < nr_pages; i++, dst += PAGE_SIZE) { in pvr2fb_write()
737 for (i = 0; i < nr_pages; i++) in pvr2fb_write()
Dxen-fbfront.c46 int nr_pages; member
403 info->nr_pages = (fb_size + PAGE_SIZE - 1) >> PAGE_SHIFT; in xenfb_probe()
405 info->gfns = vmalloc(sizeof(unsigned long) * info->nr_pages); in xenfb_probe()
551 for (i = 0; i < info->nr_pages; i++) in xenfb_init_shared_page()
554 for (i = 0; i * epd < info->nr_pages; i++) in xenfb_init_shared_page()
/linux-4.4.14/drivers/mtd/devices/
Dmtd_dataflash.c622 static int add_dataflash_otp(struct spi_device *spi, char *name, int nr_pages, in add_dataflash_otp() argument
648 device->size = nr_pages * pagesize; in add_dataflash_otp()
681 int nr_pages, int pagesize, int pageoffset) in add_dataflash() argument
683 return add_dataflash_otp(spi, name, nr_pages, pagesize, in add_dataflash()
696 unsigned nr_pages; member
839 return add_dataflash_otp(spi, info->name, info->nr_pages, in dataflash_probe()
Dsst25l.c60 unsigned nr_pages; member
384 flash->mtd.size = flash_info->page_size * flash_info->nr_pages; in sst25l_probe()
/linux-4.4.14/fs/gfs2/
Daops.c236 int nr_pages, pgoff_t end, in gfs2_write_jdata_pagevec() argument
241 unsigned nrblocks = nr_pages * (PAGE_CACHE_SIZE/inode->i_sb->s_blocksize); in gfs2_write_jdata_pagevec()
249 for(i = 0; i < nr_pages; i++) { in gfs2_write_jdata_pagevec()
350 int nr_pages; in gfs2_write_cache_jdata() local
385 nr_pages = pagevec_lookup_tag(&pvec, mapping, &index, tag, in gfs2_write_cache_jdata()
387 if (nr_pages == 0) in gfs2_write_cache_jdata()
390 ret = gfs2_write_jdata_pagevec(mapping, wbc, &pvec, nr_pages, end, &done_index); in gfs2_write_cache_jdata()
608 struct list_head *pages, unsigned nr_pages) in gfs2_readpages() argument
621 ret = mpage_readpages(mapping, pages, nr_pages, gfs2_block_map); in gfs2_readpages()
/linux-4.4.14/drivers/gpu/drm/exynos/
Dexynos_drm_fbdev.c84 unsigned int nr_pages; in exynos_drm_fbdev_update() local
100 nr_pages = exynos_gem->size >> PAGE_SHIFT; in exynos_drm_fbdev_update()
102 exynos_gem->kvaddr = (void __iomem *) vmap(exynos_gem->pages, nr_pages, in exynos_drm_fbdev_update()
Dexynos_drm_gem.c27 unsigned int nr_pages; in exynos_drm_alloc_buf() local
59 nr_pages = exynos_gem->size >> PAGE_SHIFT; in exynos_drm_alloc_buf()
61 exynos_gem->pages = drm_calloc_large(nr_pages, sizeof(struct page *)); in exynos_drm_alloc_buf()
84 nr_pages)) { in exynos_drm_alloc_buf()
/linux-4.4.14/drivers/base/
Dmemory.c228 unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; in memory_block_action() local
240 ret = online_pages(start_pfn, nr_pages, online_type); in memory_block_action()
243 ret = offline_pages(start_pfn, nr_pages); in memory_block_action()
391 unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; in show_valid_zones() local
396 end_pfn = start_pfn + nr_pages; in show_valid_zones()
Dfirmware_class.c149 int nr_pages; member
258 for (i = 0; i < buf->nr_pages; i++) in __fw_free_buf()
623 buf->data = vmap(buf->pages, buf->nr_pages, 0, PAGE_KERNEL_RO); in fw_map_pages_buf()
661 for (i = 0; i < fw_buf->nr_pages; i++) in firmware_loading_store()
666 fw_buf->nr_pages = 0; in firmware_loading_store()
788 while (buf->nr_pages < pages_needed) { in fw_realloc_buffer()
789 buf->pages[buf->nr_pages] = in fw_realloc_buffer()
792 if (!buf->pages[buf->nr_pages]) { in fw_realloc_buffer()
796 buf->nr_pages++; in fw_realloc_buffer()
/linux-4.4.14/fs/f2fs/
Ddata.c875 unsigned nr_pages) in f2fs_mpage_readpages() argument
895 for (page_idx = 0; nr_pages; page_idx++, nr_pages--) { in f2fs_mpage_readpages()
907 last_block = block_in_file + nr_pages; in f2fs_mpage_readpages()
976 min_t(int, nr_pages, BIO_MAX_PAGES)); in f2fs_mpage_readpages()
1031 struct list_head *pages, unsigned nr_pages) in f2fs_read_data_pages() argument
1036 trace_f2fs_readpages(inode, page, nr_pages); in f2fs_read_data_pages()
1042 return f2fs_mpage_readpages(mapping, pages, NULL, nr_pages); in f2fs_read_data_pages()
1214 int nr_pages; in f2fs_write_cache_pages() local
1252 nr_pages = pagevec_lookup_tag(&pvec, mapping, &index, tag, in f2fs_write_cache_pages()
1254 if (nr_pages == 0) in f2fs_write_cache_pages()
[all …]
/linux-4.4.14/include/media/
Dvideobuf-dma-sg.h66 int nr_pages; member
/linux-4.4.14/drivers/md/
Ddm-kcopyd.c285 static int client_reserve_pages(struct dm_kcopyd_client *kc, unsigned nr_pages) in client_reserve_pages() argument
290 for (i = 0; i < nr_pages; i++) { in client_reserve_pages()
301 kc->nr_reserved_pages += nr_pages; in client_reserve_pages()
521 unsigned nr_pages = dm_div_up(job->dests[0].count, PAGE_SIZE >> 9); in run_pages_job() local
523 r = kcopyd_get_pages(job->kc, nr_pages, &job->pages); in run_pages_job()
/linux-4.4.14/fs/proc/
Dkcore.c168 unsigned long nr_pages = ent->size >> PAGE_SHIFT; in get_sparsemem_vmemmap_info() local
174 end = ((unsigned long)pfn_to_page(pfn + nr_pages)) - 1; in get_sparsemem_vmemmap_info()
206 kclist_add_private(unsigned long pfn, unsigned long nr_pages, void *arg) in kclist_add_private() argument
215 ent->size = nr_pages << PAGE_SHIFT; in kclist_add_private()
Dtask_mmu.c1389 unsigned long nr_pages) in gather_stats() argument
1393 md->pages += nr_pages; in gather_stats()
1395 md->dirty += nr_pages; in gather_stats()
1398 md->swapcache += nr_pages; in gather_stats()
1401 md->active += nr_pages; in gather_stats()
1404 md->writeback += nr_pages; in gather_stats()
1407 md->anon += nr_pages; in gather_stats()
1412 md->node[page_to_nid(page)] += nr_pages; in gather_stats()
/linux-4.4.14/drivers/block/
Dbrd.c165 int nr_pages; in brd_free_pages() local
170 nr_pages = radix_tree_gang_lookup(&brd->brd_pages, in brd_free_pages()
173 for (i = 0; i < nr_pages; i++) { in brd_free_pages()
190 } while (nr_pages == FREE_BATCH); in brd_free_pages()
/linux-4.4.14/fs/fscache/
Dpage.c541 unsigned *nr_pages, in __fscache_read_or_alloc_pages() argument
551 _enter("%p,,%d,,,", cookie, *nr_pages); in __fscache_read_or_alloc_pages()
564 ASSERTCMP(*nr_pages, >, 0); in __fscache_read_or_alloc_pages()
573 atomic_set(&op->n_pages, *nr_pages); in __fscache_read_or_alloc_pages()
606 op, pages, nr_pages, gfp); in __fscache_read_or_alloc_pages()
611 op, pages, nr_pages, gfp); in __fscache_read_or_alloc_pages()
/linux-4.4.14/arch/arm64/mm/
Ddma-mapping.c311 unsigned long nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; in __swiotlb_mmap() local
321 if (off < nr_pages && nr_vma_pages <= (nr_pages - off)) { in __swiotlb_mmap()
364 unsigned long nr_pages = atomic_pool_size >> PAGE_SHIFT; in atomic_pool_init() local
370 page = dma_alloc_from_contiguous(NULL, nr_pages, in atomic_pool_init()
414 if (!dma_release_from_contiguous(NULL, page, nr_pages)) in atomic_pool_init()
/linux-4.4.14/arch/ia64/mm/
Dinit.c653 unsigned long nr_pages = size >> PAGE_SHIFT; in arch_add_memory() local
660 ret = __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory()
673 unsigned long nr_pages = size >> PAGE_SHIFT; in arch_remove_memory() local
678 ret = __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
/linux-4.4.14/fs/hpfs/
Dfile.c128 struct list_head *pages, unsigned nr_pages) in hpfs_readpages() argument
130 return mpage_readpages(mapping, pages, nr_pages, hpfs_get_block); in hpfs_readpages()
/linux-4.4.14/arch/arm/include/asm/xen/
Dpage.h107 unsigned long nr_pages);
/linux-4.4.14/drivers/net/ethernet/broadcom/bnxt/
Dbnxt_sriov.c305 u32 nr_pages, size, i, j, k = 0; in bnxt_alloc_vf_resources() local
314 nr_pages = size / BNXT_PAGE_SIZE; in bnxt_alloc_vf_resources()
316 nr_pages++; in bnxt_alloc_vf_resources()
318 for (i = 0; i < nr_pages; i++) { in bnxt_alloc_vf_resources()
344 bp->pf.hwrm_cmd_req_pages = nr_pages; in bnxt_alloc_vf_resources()
/linux-4.4.14/arch/powerpc/platforms/pseries/
Dcmm.c476 unsigned long end = start + (marg->nr_pages << PAGE_SHIFT); in cmm_count_pages()
529 unsigned long end_page = start_page + (marg->nr_pages << PAGE_SHIFT); in cmm_mem_going_offline()
535 start_page, marg->nr_pages); in cmm_mem_going_offline()
/linux-4.4.14/drivers/gpu/drm/savage/
Dsavage_bci.c367 unsigned int nr_pages = (n - rest + SAVAGE_DMA_PAGE_SIZE - 1) / in savage_dma_alloc() local
373 cur, dev_priv->dma_pages[cur].used, n, rest, nr_pages); in savage_dma_alloc()
375 if (cur + nr_pages < dev_priv->nr_dma_pages) { in savage_dma_alloc()
385 nr_pages = in savage_dma_alloc()
395 for (i = cur; nr_pages > 0; ++i, --nr_pages) { in savage_dma_alloc()
/linux-4.4.14/arch/sh/kernel/cpu/sh4/
Dsq.c375 unsigned int nr_pages = 0x04000000 >> PAGE_SHIFT; in sq_api_init() local
376 unsigned int size = (nr_pages + (BITS_PER_LONG - 1)) / BITS_PER_LONG; in sq_api_init()
/linux-4.4.14/fs/xfs/
Dxfs_buf.c1138 int nr_pages; in xfs_buf_ioapply_map() local
1164 nr_pages = BIO_MAX_SECTORS >> (PAGE_SHIFT - BBSHIFT); in xfs_buf_ioapply_map()
1165 if (nr_pages > total_nr_pages) in xfs_buf_ioapply_map()
1166 nr_pages = total_nr_pages; in xfs_buf_ioapply_map()
1168 bio = bio_alloc(GFP_NOIO, nr_pages); in xfs_buf_ioapply_map()
1175 for (; size && nr_pages; nr_pages--, page_index++) { in xfs_buf_ioapply_map()
/linux-4.4.14/include/linux/ceph/
Dosd_client.h358 struct page **pages, int nr_pages,
368 struct page **pages, int nr_pages);
/linux-4.4.14/fs/jfs/
Dinode.c300 struct list_head *pages, unsigned nr_pages) in jfs_readpages() argument
302 return mpage_readpages(mapping, pages, nr_pages, jfs_get_block); in jfs_readpages()
/linux-4.4.14/drivers/gpu/drm/
Ddrm_prime.c678 struct sg_table *drm_prime_pages_to_sg(struct page **pages, unsigned int nr_pages) in drm_prime_pages_to_sg() argument
689 ret = sg_alloc_table_from_pages(sg, pages, nr_pages, 0, in drm_prime_pages_to_sg()
690 nr_pages << PAGE_SHIFT, GFP_KERNEL); in drm_prime_pages_to_sg()
/linux-4.4.14/Documentation/ABI/testing/
Dsysfs-bus-intel_th-devices-msc20 What: /sys/bus/intel_th/devices/<intel_th_id>-msc<msc-id>/nr_pages

12