Home
last modified time | relevance | path

Searched refs:iter (Results 1 – 200 of 369) sorted by relevance

12

/linux-4.1.27/block/
Dt10-pi.c49 static int t10_pi_generate(struct blk_integrity_iter *iter, csum_fn *fn, in t10_pi_generate() argument
54 for (i = 0 ; i < iter->data_size ; i += iter->interval) { in t10_pi_generate()
55 struct t10_pi_tuple *pi = iter->prot_buf; in t10_pi_generate()
57 pi->guard_tag = fn(iter->data_buf, iter->interval); in t10_pi_generate()
61 pi->ref_tag = cpu_to_be32(lower_32_bits(iter->seed)); in t10_pi_generate()
65 iter->data_buf += iter->interval; in t10_pi_generate()
66 iter->prot_buf += sizeof(struct t10_pi_tuple); in t10_pi_generate()
67 iter->seed++; in t10_pi_generate()
73 static int t10_pi_verify(struct blk_integrity_iter *iter, csum_fn *fn, in t10_pi_verify() argument
78 for (i = 0 ; i < iter->data_size ; i += iter->interval) { in t10_pi_verify()
[all …]
Dblk-map.c65 const struct iov_iter *iter, gfp_t gfp_mask) in blk_rq_map_user_iov() argument
72 if (!iter || !iter->count) in blk_rq_map_user_iov()
75 iov_for_each(iov, i, *iter) { in blk_rq_map_user_iov()
88 if (unaligned || (q->dma_pad_mask & iter->count) || map_data) in blk_rq_map_user_iov()
89 bio = bio_copy_user_iov(q, map_data, iter, gfp_mask); in blk_rq_map_user_iov()
91 bio = bio_map_user_iov(q, iter, gfp_mask); in blk_rq_map_user_iov()
99 if (bio->bi_iter.bi_size != iter->count) { in blk_rq_map_user_iov()
Dbio.c506 struct bvec_iter iter; in zero_fill_bio() local
508 bio_for_each_segment(bv, bio, iter) { in zero_fill_bio()
618 struct bvec_iter iter; in bio_clone_bioset() local
661 bio_for_each_segment(bv, bio_src, iter) in bio_clone_bioset()
1025 struct iov_iter iter; member
1047 static int bio_copy_from_iter(struct bio *bio, struct iov_iter iter) in bio_copy_from_iter() argument
1058 &iter); in bio_copy_from_iter()
1060 if (!iov_iter_count(&iter)) in bio_copy_from_iter()
1078 static int bio_copy_to_iter(struct bio *bio, struct iov_iter iter) in bio_copy_to_iter() argument
1089 &iter); in bio_copy_to_iter()
[all …]
Dbio-integrity.c218 struct blk_integrity_iter iter; in bio_integrity_process() local
226 iter.disk_name = bio->bi_bdev->bd_disk->disk_name; in bio_integrity_process()
227 iter.interval = bi->interval; in bio_integrity_process()
228 iter.seed = bip_get_seed(bip); in bio_integrity_process()
229 iter.prot_buf = prot_buf; in bio_integrity_process()
234 iter.data_buf = kaddr + bv.bv_offset; in bio_integrity_process()
235 iter.data_size = bv.bv_len; in bio_integrity_process()
237 ret = proc_fn(&iter); in bio_integrity_process()
Dgenhd.c740 struct class_dev_iter iter; in printk_all_partitions() local
743 class_dev_iter_init(&iter, &block_class, NULL, &disk_type); in printk_all_partitions()
744 while ((dev = class_dev_iter_next(&iter))) { in printk_all_partitions()
785 class_dev_iter_exit(&iter); in printk_all_partitions()
793 struct class_dev_iter *iter; in disk_seqf_start() local
796 iter = kmalloc(sizeof(*iter), GFP_KERNEL); in disk_seqf_start()
797 if (!iter) in disk_seqf_start()
800 seqf->private = iter; in disk_seqf_start()
801 class_dev_iter_init(iter, &block_class, NULL, &disk_type); in disk_seqf_start()
803 dev = class_dev_iter_next(iter); in disk_seqf_start()
[all …]
Dbounce.c105 struct bvec_iter iter; in copy_to_high_bio_irq() local
107 bio_for_each_segment(tovec, to, iter) { in copy_to_high_bio_irq()
205 struct bvec_iter iter; in __blk_queue_bounce() local
210 bio_for_each_segment(from, *bio_orig, iter) in __blk_queue_bounce()
Dblk-merge.c20 struct bvec_iter iter; in __blk_recalc_rq_segments() local
41 bio_for_each_segment(bv, bio, iter) { in __blk_recalc_rq_segments()
127 struct bvec_iter iter; in blk_phys_contig_segment() local
139 bio_for_each_segment(end_bv, bio, iter) in blk_phys_contig_segment()
140 if (end_bv.bv_len == iter.bi_size) in blk_phys_contig_segment()
206 struct bvec_iter iter; in __blk_bios_map_sg() local
236 bio_for_each_segment(bvec, bio, iter) in __blk_bios_map_sg()
Dblk-integrity.c49 struct bvec_iter iter; in blk_rq_count_integrity_sg() local
52 bio_for_each_integrity_vec(iv, bio, iter) { in blk_rq_count_integrity_sg()
95 struct bvec_iter iter; in blk_rq_map_integrity_sg() local
98 bio_for_each_integrity_vec(iv, bio, iter) { in blk_rq_map_integrity_sg()
/linux-4.1.27/net/netlabel/
Dnetlabel_addrlist.c63 struct netlbl_af4list *iter; in netlbl_af4list_search() local
65 list_for_each_entry_rcu(iter, head, list) in netlbl_af4list_search()
66 if (iter->valid && (addr & iter->mask) == iter->addr) in netlbl_af4list_search()
67 return iter; in netlbl_af4list_search()
88 struct netlbl_af4list *iter; in netlbl_af4list_search_exact() local
90 list_for_each_entry_rcu(iter, head, list) in netlbl_af4list_search_exact()
91 if (iter->valid && iter->addr == addr && iter->mask == mask) in netlbl_af4list_search_exact()
92 return iter; in netlbl_af4list_search_exact()
113 struct netlbl_af6list *iter; in netlbl_af6list_search() local
115 list_for_each_entry_rcu(iter, head, list) in netlbl_af6list_search()
[all …]
Dnetlabel_addrlist.h96 #define netlbl_af4list_foreach(iter, head) \ argument
97 for (iter = __af4list_valid((head)->next, head); \
98 &iter->list != (head); \
99 iter = __af4list_valid(iter->list.next, head))
101 #define netlbl_af4list_foreach_rcu(iter, head) \ argument
102 for (iter = __af4list_valid_rcu((head)->next, head); \
103 &iter->list != (head); \
104 iter = __af4list_valid_rcu(iter->list.next, head))
106 #define netlbl_af4list_foreach_safe(iter, tmp, head) \ argument
107 for (iter = __af4list_valid((head)->next, head), \
[all …]
Dnetlabel_kapi.c431 struct netlbl_lsm_catmap *iter = *catmap; in _netlbl_catmap_getnode() local
434 if (iter == NULL) in _netlbl_catmap_getnode()
436 if (offset < iter->startbit) in _netlbl_catmap_getnode()
438 while (iter && offset >= (iter->startbit + NETLBL_CATMAP_SIZE)) { in _netlbl_catmap_getnode()
439 prev = iter; in _netlbl_catmap_getnode()
440 iter = iter->next; in _netlbl_catmap_getnode()
442 if (iter == NULL || offset < iter->startbit) in _netlbl_catmap_getnode()
445 return iter; in _netlbl_catmap_getnode()
449 return iter; in _netlbl_catmap_getnode()
454 iter = netlbl_catmap_alloc(gfp_flags); in _netlbl_catmap_getnode()
[all …]
Dnetlabel_cipso_v4.c105 u32 iter = 0; in netlbl_cipsov4_add_common() local
116 if (iter >= CIPSO_V4_TAG_MAXCNT) in netlbl_cipsov4_add_common()
118 doi_def->tags[iter++] = nla_get_u8(nla); in netlbl_cipsov4_add_common()
120 while (iter < CIPSO_V4_TAG_MAXCNT) in netlbl_cipsov4_add_common()
121 doi_def->tags[iter++] = CIPSO_V4_TAG_INVALID; in netlbl_cipsov4_add_common()
150 u32 iter; in netlbl_cipsov4_add_std() local
220 for (iter = 0; iter < doi_def->map.std->lvl.local_size; iter++) in netlbl_cipsov4_add_std()
221 doi_def->map.std->lvl.local[iter] = CIPSO_V4_INV_LVL; in netlbl_cipsov4_add_std()
222 for (iter = 0; iter < doi_def->map.std->lvl.cipso_size; iter++) in netlbl_cipsov4_add_std()
223 doi_def->map.std->lvl.cipso[iter] = CIPSO_V4_INV_LVL; in netlbl_cipsov4_add_std()
[all …]
Dnetlabel_domainhash.c117 u32 iter; in netlbl_domhsh_hash() local
124 for (iter = 0, val = 0, len = strlen(key); iter < len; iter++) in netlbl_domhsh_hash()
125 val = (val << 4 | (val >> (8 * sizeof(u32) - 4))) ^ key[iter]; in netlbl_domhsh_hash()
144 struct netlbl_dom_map *iter; in netlbl_domhsh_search() local
149 list_for_each_entry_rcu(iter, bkt_list, list) in netlbl_domhsh_search()
150 if (iter->valid && strcmp(iter->domain, domain) == 0) in netlbl_domhsh_search()
151 return iter; in netlbl_domhsh_search()
327 u32 iter; in netlbl_domhsh_init() local
344 for (iter = 0; iter < hsh_tbl->size; iter++) in netlbl_domhsh_init()
345 INIT_LIST_HEAD(&hsh_tbl->tbl[iter]); in netlbl_domhsh_init()
Dnetlabel_domainhash.h56 #define netlbl_domhsh_addr4_entry(iter) \ argument
57 container_of(iter, struct netlbl_domaddr4_map, list)
63 #define netlbl_domhsh_addr6_entry(iter) \ argument
64 container_of(iter, struct netlbl_domaddr6_map, list)
Dnetlabel_unlabeled.c80 #define netlbl_unlhsh_addr4_entry(iter) \ argument
81 container_of(iter, struct netlbl_unlhsh_addr4, list)
88 #define netlbl_unlhsh_addr6_entry(iter) \ argument
89 container_of(iter, struct netlbl_unlhsh_addr6, list)
226 struct netlbl_unlhsh_iface *iter; in netlbl_unlhsh_search_iface() local
230 list_for_each_entry_rcu(iter, bkt_list, list) in netlbl_unlhsh_search_iface()
231 if (iter->valid && iter->ifindex == ifindex) in netlbl_unlhsh_search_iface()
232 return iter; in netlbl_unlhsh_search_iface()
1420 u32 iter; in netlbl_unlabel_init() local
1437 for (iter = 0; iter < hsh_tbl->size; iter++) in netlbl_unlabel_init()
[all …]
/linux-4.1.27/kernel/gcov/
Dgcc_3_4.c346 static struct gcov_fn_info *get_func(struct gcov_iterator *iter) in get_func() argument
348 return get_fn_info(iter->info, iter->function); in get_func()
351 static struct type_info *get_type(struct gcov_iterator *iter) in get_type() argument
353 return &iter->type_info[iter->type]; in get_type()
364 struct gcov_iterator *iter; in gcov_iter_new() local
366 iter = kzalloc(sizeof(struct gcov_iterator) + in gcov_iter_new()
369 if (iter) in gcov_iter_new()
370 iter->info = info; in gcov_iter_new()
372 return iter; in gcov_iter_new()
379 void gcov_iter_free(struct gcov_iterator *iter) in gcov_iter_free() argument
[all …]
Dgcc_4_7.c476 struct gcov_iterator *iter; in gcov_iter_new() local
478 iter = kzalloc(sizeof(struct gcov_iterator), GFP_KERNEL); in gcov_iter_new()
479 if (!iter) in gcov_iter_new()
482 iter->info = info; in gcov_iter_new()
484 iter->size = convert_to_gcda(NULL, info); in gcov_iter_new()
485 iter->buffer = vmalloc(iter->size); in gcov_iter_new()
486 if (!iter->buffer) in gcov_iter_new()
489 convert_to_gcda(iter->buffer, info); in gcov_iter_new()
491 return iter; in gcov_iter_new()
494 kfree(iter); in gcov_iter_new()
[all …]
Dgcov.h63 void gcov_iter_free(struct gcov_iterator *iter);
64 void gcov_iter_start(struct gcov_iterator *iter);
65 int gcov_iter_next(struct gcov_iterator *iter);
66 int gcov_iter_write(struct gcov_iterator *iter, struct seq_file *seq);
67 struct gcov_info *gcov_iter_get_info(struct gcov_iterator *iter);
Dfs.c109 struct gcov_iterator *iter = data; in gcov_seq_next() local
111 if (gcov_iter_next(iter)) in gcov_seq_next()
115 return iter; in gcov_seq_next()
121 struct gcov_iterator *iter = data; in gcov_seq_show() local
123 if (gcov_iter_write(iter, seq)) in gcov_seq_show()
181 struct gcov_iterator *iter; in gcov_seq_open() local
195 iter = gcov_iter_new(info); in gcov_seq_open()
196 if (!iter) in gcov_seq_open()
202 seq->private = iter; in gcov_seq_open()
208 gcov_iter_free(iter); in gcov_seq_open()
[all …]
/linux-4.1.27/drivers/gpu/drm/
Ddrm_vma_manager.c162 struct rb_node *iter; in drm_vma_offset_lookup_locked() local
165 iter = mgr->vm_addr_space_rb.rb_node; in drm_vma_offset_lookup_locked()
168 while (likely(iter)) { in drm_vma_offset_lookup_locked()
169 node = rb_entry(iter, struct drm_vma_offset_node, vm_rb); in drm_vma_offset_lookup_locked()
172 iter = iter->rb_right; in drm_vma_offset_lookup_locked()
177 iter = iter->rb_left; in drm_vma_offset_lookup_locked()
196 struct rb_node **iter = &mgr->vm_addr_space_rb.rb_node; in _drm_vma_offset_add_rb() local
200 while (likely(*iter)) { in _drm_vma_offset_add_rb()
201 parent = *iter; in _drm_vma_offset_add_rb()
202 iter_node = rb_entry(*iter, struct drm_vma_offset_node, vm_rb); in _drm_vma_offset_add_rb()
[all …]
/linux-4.1.27/include/linux/
Dradix-tree.h341 radix_tree_iter_init(struct radix_tree_iter *iter, unsigned long start) in radix_tree_iter_init() argument
351 iter->index = 0; in radix_tree_iter_init()
352 iter->next_index = start; in radix_tree_iter_init()
370 struct radix_tree_iter *iter, unsigned flags);
382 void **radix_tree_iter_retry(struct radix_tree_iter *iter) in radix_tree_iter_retry() argument
384 iter->next_index = iter->index; in radix_tree_iter_retry()
395 radix_tree_chunk_size(struct radix_tree_iter *iter) in radix_tree_chunk_size() argument
397 return iter->next_index - iter->index; in radix_tree_chunk_size()
412 radix_tree_next_slot(void **slot, struct radix_tree_iter *iter, unsigned flags) in radix_tree_next_slot() argument
415 iter->tags >>= 1; in radix_tree_next_slot()
[all …]
Dbio.h64 #define __bvec_iter_bvec(bvec, iter) (&(bvec)[(iter).bi_idx]) argument
66 #define bvec_iter_page(bvec, iter) \ argument
67 (__bvec_iter_bvec((bvec), (iter))->bv_page)
69 #define bvec_iter_len(bvec, iter) \ argument
70 min((iter).bi_size, \
71 __bvec_iter_bvec((bvec), (iter))->bv_len - (iter).bi_bvec_done)
73 #define bvec_iter_offset(bvec, iter) \ argument
74 (__bvec_iter_bvec((bvec), (iter))->bv_offset + (iter).bi_bvec_done)
76 #define bvec_iter_bvec(bvec, iter) \ argument
78 .bv_page = bvec_iter_page((bvec), (iter)), \
[all …]
Duio.h57 static inline struct iovec iov_iter_iovec(const struct iov_iter *iter) in iov_iter_iovec() argument
60 .iov_base = iter->iov->iov_base + iter->iov_offset, in iov_iter_iovec()
61 .iov_len = min(iter->count, in iov_iter_iovec()
62 iter->iov->iov_len - iter->iov_offset), in iov_iter_iovec()
66 #define iov_for_each(iov, iter, start) \ argument
68 for (iter = (start); \
69 (iter).count && \
70 ((iov = iov_iter_iovec(&(iter))), 1); \
71 iov_iter_advance(&(iter), (iov).iov_len))
Dring_buffer.h130 void ring_buffer_read_start(struct ring_buffer_iter *iter);
131 void ring_buffer_read_finish(struct ring_buffer_iter *iter);
134 ring_buffer_iter_peek(struct ring_buffer_iter *iter, u64 *ts);
136 ring_buffer_read(struct ring_buffer_iter *iter, u64 *ts);
137 void ring_buffer_iter_reset(struct ring_buffer_iter *iter);
138 int ring_buffer_iter_empty(struct ring_buffer_iter *iter);
Dftrace.h417 struct ftrace_rec_iter *ftrace_rec_iter_next(struct ftrace_rec_iter *iter);
418 struct dyn_ftrace *ftrace_rec_iter_record(struct ftrace_rec_iter *iter);
420 #define for_ftrace_rec_iter(iter) \ argument
421 for (iter = ftrace_rec_iter_start(); \
422 iter; \
423 iter = ftrace_rec_iter_next(iter))
Dnetdevice.h3484 struct list_head **iter);
3486 struct list_head **iter);
3489 #define netdev_for_each_upper_dev_rcu(dev, updev, iter) \ argument
3490 for (iter = &(dev)->adj_list.upper, \
3491 updev = netdev_upper_get_next_dev_rcu(dev, &(iter)); \
3493 updev = netdev_upper_get_next_dev_rcu(dev, &(iter)))
3496 #define netdev_for_each_all_upper_dev_rcu(dev, updev, iter) \ argument
3497 for (iter = &(dev)->all_adj_list.upper, \
3498 updev = netdev_all_upper_get_next_dev_rcu(dev, &(iter)); \
3500 updev = netdev_all_upper_get_next_dev_rcu(dev, &(iter)))
[all …]
Dpnfs_osd_xdr.h294 struct pnfs_osd_xdr_decode_layout_iter *iter, struct xdr_stream *xdr);
297 struct pnfs_osd_xdr_decode_layout_iter *iter, struct xdr_stream *xdr,
Dftrace_event.h55 int ftrace_raw_output_prep(struct trace_iterator *iter,
118 typedef enum print_line_t (*trace_print_func)(struct trace_iterator *iter,
186 int ftrace_output_call(struct trace_iterator *iter, char *name, char *fmt, ...);
222 int ftrace_output_event(struct trace_iterator *iter, struct ftrace_event_call *event,
Drhashtable.h347 int rhashtable_walk_init(struct rhashtable *ht, struct rhashtable_iter *iter);
348 void rhashtable_walk_exit(struct rhashtable_iter *iter);
349 int rhashtable_walk_start(struct rhashtable_iter *iter) __acquires(RCU);
350 void *rhashtable_walk_next(struct rhashtable_iter *iter);
351 void rhashtable_walk_stop(struct rhashtable_iter *iter) __releases(RCU);
/linux-4.1.27/kernel/trace/
Dtrace_kdb.c22 static struct trace_iterator iter; in ftrace_dump_buf() local
27 trace_init_global_iter(&iter); in ftrace_dump_buf()
28 iter.buffer_iter = buffer_iter; in ftrace_dump_buf()
31 atomic_inc(&per_cpu_ptr(iter.trace_buffer->data, cpu)->disabled); in ftrace_dump_buf()
42 memset(&iter.seq, 0, in ftrace_dump_buf()
45 iter.iter_flags |= TRACE_FILE_LAT_FMT; in ftrace_dump_buf()
46 iter.pos = -1; in ftrace_dump_buf()
50 iter.buffer_iter[cpu] = in ftrace_dump_buf()
51 ring_buffer_read_prepare(iter.trace_buffer->buffer, cpu); in ftrace_dump_buf()
52 ring_buffer_read_start(iter.buffer_iter[cpu]); in ftrace_dump_buf()
[all …]
Dtrace_output.c23 enum print_line_t trace_print_bputs_msg_only(struct trace_iterator *iter) in trace_print_bputs_msg_only() argument
25 struct trace_seq *s = &iter->seq; in trace_print_bputs_msg_only()
26 struct trace_entry *entry = iter->ent; in trace_print_bputs_msg_only()
36 enum print_line_t trace_print_bprintk_msg_only(struct trace_iterator *iter) in trace_print_bprintk_msg_only() argument
38 struct trace_seq *s = &iter->seq; in trace_print_bprintk_msg_only()
39 struct trace_entry *entry = iter->ent; in trace_print_bprintk_msg_only()
49 enum print_line_t trace_print_printk_msg_only(struct trace_iterator *iter) in trace_print_printk_msg_only() argument
51 struct trace_seq *s = &iter->seq; in trace_print_printk_msg_only()
52 struct trace_entry *entry = iter->ent; in trace_print_printk_msg_only()
225 int ftrace_raw_output_prep(struct trace_iterator *iter, in ftrace_raw_output_prep() argument
[all …]
Dtrace.c1130 static int wait_on_pipe(struct trace_iterator *iter, bool full) in wait_on_pipe() argument
1133 if (trace_buffer_iter(iter, iter->cpu_file)) in wait_on_pipe()
1136 return ring_buffer_wait(iter->trace_buffer->buffer, iter->cpu_file, in wait_on_pipe()
2278 static void trace_iterator_increment(struct trace_iterator *iter) in trace_iterator_increment() argument
2280 struct ring_buffer_iter *buf_iter = trace_buffer_iter(iter, iter->cpu); in trace_iterator_increment()
2282 iter->idx++; in trace_iterator_increment()
2288 peek_next_entry(struct trace_iterator *iter, int cpu, u64 *ts, in peek_next_entry() argument
2292 struct ring_buffer_iter *buf_iter = trace_buffer_iter(iter, cpu); in peek_next_entry()
2297 event = ring_buffer_peek(iter->trace_buffer->buffer, cpu, ts, in peek_next_entry()
2301 iter->ent_size = ring_buffer_event_length(event); in peek_next_entry()
[all …]
Dtrace_functions_graph.c571 get_return_for_leaf(struct trace_iterator *iter, in get_return_for_leaf() argument
574 struct fgraph_data *data = iter->private; in get_return_for_leaf()
588 ring_iter = trace_buffer_iter(iter, iter->cpu); in get_return_for_leaf()
598 ring_buffer_consume(iter->trace_buffer->buffer, iter->cpu, in get_return_for_leaf()
600 event = ring_buffer_peek(iter->trace_buffer->buffer, iter->cpu, in get_return_for_leaf()
653 print_graph_irq(struct trace_iterator *iter, unsigned long addr, in print_graph_irq() argument
656 struct trace_seq *s = &iter->seq; in print_graph_irq()
657 struct trace_entry *ent = iter->ent; in print_graph_irq()
666 print_graph_abs_time(iter->ts, s); in print_graph_irq()
761 print_graph_entry_leaf(struct trace_iterator *iter, in print_graph_entry_leaf() argument
[all …]
Dtrace_mmiotrace.c102 static void mmio_pipe_open(struct trace_iterator *iter) in mmio_pipe_open() argument
105 struct trace_seq *s = &iter->seq; in mmio_pipe_open()
114 iter->private = hiter; in mmio_pipe_open()
118 static void mmio_close(struct trace_iterator *iter) in mmio_close() argument
120 struct header_iter *hiter = iter->private; in mmio_close()
122 iter->private = NULL; in mmio_close()
125 static unsigned long count_overruns(struct trace_iterator *iter) in count_overruns() argument
128 unsigned long over = ring_buffer_overruns(iter->trace_buffer->buffer); in count_overruns()
136 static ssize_t mmio_read(struct trace_iterator *iter, struct file *filp, in mmio_read() argument
140 struct header_iter *hiter = iter->private; in mmio_read()
[all …]
Dtrace_output.h8 trace_print_bputs_msg_only(struct trace_iterator *iter);
10 trace_print_bprintk_msg_only(struct trace_iterator *iter);
12 trace_print_printk_msg_only(struct trace_iterator *iter);
22 extern int trace_print_context(struct trace_iterator *iter);
23 extern int trace_print_lat_context(struct trace_iterator *iter);
29 extern enum print_line_t trace_nop_print(struct trace_iterator *iter,
Dftrace.c2357 struct ftrace_rec_iter *iter = &ftrace_rec_iter; in ftrace_rec_iter_start() local
2359 iter->pg = ftrace_pages_start; in ftrace_rec_iter_start()
2360 iter->index = 0; in ftrace_rec_iter_start()
2363 while (iter->pg && !iter->pg->index) in ftrace_rec_iter_start()
2364 iter->pg = iter->pg->next; in ftrace_rec_iter_start()
2366 if (!iter->pg) in ftrace_rec_iter_start()
2369 return iter; in ftrace_rec_iter_start()
2378 struct ftrace_rec_iter *ftrace_rec_iter_next(struct ftrace_rec_iter *iter) in ftrace_rec_iter_next() argument
2380 iter->index++; in ftrace_rec_iter_next()
2382 if (iter->index >= iter->pg->index) { in ftrace_rec_iter_next()
[all …]
Dtrace_printk.c49 const char **iter; in hold_module_trace_bprintk_format() local
57 for (iter = start; iter < end; iter++) { in hold_module_trace_bprintk_format()
58 struct trace_bprintk_fmt *tb_fmt = lookup_format(*iter); in hold_module_trace_bprintk_format()
60 *iter = tb_fmt->fmt; in hold_module_trace_bprintk_format()
67 fmt = kmalloc(strlen(*iter) + 1, GFP_KERNEL); in hold_module_trace_bprintk_format()
70 strcpy(fmt, *iter); in hold_module_trace_bprintk_format()
75 *iter = fmt; in hold_module_trace_bprintk_format()
Dring_buffer.c1884 rb_iter_head_event(struct ring_buffer_iter *iter) in rb_iter_head_event() argument
1886 return __rb_page_index(iter->head_page, iter->head); in rb_iter_head_event()
1987 static void rb_inc_iter(struct ring_buffer_iter *iter) in rb_inc_iter() argument
1989 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in rb_inc_iter()
1997 if (iter->head_page == cpu_buffer->reader_page) in rb_inc_iter()
1998 iter->head_page = rb_set_head_page(cpu_buffer); in rb_inc_iter()
2000 rb_inc_page(cpu_buffer, &iter->head_page); in rb_inc_iter()
2002 iter->read_stamp = iter->head_page->page->time_stamp; in rb_inc_iter()
2003 iter->head = 0; in rb_inc_iter()
3402 static void rb_iter_reset(struct ring_buffer_iter *iter) in rb_iter_reset() argument
[all …]
Dtrace.h363 void (*open)(struct trace_iterator *iter);
364 void (*pipe_open)(struct trace_iterator *iter);
365 void (*close)(struct trace_iterator *iter);
366 void (*pipe_close)(struct trace_iterator *iter);
367 ssize_t (*read)(struct trace_iterator *iter,
370 ssize_t (*splice_read)(struct trace_iterator *iter,
381 enum print_line_t (*print_line)(struct trace_iterator *iter);
525 trace_buffer_iter(struct trace_iterator *iter, int cpu) in trace_buffer_iter() argument
527 if (iter->buffer_iter && iter->buffer_iter[cpu]) in trace_buffer_iter()
528 return iter->buffer_iter[cpu]; in trace_buffer_iter()
[all …]
Dblktrace.c1145 typedef void (blk_log_action_t) (struct trace_iterator *iter, const char *act);
1147 static void blk_log_action_classic(struct trace_iterator *iter, const char *act) in blk_log_action_classic() argument
1150 unsigned long long ts = iter->ts; in blk_log_action_classic()
1153 const struct blk_io_trace *t = te_blk_io_trace(iter->ent); in blk_log_action_classic()
1157 trace_seq_printf(&iter->seq, in blk_log_action_classic()
1159 MAJOR(t->device), MINOR(t->device), iter->cpu, in blk_log_action_classic()
1160 secs, nsec_rem, iter->ent->pid, act, rwbs); in blk_log_action_classic()
1163 static void blk_log_action(struct trace_iterator *iter, const char *act) in blk_log_action() argument
1166 const struct blk_io_trace *t = te_blk_io_trace(iter->ent); in blk_log_action()
1169 trace_seq_printf(&iter->seq, "%3d,%-3d %2s %3s ", in blk_log_action()
[all …]
Dtrace_irqsoff.c210 static void irqsoff_trace_open(struct trace_iterator *iter) in irqsoff_trace_open() argument
213 graph_trace_open(iter); in irqsoff_trace_open()
217 static void irqsoff_trace_close(struct trace_iterator *iter) in irqsoff_trace_close() argument
219 if (iter->private) in irqsoff_trace_close()
220 graph_trace_close(iter); in irqsoff_trace_close()
228 static enum print_line_t irqsoff_print_line(struct trace_iterator *iter) in irqsoff_print_line() argument
235 return print_graph_function_flags(iter, GRAPH_TRACER_FLAGS); in irqsoff_print_line()
273 static enum print_line_t irqsoff_print_line(struct trace_iterator *iter) in irqsoff_print_line() argument
279 static void irqsoff_trace_open(struct trace_iterator *iter) { } in irqsoff_trace_open() argument
280 static void irqsoff_trace_close(struct trace_iterator *iter) { } in irqsoff_trace_close() argument
Dtrace_sched_wakeup.c260 static void wakeup_trace_open(struct trace_iterator *iter) in wakeup_trace_open() argument
263 graph_trace_open(iter); in wakeup_trace_open()
266 static void wakeup_trace_close(struct trace_iterator *iter) in wakeup_trace_close() argument
268 if (iter->private) in wakeup_trace_close()
269 graph_trace_close(iter); in wakeup_trace_close()
276 static enum print_line_t wakeup_print_line(struct trace_iterator *iter) in wakeup_print_line() argument
283 return print_graph_function_flags(iter, GRAPH_TRACER_FLAGS); in wakeup_print_line()
320 static enum print_line_t wakeup_print_line(struct trace_iterator *iter) in wakeup_print_line() argument
326 static void wakeup_trace_open(struct trace_iterator *iter) { } in wakeup_trace_open() argument
327 static void wakeup_trace_close(struct trace_iterator *iter) { } in wakeup_trace_close() argument
Dtrace_branch.c149 static enum print_line_t trace_branch_print(struct trace_iterator *iter, in trace_branch_print() argument
154 trace_assign_type(field, iter->ent); in trace_branch_print()
156 trace_seq_printf(&iter->seq, "[%s] %s:%s:%d\n", in trace_branch_print()
162 return trace_handle_return(&iter->seq); in trace_branch_print()
Dtrace_syscalls.c110 print_syscall_enter(struct trace_iterator *iter, int flags, in print_syscall_enter() argument
113 struct trace_seq *s = &iter->seq; in print_syscall_enter()
114 struct trace_entry *ent = iter->ent; in print_syscall_enter()
156 print_syscall_exit(struct trace_iterator *iter, int flags, in print_syscall_exit() argument
159 struct trace_seq *s = &iter->seq; in print_syscall_exit()
160 struct trace_entry *ent = iter->ent; in print_syscall_exit()
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/
Dicm.h88 struct mlx4_icm_iter *iter) in mlx4_icm_first() argument
90 iter->icm = icm; in mlx4_icm_first()
91 iter->chunk = list_empty(&icm->chunk_list) ? in mlx4_icm_first()
94 iter->page_idx = 0; in mlx4_icm_first()
97 static inline int mlx4_icm_last(struct mlx4_icm_iter *iter) in mlx4_icm_last() argument
99 return !iter->chunk; in mlx4_icm_last()
102 static inline void mlx4_icm_next(struct mlx4_icm_iter *iter) in mlx4_icm_next() argument
104 if (++iter->page_idx >= iter->chunk->nsg) { in mlx4_icm_next()
105 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mlx4_icm_next()
106 iter->chunk = NULL; in mlx4_icm_next()
[all …]
/linux-4.1.27/kernel/
Dkallsyms.c458 static int get_ksymbol_mod(struct kallsym_iter *iter) in get_ksymbol_mod() argument
460 if (module_get_kallsym(iter->pos - kallsyms_num_syms, &iter->value, in get_ksymbol_mod()
461 &iter->type, iter->name, iter->module_name, in get_ksymbol_mod()
462 &iter->exported) < 0) in get_ksymbol_mod()
468 static unsigned long get_ksymbol_core(struct kallsym_iter *iter) in get_ksymbol_core() argument
470 unsigned off = iter->nameoff; in get_ksymbol_core()
472 iter->module_name[0] = '\0'; in get_ksymbol_core()
473 iter->value = kallsyms_addresses[iter->pos]; in get_ksymbol_core()
475 iter->type = kallsyms_get_symbol_type(off); in get_ksymbol_core()
477 off = kallsyms_expand_symbol(off, iter->name, ARRAY_SIZE(iter->name)); in get_ksymbol_core()
[all …]
Djump_label.c140 struct jump_entry *iter; in __jump_label_text_reserved() local
142 iter = iter_start; in __jump_label_text_reserved()
143 while (iter < iter_stop) { in __jump_label_text_reserved()
144 if (addr_conflict(iter, start, end)) in __jump_label_text_reserved()
146 iter++; in __jump_label_text_reserved()
197 struct jump_entry *iter; in jump_label_init() local
202 for (iter = iter_start; iter < iter_stop; iter++) { in jump_label_init()
205 iterk = (struct static_key *)(unsigned long)iter->key; in jump_label_init()
206 arch_jump_label_transform_static(iter, jump_label_type(iterk)); in jump_label_init()
214 *((unsigned long *)&key->entries) += (unsigned long)iter; in jump_label_init()
[all …]
Dtracepoint.c353 struct tracepoint * const *iter; in tp_module_going_check_quiescent() local
357 for (iter = begin; iter < end; iter++) in tp_module_going_check_quiescent()
358 WARN_ON_ONCE((*iter)->funcs); in tp_module_going_check_quiescent()
467 struct tracepoint * const *iter; in for_each_tracepoint_range() local
471 for (iter = begin; iter < end; iter++) in for_each_tracepoint_range()
472 fct(*iter, priv); in for_each_tracepoint_range()
/linux-4.1.27/net/rds/
Dinfo.c104 void rds_info_iter_unmap(struct rds_info_iterator *iter) in rds_info_iter_unmap() argument
106 if (iter->addr) { in rds_info_iter_unmap()
107 kunmap_atomic(iter->addr); in rds_info_iter_unmap()
108 iter->addr = NULL; in rds_info_iter_unmap()
115 void rds_info_copy(struct rds_info_iterator *iter, void *data, in rds_info_copy() argument
121 if (!iter->addr) in rds_info_copy()
122 iter->addr = kmap_atomic(*iter->pages); in rds_info_copy()
124 this = min(bytes, PAGE_SIZE - iter->offset); in rds_info_copy()
127 "bytes %lu\n", *iter->pages, iter->addr, in rds_info_copy()
128 iter->offset, this, data, bytes); in rds_info_copy()
[all …]
Dstats.c81 void rds_stats_info_copy(struct rds_info_iterator *iter, in rds_stats_info_copy() argument
93 rds_info_copy(iter, &ctr, sizeof(ctr)); in rds_stats_info_copy()
109 struct rds_info_iterator *iter, in rds_stats_info() argument
133 rds_stats_info_copy(iter, (uint64_t *)&stats, rds_stat_names, in rds_stats_info()
139 lens->nr = rds_trans_stats_info_copy(iter, avail) + in rds_stats_info()
Dinfo.h18 struct rds_info_iterator *iter,
25 void rds_info_copy(struct rds_info_iterator *iter, void *data,
27 void rds_info_iter_unmap(struct rds_info_iterator *iter);
Dconnection.c393 struct rds_info_iterator *iter, in rds_conn_message_info() argument
423 rds_inc_info_copy(&rm->m_inc, iter, in rds_conn_message_info()
438 struct rds_info_iterator *iter, in rds_conn_message_info_send() argument
441 rds_conn_message_info(sock, len, iter, lens, 1); in rds_conn_message_info_send()
446 struct rds_info_iterator *iter, in rds_conn_message_info_retrans() argument
449 rds_conn_message_info(sock, len, iter, lens, 0); in rds_conn_message_info_retrans()
453 struct rds_info_iterator *iter, in rds_for_each_conn_info() argument
480 rds_info_copy(iter, buffer, item_len); in rds_for_each_conn_info()
516 struct rds_info_iterator *iter, in rds_conn_info() argument
519 rds_for_each_conn_info(sock, len, iter, lens, in rds_conn_info()
Dtransport.c110 unsigned int rds_trans_stats_info_copy(struct rds_info_iterator *iter, in rds_trans_stats_info_copy() argument
119 rds_info_iter_unmap(iter); in rds_trans_stats_info_copy()
128 part = trans->stats_info_copy(iter, avail); in rds_trans_stats_info_copy()
Dtcp_stats.c51 unsigned int rds_tcp_stats_info_copy(struct rds_info_iterator *iter, in rds_tcp_stats_info_copy() argument
70 rds_stats_info_copy(iter, (uint64_t *)&stats, rds_tcp_stat_names, in rds_tcp_stats_info_copy()
Diw_stats.c72 unsigned int rds_iw_stats_info_copy(struct rds_info_iterator *iter, in rds_iw_stats_info_copy() argument
91 rds_stats_info_copy(iter, (uint64_t *)&stats, rds_iw_stat_names, in rds_iw_stats_info_copy()
Dib_stats.c74 unsigned int rds_ib_stats_info_copy(struct rds_info_iterator *iter, in rds_ib_stats_info_copy() argument
93 rds_stats_info_copy(iter, (uint64_t *)&stats, rds_ib_stat_names, in rds_ib_stats_info_copy()
Daf_rds.c467 struct rds_info_iterator *iter, in rds_sock_inc_info() argument
485 rds_inc_info_copy(inc, iter, inc->i_saddr, in rds_sock_inc_info()
499 struct rds_info_iterator *iter, in rds_sock_info() argument
521 rds_info_copy(iter, &sinfo, sizeof(sinfo)); in rds_sock_info()
Drds.h444 unsigned int (*stats_info_copy)(struct rds_info_iterator *iter,
624 struct rds_info_iterator *iter,
711 struct rds_info_iterator *iter,
770 void rds_stats_info_copy(struct rds_info_iterator *iter,
804 unsigned int rds_trans_stats_info_copy(struct rds_info_iterator *iter,
/linux-4.1.27/drivers/infiniband/ulp/ipoib/
Dipoib_fs.c60 struct ipoib_mcast_iter *iter; in ipoib_mcg_seq_start() local
63 iter = ipoib_mcast_iter_init(file->private); in ipoib_mcg_seq_start()
64 if (!iter) in ipoib_mcg_seq_start()
68 if (ipoib_mcast_iter_next(iter)) { in ipoib_mcg_seq_start()
69 kfree(iter); in ipoib_mcg_seq_start()
74 return iter; in ipoib_mcg_seq_start()
80 struct ipoib_mcast_iter *iter = iter_ptr; in ipoib_mcg_seq_next() local
84 if (ipoib_mcast_iter_next(iter)) { in ipoib_mcg_seq_next()
85 kfree(iter); in ipoib_mcg_seq_next()
89 return iter; in ipoib_mcg_seq_next()
[all …]
Dipoib_multicast.c935 struct ipoib_mcast_iter *iter; in ipoib_mcast_iter_init() local
937 iter = kmalloc(sizeof *iter, GFP_KERNEL); in ipoib_mcast_iter_init()
938 if (!iter) in ipoib_mcast_iter_init()
941 iter->dev = dev; in ipoib_mcast_iter_init()
942 memset(iter->mgid.raw, 0, 16); in ipoib_mcast_iter_init()
944 if (ipoib_mcast_iter_next(iter)) { in ipoib_mcast_iter_init()
945 kfree(iter); in ipoib_mcast_iter_init()
949 return iter; in ipoib_mcast_iter_init()
952 int ipoib_mcast_iter_next(struct ipoib_mcast_iter *iter) in ipoib_mcast_iter_next() argument
954 struct ipoib_dev_priv *priv = netdev_priv(iter->dev); in ipoib_mcast_iter_next()
[all …]
Dipoib_main.c332 struct ipoib_path_iter *iter; in ipoib_path_iter_init() local
334 iter = kmalloc(sizeof *iter, GFP_KERNEL); in ipoib_path_iter_init()
335 if (!iter) in ipoib_path_iter_init()
338 iter->dev = dev; in ipoib_path_iter_init()
339 memset(iter->path.pathrec.dgid.raw, 0, 16); in ipoib_path_iter_init()
341 if (ipoib_path_iter_next(iter)) { in ipoib_path_iter_init()
342 kfree(iter); in ipoib_path_iter_init()
346 return iter; in ipoib_path_iter_init()
349 int ipoib_path_iter_next(struct ipoib_path_iter *iter) in ipoib_path_iter_next() argument
351 struct ipoib_dev_priv *priv = netdev_priv(iter->dev); in ipoib_path_iter_next()
[all …]
/linux-4.1.27/drivers/infiniband/hw/mthca/
Dmthca_memfree.h100 struct mthca_icm_iter *iter) in mthca_icm_first() argument
102 iter->icm = icm; in mthca_icm_first()
103 iter->chunk = list_empty(&icm->chunk_list) ? in mthca_icm_first()
106 iter->page_idx = 0; in mthca_icm_first()
109 static inline int mthca_icm_last(struct mthca_icm_iter *iter) in mthca_icm_last() argument
111 return !iter->chunk; in mthca_icm_last()
114 static inline void mthca_icm_next(struct mthca_icm_iter *iter) in mthca_icm_next() argument
116 if (++iter->page_idx >= iter->chunk->nsg) { in mthca_icm_next()
117 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mthca_icm_next()
118 iter->chunk = NULL; in mthca_icm_next()
[all …]
/linux-4.1.27/drivers/s390/cio/
Dblacklist.c287 struct ccwdev_iter *iter = s->private; in cio_ignore_proc_seq_start() local
291 memset(iter, 0, sizeof(*iter)); in cio_ignore_proc_seq_start()
292 iter->ssid = *offset / (__MAX_SUBCHANNEL + 1); in cio_ignore_proc_seq_start()
293 iter->devno = *offset % (__MAX_SUBCHANNEL + 1); in cio_ignore_proc_seq_start()
294 return iter; in cio_ignore_proc_seq_start()
305 struct ccwdev_iter *iter; in cio_ignore_proc_seq_next() local
309 iter = it; in cio_ignore_proc_seq_next()
310 if (iter->devno == __MAX_SUBCHANNEL) { in cio_ignore_proc_seq_next()
311 iter->devno = 0; in cio_ignore_proc_seq_next()
312 iter->ssid++; in cio_ignore_proc_seq_next()
[all …]
/linux-4.1.27/tools/perf/util/
Dhist.c478 iter_next_nop_entry(struct hist_entry_iter *iter __maybe_unused, in iter_next_nop_entry()
485 iter_add_next_nop_entry(struct hist_entry_iter *iter __maybe_unused, in iter_add_next_nop_entry()
492 iter_prepare_mem_entry(struct hist_entry_iter *iter, struct addr_location *al) in iter_prepare_mem_entry() argument
494 struct perf_sample *sample = iter->sample; in iter_prepare_mem_entry()
501 iter->priv = mi; in iter_prepare_mem_entry()
506 iter_add_single_mem_entry(struct hist_entry_iter *iter, struct addr_location *al) in iter_add_single_mem_entry() argument
509 struct mem_info *mi = iter->priv; in iter_add_single_mem_entry()
510 struct hists *hists = evsel__hists(iter->evsel); in iter_add_single_mem_entry()
516 cost = iter->sample->weight; in iter_add_single_mem_entry()
527 he = __hists__add_entry(hists, al, iter->parent, NULL, mi, in iter_add_single_mem_entry()
[all …]
Dcomm.c50 struct comm_str *iter, *new; in comm_str__findnew() local
55 iter = rb_entry(parent, struct comm_str, rb_node); in comm_str__findnew()
57 cmp = strcmp(str, iter->str); in comm_str__findnew()
59 return iter; in comm_str__findnew()
Dordered-events.c188 struct ordered_event *tmp, *iter; in __ordered_events__flush() local
201 list_for_each_entry_safe(iter, tmp, head, list) { in __ordered_events__flush()
205 if (iter->timestamp > limit) in __ordered_events__flush()
207 ret = oe->deliver(oe, iter); in __ordered_events__flush()
211 ordered_events__delete(oe, iter); in __ordered_events__flush()
212 oe->last_flush = iter->timestamp; in __ordered_events__flush()
/linux-4.1.27/drivers/dma/ppc4xx/
Dadma.c186 struct ppc440spe_adma_desc_slot *iter) in print_cb_list() argument
188 for (; iter; iter = iter->hw_next) in print_cb_list()
189 print_cb(chan, iter->hw_desc); in print_cb_list()
334 struct ppc440spe_adma_desc_slot *iter; in ppc440spe_desc_init_dma01pq() local
348 list_for_each_entry(iter, &desc->group_list, chain_node) { in ppc440spe_desc_init_dma01pq()
349 hw_desc = iter->hw_desc; in ppc440spe_desc_init_dma01pq()
350 memset(iter->hw_desc, 0, sizeof(struct dma_cdb)); in ppc440spe_desc_init_dma01pq()
352 if (likely(!list_is_last(&iter->chain_node, in ppc440spe_desc_init_dma01pq()
355 iter->hw_next = list_entry(iter->chain_node.next, in ppc440spe_desc_init_dma01pq()
357 clear_bit(PPC440SPE_DESC_INT, &iter->flags); in ppc440spe_desc_init_dma01pq()
[all …]
/linux-4.1.27/lib/
Drhashtable.c498 int rhashtable_walk_init(struct rhashtable *ht, struct rhashtable_iter *iter) in rhashtable_walk_init() argument
500 iter->ht = ht; in rhashtable_walk_init()
501 iter->p = NULL; in rhashtable_walk_init()
502 iter->slot = 0; in rhashtable_walk_init()
503 iter->skip = 0; in rhashtable_walk_init()
505 iter->walker = kmalloc(sizeof(*iter->walker), GFP_KERNEL); in rhashtable_walk_init()
506 if (!iter->walker) in rhashtable_walk_init()
510 iter->walker->tbl = in rhashtable_walk_init()
512 list_add(&iter->walker->list, &iter->walker->tbl->walkers); in rhashtable_walk_init()
525 void rhashtable_walk_exit(struct rhashtable_iter *iter) in rhashtable_walk_exit() argument
[all …]
Dcordic.c59 unsigned iter; in cordic_calc_iq() local
80 for (iter = 0; iter < CORDIC_NUM_ITER; iter++) { in cordic_calc_iq()
82 valtmp = coord.i - (coord.q >> iter); in cordic_calc_iq()
83 coord.q += (coord.i >> iter); in cordic_calc_iq()
84 angle += arctan_table[iter]; in cordic_calc_iq()
86 valtmp = coord.i + (coord.q >> iter); in cordic_calc_iq()
87 coord.q -= (coord.i >> iter); in cordic_calc_iq()
88 angle -= arctan_table[iter]; in cordic_calc_iq()
Dplist.c76 struct plist_node *first, *iter, *prev = NULL; in plist_add() local
86 first = iter = plist_first(head); in plist_add()
89 if (node->prio < iter->prio) { in plist_add()
90 node_next = &iter->node_list; in plist_add()
94 prev = iter; in plist_add()
95 iter = list_entry(iter->prio_list.next, in plist_add()
97 } while (iter != first); in plist_add()
100 list_add_tail(&node->prio_list, &iter->prio_list); in plist_add()
148 struct plist_node *iter; in plist_requeue() local
158 iter = plist_next(node); in plist_requeue()
[all …]
Ddynamic_debug.c684 static struct _ddebug *ddebug_iter_first(struct ddebug_iter *iter) in ddebug_iter_first() argument
687 iter->table = NULL; in ddebug_iter_first()
688 iter->idx = 0; in ddebug_iter_first()
691 iter->table = list_entry(ddebug_tables.next, in ddebug_iter_first()
693 iter->idx = 0; in ddebug_iter_first()
694 return &iter->table->ddebugs[iter->idx]; in ddebug_iter_first()
703 static struct _ddebug *ddebug_iter_next(struct ddebug_iter *iter) in ddebug_iter_next() argument
705 if (iter->table == NULL) in ddebug_iter_next()
707 if (++iter->idx == iter->table->num_ddebugs) { in ddebug_iter_next()
709 iter->idx = 0; in ddebug_iter_next()
[all …]
Dradix-tree.c748 struct radix_tree_iter *iter, unsigned flags) in radix_tree_next_chunk() argument
766 index = iter->next_index; in radix_tree_next_chunk()
767 if (!index && iter->index) in radix_tree_next_chunk()
775 iter->index = 0; in radix_tree_next_chunk()
776 iter->next_index = 1; in radix_tree_next_chunk()
777 iter->tags = 1; in radix_tree_next_chunk()
831 iter->index = index; in radix_tree_next_chunk()
832 iter->next_index = (index | RADIX_TREE_MAP_MASK) + 1; in radix_tree_next_chunk()
840 iter->tags = node->tags[tag][tag_long] >> tag_bit; in radix_tree_next_chunk()
845 iter->tags |= node->tags[tag][tag_long + 1] << in radix_tree_next_chunk()
[all …]
/linux-4.1.27/drivers/net/wireless/libertas/
Dfirmware.c85 const struct lbs_fw_table *iter; in load_next_firmware_from_table() local
88 iter = priv->fw_table; in load_next_firmware_from_table()
90 iter = ++priv->fw_iter; in load_next_firmware_from_table()
98 if (!iter->helper) { in load_next_firmware_from_table()
104 if (iter->model != priv->fw_model) { in load_next_firmware_from_table()
105 iter++; in load_next_firmware_from_table()
109 priv->fw_iter = iter; in load_next_firmware_from_table()
110 do_load_firmware(priv, iter->helper, helper_firmware_cb); in load_next_firmware_from_table()
176 const struct lbs_fw_table *iter; in lbs_get_firmware() local
183 iter = fw_table; in lbs_get_firmware()
[all …]
/linux-4.1.27/kernel/time/
Dtimer_list.c293 struct timer_list_iter *iter = v; in timer_list_show() local
295 if (iter->cpu == -1 && !iter->second_pass) in timer_list_show()
296 timer_list_header(m, iter->now); in timer_list_show()
297 else if (!iter->second_pass) in timer_list_show()
298 print_cpu(m, iter->cpu, iter->now); in timer_list_show()
300 else if (iter->cpu == -1 && iter->second_pass) in timer_list_show()
303 print_tickdevice(m, tick_get_device(iter->cpu), iter->cpu); in timer_list_show()
326 static void *move_iter(struct timer_list_iter *iter, loff_t offset) in move_iter() argument
329 iter->cpu = cpumask_next(iter->cpu, cpu_online_mask); in move_iter()
330 if (iter->cpu >= nr_cpu_ids) { in move_iter()
[all …]
/linux-4.1.27/drivers/staging/lustre/lustre/lov/
Dlov_pool.c175 struct pool_iterator *iter = (struct pool_iterator *)s->private; in pool_proc_next() local
178 LASSERTF(iter->magic == POOL_IT_MAGIC, "%08X", iter->magic); in pool_proc_next()
181 if (*pos >= pool_tgt_count(iter->pool)) in pool_proc_next()
185 prev_idx = iter->idx; in pool_proc_next()
186 down_read(&pool_tgt_rw_sem(iter->pool)); in pool_proc_next()
187 iter->idx++; in pool_proc_next()
188 if (iter->idx == pool_tgt_count(iter->pool)) { in pool_proc_next()
189 iter->idx = prev_idx; /* we stay on the last entry */ in pool_proc_next()
190 up_read(&pool_tgt_rw_sem(iter->pool)); in pool_proc_next()
193 up_read(&pool_tgt_rw_sem(iter->pool)); in pool_proc_next()
[all …]
/linux-4.1.27/mm/
Dprocess_vm_access.c36 struct iov_iter *iter, in process_vm_rw_pages() argument
40 while (len && iov_iter_count(iter)) { in process_vm_rw_pages()
49 copied = copy_page_from_iter(page, offset, copy, iter); in process_vm_rw_pages()
52 copied = copy_page_to_iter(page, offset, copy, iter); in process_vm_rw_pages()
55 if (copied < copy && iov_iter_count(iter)) in process_vm_rw_pages()
79 struct iov_iter *iter, in process_vm_rw_single_vec() argument
97 while (!rc && nr_pages && iov_iter_count(iter)) { in process_vm_rw_single_vec()
112 start_offset, bytes, iter, in process_vm_rw_single_vec()
141 static ssize_t process_vm_rw_core(pid_t pid, struct iov_iter *iter, in process_vm_rw_core() argument
155 size_t total_len = iov_iter_count(iter); in process_vm_rw_core()
[all …]
Dslab.h170 #define for_each_memcg_cache(iter, root) \ argument
171 list_for_each_entry(iter, &(root)->memcg_params.list, \
174 #define for_each_memcg_cache_safe(iter, tmp, root) \ argument
175 list_for_each_entry_safe(iter, tmp, &(root)->memcg_params.list, \
256 #define for_each_memcg_cache(iter, root) \ argument
257 for ((void)(iter), (void)(root); 0; )
258 #define for_each_memcg_cache_safe(iter, tmp, root) \ argument
259 for ((void)(iter), (void)(tmp), (void)(root); 0; )
Dfilemap.c1161 struct radix_tree_iter iter; in find_get_entries() local
1168 radix_tree_for_each_slot(slot, &mapping->page_tree, &iter, start) { in find_get_entries()
1193 indices[ret] = iter.index; in find_get_entries()
1221 struct radix_tree_iter iter; in find_get_pages() local
1230 radix_tree_for_each_slot(slot, &mapping->page_tree, &iter, start) { in find_get_pages()
1244 WARN_ON(iter.index); in find_get_pages()
1288 struct radix_tree_iter iter; in find_get_pages_contig() local
1297 radix_tree_for_each_contig(slot, &mapping->page_tree, &iter, index) { in find_get_pages_contig()
1336 if (page->mapping == NULL || page->index != iter.index) { in find_get_pages_contig()
1364 struct radix_tree_iter iter; in find_get_pages_tag() local
[all …]
Dmemcontrol.c148 struct reclaim_iter iter[DEF_PRIORITY + 1]; member
1013 struct reclaim_iter *uninitialized_var(iter); in mem_cgroup_iter()
1039 iter = &mz->iter[reclaim->priority]; in mem_cgroup_iter()
1041 if (prev && reclaim->generation != iter->generation) in mem_cgroup_iter()
1045 pos = READ_ONCE(iter->position); in mem_cgroup_iter()
1097 if (cmpxchg(&iter->position, pos, memcg) == pos) { in mem_cgroup_iter()
1112 iter->generation++; in mem_cgroup_iter()
1114 reclaim->generation = iter->generation; in mem_cgroup_iter()
1145 #define for_each_mem_cgroup_tree(iter, root) \ argument
1146 for (iter = mem_cgroup_iter(root, NULL, NULL); \
[all …]
/linux-4.1.27/net/ipv4/
Dcipso_ipv4.c253 u32 iter; in cipso_v4_cache_init() local
261 for (iter = 0; iter < CIPSO_V4_CACHE_BUCKETS; iter++) { in cipso_v4_cache_init()
262 spin_lock_init(&cipso_v4_cache[iter].lock); in cipso_v4_cache_init()
263 cipso_v4_cache[iter].size = 0; in cipso_v4_cache_init()
264 INIT_LIST_HEAD(&cipso_v4_cache[iter].list); in cipso_v4_cache_init()
281 u32 iter; in cipso_v4_cache_invalidate() local
283 for (iter = 0; iter < CIPSO_V4_CACHE_BUCKETS; iter++) { in cipso_v4_cache_invalidate()
284 spin_lock_bh(&cipso_v4_cache[iter].lock); in cipso_v4_cache_invalidate()
287 &cipso_v4_cache[iter].list, list) { in cipso_v4_cache_invalidate()
291 cipso_v4_cache[iter].size = 0; in cipso_v4_cache_invalidate()
[all …]
Dfib_trie.c2010 static struct key_vector *fib_trie_get_next(struct fib_trie_iter *iter) in fib_trie_get_next() argument
2012 unsigned long cindex = iter->index; in fib_trie_get_next()
2013 struct key_vector *pn = iter->tnode; in fib_trie_get_next()
2017 iter->tnode, iter->index, iter->depth); in fib_trie_get_next()
2027 iter->tnode = pn; in fib_trie_get_next()
2028 iter->index = cindex; in fib_trie_get_next()
2031 iter->tnode = n; in fib_trie_get_next()
2032 iter->index = 0; in fib_trie_get_next()
2033 ++iter->depth; in fib_trie_get_next()
2043 --iter->depth; in fib_trie_get_next()
[all …]
/linux-4.1.27/net/ipv6/
Dip6_fib.c694 struct rt6_info *iter = NULL; in fib6_add_rt2node() local
707 for (iter = fn->leaf; iter; iter = iter->dst.rt6_next) { in fib6_add_rt2node()
712 if (iter->rt6i_metric == rt->rt6i_metric) { in fib6_add_rt2node()
720 if (rt_can_ecmp == rt6_qualify_for_ecmp(iter)) { in fib6_add_rt2node()
729 if (iter->dst.dev == rt->dst.dev && in fib6_add_rt2node()
730 iter->rt6i_idev == rt->rt6i_idev && in fib6_add_rt2node()
731 ipv6_addr_equal(&iter->rt6i_gateway, in fib6_add_rt2node()
735 if (!(iter->rt6i_flags & RTF_EXPIRES)) in fib6_add_rt2node()
738 rt6_clean_expires(iter); in fib6_add_rt2node()
740 rt6_set_expires(iter, rt->dst.expires); in fib6_add_rt2node()
[all …]
Dip6mr.c389 struct ipmr_vif_iter *iter, in ip6mr_vif_seq_idx() argument
392 struct mr6_table *mrt = iter->mrt; in ip6mr_vif_seq_idx()
394 for (iter->ct = 0; iter->ct < mrt->maxvif; ++iter->ct) { in ip6mr_vif_seq_idx()
395 if (!MIF_EXISTS(mrt, iter->ct)) in ip6mr_vif_seq_idx()
398 return &mrt->vif6_table[iter->ct]; in ip6mr_vif_seq_idx()
406 struct ipmr_vif_iter *iter = seq->private; in ip6mr_vif_seq_start() local
414 iter->mrt = mrt; in ip6mr_vif_seq_start()
423 struct ipmr_vif_iter *iter = seq->private; in ip6mr_vif_seq_next() local
425 struct mr6_table *mrt = iter->mrt; in ip6mr_vif_seq_next()
429 return ip6mr_vif_seq_idx(net, iter, 0); in ip6mr_vif_seq_next()
[all …]
/linux-4.1.27/drivers/base/
Dclass.c294 void class_dev_iter_init(struct class_dev_iter *iter, struct class *class, in class_dev_iter_init() argument
301 klist_iter_init_node(&class->p->klist_devices, &iter->ki, start_knode); in class_dev_iter_init()
302 iter->type = type; in class_dev_iter_init()
318 struct device *class_dev_iter_next(struct class_dev_iter *iter) in class_dev_iter_next() argument
324 knode = klist_next(&iter->ki); in class_dev_iter_next()
328 if (!iter->type || iter->type == dev->type) in class_dev_iter_next()
341 void class_dev_iter_exit(struct class_dev_iter *iter) in class_dev_iter_exit() argument
343 klist_iter_exit(&iter->ki); in class_dev_iter_exit()
368 struct class_dev_iter iter; in class_for_each_device() local
380 class_dev_iter_init(&iter, class, start, NULL); in class_for_each_device()
[all …]
Dattribute_container.c183 #define klist_for_each_entry(pos, head, member, iter) \ argument
184 for (klist_iter_init(head, iter); (pos = ({ \
185 struct klist_node *n = klist_next(iter); \
187 ({ klist_iter_exit(iter) ; NULL; }); \
217 struct klist_iter iter; in attribute_container_remove_device() local
225 klist_for_each_entry(ic, &cont->containers, node, &iter) { in attribute_container_remove_device()
261 struct klist_iter iter; in attribute_container_device_trigger() local
271 klist_for_each_entry(ic, &cont->containers, node, &iter) { in attribute_container_device_trigger()
429 struct klist_iter iter; in attribute_container_find_class_device() local
431 klist_for_each_entry(ic, &cont->containers, node, &iter) { in attribute_container_find_class_device()
[all …]
Dbus.c1062 void subsys_dev_iter_init(struct subsys_dev_iter *iter, struct bus_type *subsys, in subsys_dev_iter_init() argument
1069 klist_iter_init_node(&subsys->p->klist_devices, &iter->ki, start_knode); in subsys_dev_iter_init()
1070 iter->type = type; in subsys_dev_iter_init()
1086 struct device *subsys_dev_iter_next(struct subsys_dev_iter *iter) in subsys_dev_iter_next() argument
1092 knode = klist_next(&iter->ki); in subsys_dev_iter_next()
1096 if (!iter->type || iter->type == dev->type) in subsys_dev_iter_next()
1109 void subsys_dev_iter_exit(struct subsys_dev_iter *iter) in subsys_dev_iter_exit() argument
1111 klist_iter_exit(&iter->ki); in subsys_dev_iter_exit()
1118 struct subsys_dev_iter iter; in subsys_interface_register() local
1131 subsys_dev_iter_init(&iter, subsys, NULL, NULL); in subsys_interface_register()
[all …]
/linux-4.1.27/drivers/md/bcache/
Dbset.c55 struct btree_iter iter; in __bch_count_data() local
59 for_each_key(b, k, &iter) in __bch_count_data()
68 struct btree_iter iter; in __bch_check_keys() local
71 for_each_key(b, k, &iter) { in __bch_check_keys()
109 static void bch_btree_iter_next_check(struct btree_iter *iter) in bch_btree_iter_next_check() argument
111 struct bkey *k = iter->data->k, *next = bkey_next(k); in bch_btree_iter_next_check()
113 if (next < iter->data->end && in bch_btree_iter_next_check()
114 bkey_cmp(k, iter->b->ops->is_extents ? in bch_btree_iter_next_check()
116 bch_dump_bucket(iter->b); in bch_btree_iter_next_check()
123 static inline void bch_btree_iter_next_check(struct btree_iter *iter) {} in bch_btree_iter_next_check() argument
[all …]
Dextents.c29 static void sort_key_next(struct btree_iter *iter, in sort_key_next() argument
35 *i = iter->data[--iter->used]; in sort_key_next()
227 struct btree_iter *iter, in bch_btree_ptr_insert_fixup() argument
264 static struct bkey *bch_extent_sort_fixup(struct btree_iter *iter, in bch_extent_sort_fixup() argument
267 while (iter->used > 1) { in bch_extent_sort_fixup()
268 struct btree_iter_set *top = iter->data, *i = top + 1; in bch_extent_sort_fixup()
270 if (iter->used > 2 && in bch_extent_sort_fixup()
278 sort_key_next(iter, i); in bch_extent_sort_fixup()
279 heap_sift(iter, i - top, bch_extent_sort_cmp); in bch_extent_sort_fixup()
285 sort_key_next(iter, i); in bch_extent_sort_fixup()
[all …]
Dbtree.h200 #define for_each_cached_btree(b, c, iter) \ argument
201 for (iter = 0; \
202 iter < ARRAY_SIZE((c)->bucket_hash); \
203 iter++) \
204 hlist_for_each_entry_rcu((b), (c)->bucket_hash + iter, hash)
Dbset.h338 #define for_each_key_filter(b, k, iter, filter) \ argument
339 for (bch_btree_iter_init((b), (iter), NULL); \
340 ((k) = bch_btree_iter_next_filter((iter), (b), filter));)
342 #define for_each_key(b, k, iter) \ argument
343 for (bch_btree_iter_init((b), (iter), NULL); \
344 ((k) = bch_btree_iter_next(iter));)
Dalloc.c414 size_t iter; in bch_bucket_alloc() local
418 for (iter = 0; iter < prio_buckets(ca) * 2; iter++) in bch_bucket_alloc()
419 BUG_ON(ca->prio_buckets[iter] == (uint64_t) r); in bch_bucket_alloc()
422 fifo_for_each(i, &ca->free[j], iter) in bch_bucket_alloc()
424 fifo_for_each(i, &ca->free_inc, iter) in bch_bucket_alloc()
Dbtree.c202 struct btree_iter *iter; in bch_btree_node_read_done() local
204 iter = mempool_alloc(b->c->fill_iter, GFP_NOIO); in bch_btree_node_read_done()
205 iter->size = b->c->sb.bucket_size / b->c->sb.block_size; in bch_btree_node_read_done()
206 iter->used = 0; in bch_btree_node_read_done()
209 iter->b = &b->keys; in bch_btree_node_read_done()
247 bch_btree_iter_push(iter, i->start, bset_bkey_last(i)); in bch_btree_node_read_done()
259 bch_btree_sort_and_fix_extents(&b->keys, iter, &b->c->sort); in bch_btree_node_read_done()
271 mempool_free(iter, b->c->fill_iter); in bch_btree_node_read_done()
1250 struct btree_iter iter; in btree_gc_mark_node() local
1255 for_each_key_filter(&b->keys, k, &iter, bch_ptr_invalid) { in btree_gc_mark_node()
[all …]
Dbcache.h816 #define for_each_cache(ca, cs, iter) \ argument
817 for (iter = 0; ca = cs->cache[iter], iter < (cs)->sb.nr_in_set; iter++)
Dutil.h126 #define fifo_for_each(c, fifo, iter) \ argument
127 for (iter = (fifo)->front; \
128 c = (fifo)->data[iter], iter != (fifo)->back; \
129 iter = (iter + 1) & (fifo)->mask)
Ddebug.c110 struct bvec_iter iter; in bch_data_verify() local
122 bio_for_each_segment(bv, bio, iter) { in bch_data_verify()
124 void *p2 = page_address(check->bi_io_vec[iter.bi_idx].bv_page); in bch_data_verify()
/linux-4.1.27/drivers/infiniband/hw/qib/
Dqib_debugfs.c193 struct qib_qp_iter *iter; in DEBUGFS_FILE() local
197 iter = qib_qp_iter_init(s->private); in DEBUGFS_FILE()
198 if (!iter) in DEBUGFS_FILE()
202 if (qib_qp_iter_next(iter)) { in DEBUGFS_FILE()
203 kfree(iter); in DEBUGFS_FILE()
208 return iter; in DEBUGFS_FILE()
214 struct qib_qp_iter *iter = iter_ptr; in _qp_stats_seq_next() local
218 if (qib_qp_iter_next(iter)) { in _qp_stats_seq_next()
219 kfree(iter); in _qp_stats_seq_next()
223 return iter; in _qp_stats_seq_next()
[all …]
Dqib_qp.c1323 struct qib_qp_iter *iter; in qib_qp_iter_init() local
1325 iter = kzalloc(sizeof(*iter), GFP_KERNEL); in qib_qp_iter_init()
1326 if (!iter) in qib_qp_iter_init()
1329 iter->dev = dev; in qib_qp_iter_init()
1330 if (qib_qp_iter_next(iter)) { in qib_qp_iter_init()
1331 kfree(iter); in qib_qp_iter_init()
1335 return iter; in qib_qp_iter_init()
1338 int qib_qp_iter_next(struct qib_qp_iter *iter) in qib_qp_iter_next() argument
1340 struct qib_ibdev *dev = iter->dev; in qib_qp_iter_next()
1341 int n = iter->n; in qib_qp_iter_next()
[all …]
/linux-4.1.27/drivers/gpu/drm/radeon/
Ddrm_buffer.h130 int iter = buffer->iterator + offset * 4; in drm_buffer_pointer_to_dword() local
131 return &buffer->data[iter / PAGE_SIZE][iter & (PAGE_SIZE - 1)]; in drm_buffer_pointer_to_dword()
144 int iter = buffer->iterator + offset; in drm_buffer_pointer_to_byte() local
145 return &buffer->data[iter / PAGE_SIZE][iter & (PAGE_SIZE - 1)]; in drm_buffer_pointer_to_byte()
/linux-4.1.27/arch/sparc/prom/
Dbootstr_32.c18 int iter; in prom_getbootargs() local
30 for (iter = 1; iter < 8; iter++) { in prom_getbootargs()
31 arg = (*(romvec->pv_v0bootargs))->argv[iter]; in prom_getbootargs()
/linux-4.1.27/net/sunrpc/
Ddebugfs.c47 struct rpc_clnt_iter *iter = f->private; in tasks_start() local
49 struct rpc_clnt *clnt = iter->clnt; in tasks_start()
52 iter->pos = pos + 1; in tasks_start()
63 struct rpc_clnt_iter *iter = f->private; in tasks_next() local
64 struct rpc_clnt *clnt = iter->clnt; in tasks_next()
68 ++iter->pos; in tasks_next()
81 struct rpc_clnt_iter *iter = f->private; in tasks_stop() local
82 struct rpc_clnt *clnt = iter->clnt; in tasks_stop()
101 struct rpc_clnt_iter *iter = seq->private; in tasks_open() local
103 iter->clnt = inode->i_private; in tasks_open()
[all …]
/linux-4.1.27/fs/nfs/objlayout/
Dpnfs_osd_xdr_cli.c160 struct pnfs_osd_xdr_decode_layout_iter *iter, struct xdr_stream *xdr) in pnfs_osd_xdr_decode_layout_map() argument
164 memset(iter, 0, sizeof(*iter)); in pnfs_osd_xdr_decode_layout_map()
176 iter->total_comps = layout->olo_num_comps; in pnfs_osd_xdr_decode_layout_map()
181 struct pnfs_osd_xdr_decode_layout_iter *iter, struct xdr_stream *xdr, in pnfs_osd_xdr_decode_layout_comp() argument
184 BUG_ON(iter->decoded_comps > iter->total_comps); in pnfs_osd_xdr_decode_layout_comp()
185 if (iter->decoded_comps == iter->total_comps) in pnfs_osd_xdr_decode_layout_comp()
192 iter->decoded_comps, iter->total_comps); in pnfs_osd_xdr_decode_layout_comp()
204 iter->decoded_comps++; in pnfs_osd_xdr_decode_layout_comp()
/linux-4.1.27/ipc/
Dutil.c778 struct ipc_proc_iter *iter = s->private; in sysvipc_proc_next() local
779 struct ipc_proc_iface *iface = iter->iface; in sysvipc_proc_next()
786 return sysvipc_find_ipc(&iter->ns->ids[iface->ids], *pos, pos); in sysvipc_proc_next()
795 struct ipc_proc_iter *iter = s->private; in sysvipc_proc_start() local
796 struct ipc_proc_iface *iface = iter->iface; in sysvipc_proc_start()
799 ids = &iter->ns->ids[iface->ids]; in sysvipc_proc_start()
822 struct ipc_proc_iter *iter = s->private; in sysvipc_proc_stop() local
823 struct ipc_proc_iface *iface = iter->iface; in sysvipc_proc_stop()
830 ids = &iter->ns->ids[iface->ids]; in sysvipc_proc_stop()
837 struct ipc_proc_iter *iter = s->private; in sysvipc_proc_show() local
[all …]
/linux-4.1.27/drivers/dma/
Dmv_xor.c235 struct mv_xor_desc_slot *iter, *_iter; in mv_xor_clean_completed_slots() local
238 list_for_each_entry_safe(iter, _iter, &mv_chan->completed_slots, in mv_xor_clean_completed_slots()
241 if (async_tx_test_ack(&iter->async_tx)) { in mv_xor_clean_completed_slots()
242 list_del(&iter->completed_node); in mv_xor_clean_completed_slots()
243 mv_xor_free_slots(mv_chan, iter); in mv_xor_clean_completed_slots()
272 struct mv_xor_desc_slot *iter, *_iter; in mv_xor_slot_cleanup() local
287 list_for_each_entry_safe(iter, _iter, &mv_chan->chain, in mv_xor_slot_cleanup()
291 hw_desc = iter->hw_desc; in mv_xor_slot_cleanup()
293 cookie = mv_xor_run_tx_complete_actions(iter, mv_chan, in mv_xor_slot_cleanup()
297 mv_xor_clean_slot(iter, mv_chan); in mv_xor_slot_cleanup()
[all …]
Diop-adma.c116 struct iop_adma_desc_slot *iter, *_iter, *grp_start = NULL; in __iop_adma_slot_cleanup() local
126 list_for_each_entry_safe(iter, _iter, &iop_chan->chain, in __iop_adma_slot_cleanup()
130 iter->async_tx.cookie, iter->idx, busy, in __iop_adma_slot_cleanup()
131 iter->async_tx.phys, iop_desc_get_next_desc(iter), in __iop_adma_slot_cleanup()
132 async_tx_test_ack(&iter->async_tx)); in __iop_adma_slot_cleanup()
147 if (iter->async_tx.phys == current_desc) { in __iop_adma_slot_cleanup()
149 if (busy || iop_desc_get_next_desc(iter)) in __iop_adma_slot_cleanup()
155 slot_cnt = iter->slot_cnt; in __iop_adma_slot_cleanup()
156 slots_per_op = iter->slots_per_op; in __iop_adma_slot_cleanup()
166 grp_start = iter; in __iop_adma_slot_cleanup()
[all …]
Dfsl-edma.c541 u16 soff, doff, iter; in fsl_edma_prep_dma_cyclic() local
554 iter = period_len / nbytes; in fsl_edma_prep_dma_cyclic()
576 fsl_chan->fsc.attr, soff, nbytes, 0, iter, in fsl_edma_prep_dma_cyclic()
577 iter, doff, last_sg, true, false, true); in fsl_edma_prep_dma_cyclic()
593 u16 soff, doff, iter; in fsl_edma_prep_slave_sg() local
621 iter = sg_dma_len(sg) / nbytes; in fsl_edma_prep_slave_sg()
626 nbytes, 0, iter, iter, doff, last_sg, in fsl_edma_prep_slave_sg()
632 nbytes, 0, iter, iter, doff, last_sg, in fsl_edma_prep_slave_sg()
/linux-4.1.27/arch/unicore32/include/asm/
Dmemblock.h34 #define for_each_bank(iter, mi) \ argument
35 for (iter = 0; iter < (mi)->nr_banks; iter++)
/linux-4.1.27/net/sctp/
Dtsnmap.c158 struct sctp_tsnmap_iter *iter) in sctp_tsnmap_iter_init() argument
161 iter->start = map->cumulative_tsn_ack_point + 1; in sctp_tsnmap_iter_init()
168 struct sctp_tsnmap_iter *iter, in sctp_tsnmap_next_gap_ack() argument
175 if (TSN_lte(map->max_tsn_seen, iter->start)) in sctp_tsnmap_next_gap_ack()
178 offset = iter->start - map->base_tsn; in sctp_tsnmap_next_gap_ack()
197 iter->start = map->cumulative_tsn_ack_point + *end + 1; in sctp_tsnmap_next_gap_ack()
335 struct sctp_tsnmap_iter iter; in sctp_tsnmap_num_gabs() local
341 sctp_tsnmap_iter_init(map, &iter); in sctp_tsnmap_num_gabs()
342 while (sctp_tsnmap_next_gap_ack(map, &iter, in sctp_tsnmap_num_gabs()
/linux-4.1.27/security/selinux/
Dnetport.c239 int iter; in sel_netport_init() local
244 for (iter = 0; iter < SEL_NETPORT_HASH_SIZE; iter++) { in sel_netport_init()
245 INIT_LIST_HEAD(&sel_netport_hash[iter].list); in sel_netport_init()
246 sel_netport_hash[iter].size = 0; in sel_netport_init()
Dnetnode.c305 int iter; in sel_netnode_init() local
310 for (iter = 0; iter < SEL_NETNODE_HASH_SIZE; iter++) { in sel_netnode_init()
311 INIT_LIST_HEAD(&sel_netnode_hash[iter].list); in sel_netnode_init()
312 sel_netnode_hash[iter].size = 0; in sel_netnode_init()
Dxfrm.c456 struct dst_entry *iter; in selinux_xfrm_postroute_last() local
458 for (iter = dst; iter != NULL; iter = iter->child) { in selinux_xfrm_postroute_last()
459 struct xfrm_state *x = iter->xfrm; in selinux_xfrm_postroute_last()
/linux-4.1.27/net/netfilter/
Dnft_hash.c186 struct nft_set_iter *iter) in nft_hash_walk() argument
196 iter->err = err; in nft_hash_walk()
202 iter->err = err; in nft_hash_walk()
210 iter->err = err; in nft_hash_walk()
217 if (iter->count < iter->skip) in nft_hash_walk()
226 iter->err = iter->fn(ctx, set, iter, &elem); in nft_hash_walk()
227 if (iter->err < 0) in nft_hash_walk()
231 iter->count++; in nft_hash_walk()
Dnft_rbtree.c175 struct nft_set_iter *iter) in nft_rbtree_walk() argument
187 if (iter->count < iter->skip) in nft_rbtree_walk()
194 iter->err = iter->fn(ctx, set, iter, &elem); in nft_rbtree_walk()
195 if (iter->err < 0) { in nft_rbtree_walk()
200 iter->count++; in nft_rbtree_walk()
Dnf_tables_api.c2842 const struct nft_set_iter *iter, in nf_tables_bind_check_setelem() argument
2859 struct nft_set_iter iter; in nf_tables_bind_set() local
2874 iter.skip = 0; in nf_tables_bind_set()
2875 iter.count = 0; in nf_tables_bind_set()
2876 iter.err = 0; in nf_tables_bind_set()
2877 iter.fn = nf_tables_bind_check_setelem; in nf_tables_bind_set()
2879 set->ops->walk(ctx, set, &iter); in nf_tables_bind_set()
2880 if (iter.err < 0) { in nf_tables_bind_set()
2885 return iter.err; in nf_tables_bind_set()
3047 struct nft_set_iter iter; member
[all …]
/linux-4.1.27/drivers/scsi/qla4xxx/
Dql4_attr.c129 struct sysfs_entry *iter; in qla4_8xxx_alloc_sysfs_attr() local
132 for (iter = bin_file_entries; iter->name; iter++) { in qla4_8xxx_alloc_sysfs_attr()
134 iter->attr); in qla4_8xxx_alloc_sysfs_attr()
138 iter->name, ret); in qla4_8xxx_alloc_sysfs_attr()
145 struct sysfs_entry *iter; in qla4_8xxx_free_sysfs_attr() local
147 for (iter = bin_file_entries; iter->name; iter++) in qla4_8xxx_free_sysfs_attr()
149 iter->attr); in qla4_8xxx_free_sysfs_attr()
/linux-4.1.27/fs/nfs/
Ddirect.c256 ssize_t nfs_direct_IO(struct kiocb *iocb, struct iov_iter *iter, loff_t pos) in nfs_direct_IO() argument
264 VM_BUG_ON(iov_iter_count(iter) != PAGE_SIZE); in nfs_direct_IO()
266 if (iov_iter_rw(iter) == READ) in nfs_direct_IO()
267 return nfs_file_direct_read(iocb, iter, pos); in nfs_direct_IO()
268 return nfs_file_direct_write(iocb, iter); in nfs_direct_IO()
472 struct iov_iter *iter, in nfs_direct_read_schedule_iovec() argument
487 while (iov_iter_count(iter)) { in nfs_direct_read_schedule_iovec()
493 result = iov_iter_get_pages_alloc(iter, &pagevec, in nfs_direct_read_schedule_iovec()
499 iov_iter_advance(iter, bytes); in nfs_direct_read_schedule_iovec()
567 ssize_t nfs_file_direct_read(struct kiocb *iocb, struct iov_iter *iter, in nfs_file_direct_read() argument
[all …]
/linux-4.1.27/tools/perf/ui/gtk/
Dhists.c101 GtkTreeIter iter, new_parent; in perf_gtk__add_callchain() local
117 gtk_tree_store_append(store, &iter, &new_parent); in perf_gtk__add_callchain()
120 gtk_tree_store_set(store, &iter, 0, buf, -1); in perf_gtk__add_callchain()
123 gtk_tree_store_set(store, &iter, col, buf, -1); in perf_gtk__add_callchain()
130 new_parent = iter; in perf_gtk__add_callchain()
141 perf_gtk__add_callchain(&node->rb_root, store, &iter, col, in perf_gtk__add_callchain()
225 GtkTreeIter iter; in perf_gtk__show_hists() local
236 gtk_tree_store_append(store, &iter, NULL); in perf_gtk__show_hists()
249 gtk_tree_store_set(store, &iter, col_idx++, s, -1); in perf_gtk__show_hists()
257 perf_gtk__add_callchain(&h->sorted_chain, store, &iter, in perf_gtk__show_hists()
Dannotate.c121 GtkTreeIter iter; in perf_gtk__annotate_symbol() local
124 gtk_list_store_append(store, &iter); in perf_gtk__annotate_symbol()
140 gtk_list_store_set(store, &iter, ANN_COL__PERCENT, s, -1); in perf_gtk__annotate_symbol()
142 gtk_list_store_set(store, &iter, ANN_COL__OFFSET, s, -1); in perf_gtk__annotate_symbol()
144 gtk_list_store_set(store, &iter, ANN_COL__LINE, s, -1); in perf_gtk__annotate_symbol()
/linux-4.1.27/include/net/
Dbonding.h82 #define bond_for_each_slave(bond, pos, iter) \ argument
83 netdev_for_each_lower_private((bond)->dev, pos, iter)
86 #define bond_for_each_slave_rcu(bond, pos, iter) \ argument
87 netdev_for_each_lower_private_rcu((bond)->dev, pos, iter)
354 struct list_head *iter; in bond_slave_state_change() local
357 bond_for_each_slave(bond, tmp, iter) { in bond_slave_state_change()
367 struct list_head *iter; in bond_slave_state_notify() local
370 bond_for_each_slave(bond, tmp, iter) { in bond_slave_state_notify()
585 struct list_head *iter; in bond_slave_has_mac() local
588 bond_for_each_slave(bond, tmp, iter) in bond_slave_has_mac()
[all …]
Dnetlabel.h284 struct netlbl_lsm_catmap *iter; in netlbl_catmap_free() local
287 iter = catmap; in netlbl_catmap_free()
289 kfree(iter); in netlbl_catmap_free()
/linux-4.1.27/drivers/s390/block/
Dscm_blk_cluster.c79 struct scm_request *iter; in scm_reserve_cluster() local
86 list_for_each_entry(iter, &bdev->cluster_list, cluster.list) { in scm_reserve_cluster()
87 if (iter == scmrq) { in scm_reserve_cluster()
95 for (pos = 0; pos < iter->aob->request.msb_count; pos++) { in scm_reserve_cluster()
96 if (clusters_intersect(req, iter->request[pos]) && in scm_reserve_cluster()
98 rq_data_dir(iter->request[pos]) == WRITE)) { in scm_reserve_cluster()
136 struct req_iterator iter; in scm_prepare_cluster_request() local
183 rq_for_each_segment(bv, req, iter) { in scm_prepare_cluster_request()
Dscm_blk.c52 struct list_head *iter, *safe; in scm_free_rqs() local
56 list_for_each_safe(iter, safe, &inactive_requests) { in scm_free_rqs()
57 scmrq = list_entry(iter, struct scm_request, list); in scm_free_rqs()
188 struct req_iterator iter; in scm_request_prepare() local
203 rq_for_each_segment(bv, req, iter) { in scm_request_prepare()
Ddasd_fba.c262 struct req_iterator iter; in dasd_fba_build_cp() local
285 rq_for_each_segment(bv, req, iter) { in dasd_fba_build_cp()
326 rq_for_each_segment(bv, req, iter) { in dasd_fba_build_cp()
384 struct req_iterator iter; in dasd_fba_free_cp() local
399 rq_for_each_segment(bv, req, iter) { in dasd_fba_free_cp()
/linux-4.1.27/net/switchdev/
Dswitchdev.c52 struct list_head *iter; in netdev_switch_port_stp_update() local
58 netdev_for_each_lower_dev(dev, lower_dev, iter) { in netdev_switch_port_stp_update()
195 struct list_head *iter; in ndo_dflt_netdev_switch_port_bridge_setlink() local
201 netdev_for_each_lower_dev(dev, lower_dev, iter) { in ndo_dflt_netdev_switch_port_bridge_setlink()
225 struct list_head *iter; in ndo_dflt_netdev_switch_port_bridge_dellink() local
231 netdev_for_each_lower_dev(dev, lower_dev, iter) { in ndo_dflt_netdev_switch_port_bridge_dellink()
246 struct list_head *iter; in netdev_switch_get_lowest_dev() local
256 netdev_for_each_lower_dev(dev, lower_dev, iter) { in netdev_switch_get_lowest_dev()
/linux-4.1.27/drivers/staging/lustre/lustre/llite/
Drw26.c362 static ssize_t ll_direct_IO_26(struct kiocb *iocb, struct iov_iter *iter, in ll_direct_IO_26() argument
370 ssize_t count = iov_iter_count(iter); in ll_direct_IO_26()
390 if (iov_iter_alignment(iter) & ~CFS_PAGE_MASK) in ll_direct_IO_26()
402 if (iov_iter_rw(iter) == READ) in ll_direct_IO_26()
406 while (iov_iter_count(iter)) { in ll_direct_IO_26()
410 count = min_t(size_t, iov_iter_count(iter), size); in ll_direct_IO_26()
411 if (iov_iter_rw(iter) == READ) { in ll_direct_IO_26()
418 result = iov_iter_get_pages_alloc(iter, &pages, count, &offs); in ll_direct_IO_26()
421 result = ll_direct_IO_26_seg(env, io, iov_iter_rw(iter), in ll_direct_IO_26()
425 ll_free_user_pages(pages, n, iov_iter_rw(iter) == READ); in ll_direct_IO_26()
[all …]
/linux-4.1.27/drivers/net/wireless/ath/carl9170/
Ddebug.c301 struct carl9170_sta_tid *iter; in carl9170_debugfs_ampdu_state_read() local
307 list_for_each_entry_rcu(iter, &ar->tx_ampdu_list, list) { in carl9170_debugfs_ampdu_state_read()
309 spin_lock_bh(&iter->lock); in carl9170_debugfs_ampdu_state_read()
312 cnt, iter->tid, iter->bsn, iter->snx, iter->hsn, in carl9170_debugfs_ampdu_state_read()
313 iter->max, iter->state, iter->counter); in carl9170_debugfs_ampdu_state_read()
316 CARL9170_BAW_BITS, iter->bitmap); in carl9170_debugfs_ampdu_state_read()
325 offset = BM_STR_OFF(SEQ_DIFF(iter->snx, iter->bsn)); in carl9170_debugfs_ampdu_state_read()
328 offset = BM_STR_OFF(((int)iter->hsn - (int)iter->bsn) % in carl9170_debugfs_ampdu_state_read()
333 " currently queued:%d\n", skb_queue_len(&iter->queue)); in carl9170_debugfs_ampdu_state_read()
336 skb_queue_walk(&iter->queue, skb) { in carl9170_debugfs_ampdu_state_read()
[all …]
Dfw.c36 const struct carl9170fw_desc_head *iter; in carl9170_fw_find_desc() local
38 carl9170fw_for_each_hdr(iter, ar->fw.desc) { in carl9170_fw_find_desc()
39 if (carl9170fw_desc_cmp(iter, descid, len, in carl9170_fw_find_desc()
41 return (void *)iter; in carl9170_fw_find_desc()
45 if (carl9170fw_desc_cmp(iter, descid, len, in carl9170_fw_find_desc()
47 return (void *)iter; in carl9170_fw_find_desc()
Dtx.c605 struct carl9170_sta_tid *iter; in carl9170_tx_ampdu_timeout() local
612 list_for_each_entry_rcu(iter, &ar->tx_ampdu_list, list) { in carl9170_tx_ampdu_timeout()
613 if (iter->state < CARL9170_TID_STATE_IDLE) in carl9170_tx_ampdu_timeout()
616 spin_lock_bh(&iter->lock); in carl9170_tx_ampdu_timeout()
617 skb = skb_peek(&iter->queue); in carl9170_tx_ampdu_timeout()
627 sta = iter->sta; in carl9170_tx_ampdu_timeout()
631 ieee80211_stop_tx_ba_session(sta, iter->tid); in carl9170_tx_ampdu_timeout()
633 spin_unlock_bh(&iter->lock); in carl9170_tx_ampdu_timeout()
1397 struct sk_buff *iter; in carl9170_tx_ampdu_queue() local
1432 skb_queue_reverse_walk(&agg->queue, iter) { in carl9170_tx_ampdu_queue()
[all …]
/linux-4.1.27/net/nfc/
Dnfc.h113 static inline void nfc_device_iter_init(struct class_dev_iter *iter) in nfc_device_iter_init() argument
115 class_dev_iter_init(iter, &nfc_class, NULL, NULL); in nfc_device_iter_init()
118 static inline struct nfc_dev *nfc_device_iter_next(struct class_dev_iter *iter) in nfc_device_iter_next() argument
120 struct device *d = class_dev_iter_next(iter); in nfc_device_iter_next()
127 static inline void nfc_device_iter_exit(struct class_dev_iter *iter) in nfc_device_iter_exit() argument
129 class_dev_iter_exit(iter); in nfc_device_iter_exit()
Dnetlink.c580 struct class_dev_iter *iter = (struct class_dev_iter *) cb->args[0]; in nfc_genl_dump_devices() local
584 if (!iter) { in nfc_genl_dump_devices()
586 iter = kmalloc(sizeof(struct class_dev_iter), GFP_KERNEL); in nfc_genl_dump_devices()
587 if (!iter) in nfc_genl_dump_devices()
589 cb->args[0] = (long) iter; in nfc_genl_dump_devices()
597 nfc_device_iter_init(iter); in nfc_genl_dump_devices()
598 dev = nfc_device_iter_next(iter); in nfc_genl_dump_devices()
609 dev = nfc_device_iter_next(iter); in nfc_genl_dump_devices()
621 struct class_dev_iter *iter = (struct class_dev_iter *) cb->args[0]; in nfc_genl_dump_devices_done() local
623 nfc_device_iter_exit(iter); in nfc_genl_dump_devices_done()
[all …]
/linux-4.1.27/fs/
Ddax.c101 static ssize_t dax_io(struct inode *inode, struct iov_iter *iter, in dax_io() argument
112 if (iov_iter_rw(iter) != WRITE) in dax_io()
127 iov_iter_rw(iter) == WRITE); in dax_io()
140 hole = iov_iter_rw(iter) != WRITE && !buffer_written(bh); in dax_io()
157 if (iov_iter_rw(iter) == WRITE) in dax_io()
158 len = copy_from_iter(addr, max - pos, iter); in dax_io()
160 len = copy_to_iter(addr, max - pos, iter); in dax_io()
162 len = iov_iter_zero(max - pos, iter); in dax_io()
192 struct iov_iter *iter, loff_t pos, get_block_t get_block, in dax_do_io() argument
197 loff_t end = pos + iov_iter_count(iter); in dax_do_io()
[all …]
Dread_write.c332 ssize_t vfs_iter_read(struct file *file, struct iov_iter *iter, loff_t *ppos) in vfs_iter_read() argument
343 iter->type |= READ; in vfs_iter_read()
344 ret = file->f_op->read_iter(&kiocb, iter); in vfs_iter_read()
352 ssize_t vfs_iter_write(struct file *file, struct iov_iter *iter, loff_t *ppos) in vfs_iter_write() argument
363 iter->type |= WRITE; in vfs_iter_write()
364 ret = file->f_op->write_iter(&kiocb, iter); in vfs_iter_write()
415 struct iov_iter iter; in new_sync_read() local
420 iov_iter_init(&iter, READ, &iov, 1, len); in new_sync_read()
422 ret = filp->f_op->read_iter(&kiocb, &iter); in new_sync_read()
471 struct iov_iter iter; in new_sync_write() local
[all …]
Ddirect-io.c98 struct iov_iter *iter; member
160 ret = iov_iter_get_pages(sdio->iter, dio->pages, LONG_MAX, DIO_PAGES, in dio_refill_pages()
182 iov_iter_advance(sdio->iter, ret); in dio_refill_pages()
1099 struct block_device *bdev, struct iov_iter *iter, in do_blockdev_direct_IO() argument
1107 size_t count = iov_iter_count(iter); in do_blockdev_direct_IO()
1113 unsigned long align = offset | iov_iter_alignment(iter); in do_blockdev_direct_IO()
1129 if (iov_iter_rw(iter) == READ && !iov_iter_count(iter)) in do_blockdev_direct_IO()
1145 if (iov_iter_rw(iter) == READ) { in do_blockdev_direct_IO()
1164 if (iov_iter_rw(iter) == READ && offset >= dio->i_size) { in do_blockdev_direct_IO()
1181 iov_iter_rw(iter) == WRITE && end > i_size_read(inode)) in do_blockdev_direct_IO()
[all …]
/linux-4.1.27/drivers/hid/
Dhid-wiimote-core.c619 const __u8 *mods, *iter; in wiimote_modules_load() local
625 for (iter = mods; *iter != WIIMOD_NULL; ++iter) { in wiimote_modules_load()
626 if (wiimod_table[*iter]->flags & WIIMOD_FLAG_INPUT) { in wiimote_modules_load()
646 for (iter = mods; *iter != WIIMOD_NULL; ++iter) { in wiimote_modules_load()
647 ops = wiimod_table[*iter]; in wiimote_modules_load()
668 for ( ; iter-- != mods; ) { in wiimote_modules_load()
669 ops = wiimod_table[*iter]; in wiimote_modules_load()
682 const __u8 *mods, *iter; in wiimote_modules_unload() local
693 for (iter = mods; *iter != WIIMOD_NULL; ++iter) in wiimote_modules_unload()
701 for ( ; iter-- != mods; ) { in wiimote_modules_unload()
[all …]
/linux-4.1.27/drivers/block/aoe/
Daoecmd.c199 memset(&f->iter, 0, sizeof(f->iter)); in aoe_freetframe()
297 skb_fillup(struct sk_buff *skb, struct bio *bio, struct bvec_iter iter) in skb_fillup() argument
302 __bio_for_each_segment(bv, bio, iter, iter) in skb_fillup()
343 ah->scnt = f->iter.bi_size >> 9; in ata_rw_frameinit()
344 put_lba(ah, f->iter.bi_sector); in ata_rw_frameinit()
353 skb_fillup(skb, f->buf->bio, f->iter); in ata_rw_frameinit()
355 skb->len += f->iter.bi_size; in ata_rw_frameinit()
356 skb->data_len = f->iter.bi_size; in ata_rw_frameinit()
357 skb->truesize += f->iter.bi_size; in ata_rw_frameinit()
385 f->iter = buf->iter; in aoecmd_ata_rw()
[all …]
Daoe.h104 struct bvec_iter iter; member
123 struct bvec_iter iter; member
/linux-4.1.27/arch/powerpc/kernel/
Dcacheinfo.c166 struct cache *iter; in release_cache_debugcheck() local
168 list_for_each_entry(iter, &cache_list, list) in release_cache_debugcheck()
169 WARN_ONCE(iter->next_local == cache, in release_cache_debugcheck()
171 iter->ofnode->full_name, in release_cache_debugcheck()
172 cache_type_string(iter), in release_cache_debugcheck()
304 struct cache *iter; in cache_find_first_sibling() local
310 list_for_each_entry(iter, &cache_list, list) in cache_find_first_sibling()
311 if (iter->ofnode == cache->ofnode && iter->next_local == cache) in cache_find_first_sibling()
312 return iter; in cache_find_first_sibling()
321 struct cache *iter; in cache_lookup_by_node() local
[all …]
Dftrace.c443 struct ftrace_rec_iter *iter; in ftrace_replace_code() local
447 for (iter = ftrace_rec_iter_start(); iter; in ftrace_replace_code()
448 iter = ftrace_rec_iter_next(iter)) { in ftrace_replace_code()
449 rec = ftrace_rec_iter_record(iter); in ftrace_replace_code()
/linux-4.1.27/net/appletalk/
Daarp.c921 static struct aarp_entry *iter_next(struct aarp_iter_state *iter, loff_t *pos) in iter_next() argument
923 int ct = iter->bucket; in iter_next()
924 struct aarp_entry **table = iter->table; in iter_next()
932 iter->table = table; in iter_next()
933 iter->bucket = ct; in iter_next()
956 struct aarp_iter_state *iter = seq->private; in aarp_seq_start() local
959 iter->table = resolved; in aarp_seq_start()
960 iter->bucket = 0; in aarp_seq_start()
962 return *pos ? iter_next(iter, pos) : SEQ_START_TOKEN; in aarp_seq_start()
968 struct aarp_iter_state *iter = seq->private; in aarp_seq_next() local
[all …]
/linux-4.1.27/arch/arm/include/asm/hardware/
Diop3xx-adma.h576 struct iop3xx_desc_aau *hw_desc, *prev_hw_desc, *iter; in iop_desc_init_zero_sum() local
587 iter = iop_hw_desc_slot_idx(hw_desc, i); in iop_desc_init_zero_sum()
588 u_desc_ctrl.value = iop3xx_desc_init_xor(iter, src_cnt, flags); in iop_desc_init_zero_sum()
592 iter->desc_ctrl = u_desc_ctrl.value; in iop_desc_init_zero_sum()
695 struct iop3xx_desc_aau *hw_desc = desc->hw_desc, *iter; in iop_desc_set_zero_sum_byte_count() local
702 iter = iop_hw_desc_slot_idx(hw_desc, i); in iop_desc_set_zero_sum_byte_count()
703 iter->byte_count = IOP_ADMA_ZERO_SUM_MAX_BYTE_COUNT; in iop_desc_set_zero_sum_byte_count()
708 iter = iop_hw_desc_slot_idx(hw_desc, i); in iop_desc_set_zero_sum_byte_count()
709 iter->byte_count = len; in iop_desc_set_zero_sum_byte_count()
744 struct iop3xx_desc_aau *hw_desc = desc->hw_desc, *iter; in iop_desc_set_zero_sum_src_addr() local
[all …]
/linux-4.1.27/drivers/net/bonding/
Dbond_main.c290 struct list_head *iter; in bond_vlan_rx_add_vid() local
293 bond_for_each_slave(bond, slave, iter) { in bond_vlan_rx_add_vid()
303 bond_for_each_slave(bond, rollback_slave, iter) { in bond_vlan_rx_add_vid()
322 struct list_head *iter; in bond_vlan_rx_kill_vid() local
325 bond_for_each_slave(bond, slave, iter) in bond_vlan_rx_kill_vid()
344 struct list_head *iter; in bond_set_carrier() local
353 bond_for_each_slave(bond, slave, iter) { in bond_set_carrier()
496 struct list_head *iter; in bond_set_promiscuity() local
507 bond_for_each_slave(bond, slave, iter) { in bond_set_promiscuity()
519 struct list_head *iter; in bond_set_allmulti() local
[all …]
Dbond_procfs.c13 struct list_head *iter; in bond_info_seq_start() local
22 bond_for_each_slave_rcu(bond, slave, iter) in bond_info_seq_start()
32 struct list_head *iter; in bond_info_seq_next() local
40 bond_for_each_slave_rcu(bond, slave, iter) { in bond_info_seq_next()
Dbond_3ad.c701 struct list_head *iter; in __get_active_agg() local
704 bond_for_each_slave_rcu(bond, slave, iter) in __get_active_agg()
1311 struct list_head *iter; in ad_port_selection_logic() local
1375 bond_for_each_slave(bond, slave, iter) { in ad_port_selection_logic()
1583 struct list_head *iter; in ad_agg_selection_logic() local
1592 bond_for_each_slave_rcu(bond, slave, iter) { in ad_agg_selection_logic()
1637 bond_for_each_slave_rcu(bond, slave, iter) { in ad_agg_selection_logic()
1997 struct list_head *iter; in bond_3ad_unbind_slave() local
2030 bond_for_each_slave(bond, slave_iter, iter) { in bond_3ad_unbind_slave()
2107 bond_for_each_slave(bond, slave_iter, iter) { in bond_3ad_unbind_slave()
[all …]
Dbond_alb.c198 struct list_head *iter; in tlb_get_least_loaded_slave() local
205 bond_for_each_slave_rcu(bond, slave, iter) { in tlb_get_least_loaded_slave()
338 struct list_head *iter; in __rlb_next_rx_slave() local
341 bond_for_each_slave_rcu(bond, slave, iter) { in __rlb_next_rx_slave()
961 struct list_head *iter; in alb_send_learning_packets() local
971 netdev_for_each_all_upper_dev_rcu(bond->dev, upper, iter) { in alb_send_learning_packets()
1143 struct list_head *iter; in alb_handle_addr_collision_on_attach() local
1167 bond_for_each_slave(bond, tmp_slave1, iter) { in alb_handle_addr_collision_on_attach()
1217 struct list_head *iter; in alb_set_mac_address() local
1225 bond_for_each_slave(bond, slave, iter) { in alb_set_mac_address()
[all …]
Dbond_options.c911 struct list_head *iter; in _bond_options_arp_ip_target_set() local
915 bond_for_each_slave(bond, slave, iter) in _bond_options_arp_ip_target_set()
959 struct list_head *iter; in bond_option_arp_ip_target_rem() local
982 bond_for_each_slave(bond, slave, iter) { in bond_option_arp_ip_target_rem()
1060 struct list_head *iter; in bond_option_primary_set() local
1077 bond_for_each_slave(bond, slave, iter) { in bond_option_primary_set()
1160 struct list_head *iter; in bond_option_all_slaves_active_set() local
1166 bond_for_each_slave(bond, slave, iter) { in bond_option_all_slaves_active_set()
1241 struct list_head *iter; in bond_option_queue_id_set() local
1270 bond_for_each_slave(bond, slave, iter) { in bond_option_queue_id_set()
/linux-4.1.27/arch/arm/mach-iop13xx/include/mach/
Dadma.h372 struct iop13xx_adma_desc_hw *hw_desc = desc->hw_desc, *iter; in iop_desc_set_zero_sum_byte_count() local
379 iter = iop_hw_desc_slot_idx(hw_desc, i); in iop_desc_set_zero_sum_byte_count()
380 iter->byte_count = IOP_ADMA_ZERO_SUM_MAX_BYTE_COUNT; in iop_desc_set_zero_sum_byte_count()
386 iter = iop_hw_desc_slot_idx(hw_desc, i); in iop_desc_set_zero_sum_byte_count()
387 iter->byte_count = len; in iop_desc_set_zero_sum_byte_count()
425 struct iop13xx_adma_desc_hw *hw_desc = desc->hw_desc, *iter; in iop_desc_set_xor_src_addr() local
429 iter = iop_hw_desc_slot_idx(hw_desc, i); in iop_desc_set_xor_src_addr()
430 iter->src[src_idx].src_addr = addr; in iop_desc_set_xor_src_addr()
431 iter->src[src_idx].upper_src_addr = 0; in iop_desc_set_xor_src_addr()
445 struct iop13xx_adma_desc_hw *hw_desc = desc->hw_desc, *iter; in iop_desc_set_pq_src_addr() local
[all …]
/linux-4.1.27/net/irda/
Dirlmp.c1834 static void *irlmp_seq_hb_idx(struct irlmp_iter_state *iter, loff_t *off) in irlmp_seq_hb_idx() argument
1838 spin_lock_irq(&iter->hashbin->hb_spinlock); in irlmp_seq_hb_idx()
1839 for (element = hashbin_get_first(iter->hashbin); in irlmp_seq_hb_idx()
1841 element = hashbin_get_next(iter->hashbin)) { in irlmp_seq_hb_idx()
1847 spin_unlock_irq(&iter->hashbin->hb_spinlock); in irlmp_seq_hb_idx()
1848 iter->hashbin = NULL; in irlmp_seq_hb_idx()
1855 struct irlmp_iter_state *iter = seq->private; in irlmp_seq_start() local
1859 iter->hashbin = NULL; in irlmp_seq_start()
1863 iter->hashbin = irlmp->unconnected_lsaps; in irlmp_seq_start()
1864 v = irlmp_seq_hb_idx(iter, &off); in irlmp_seq_start()
[all …]
Dirlap.c1075 struct irlap_iter_state *iter = seq->private; in irlap_seq_start() local
1080 iter->id = 0; in irlap_seq_start()
1084 if (iter->id == *pos) in irlap_seq_start()
1086 ++iter->id; in irlap_seq_start()
1094 struct irlap_iter_state *iter = seq->private; in irlap_seq_next() local
1097 ++iter->id; in irlap_seq_next()
1108 const struct irlap_iter_state *iter = seq->private; in irlap_seq_show() local
1113 seq_printf(seq, "irlap%d ", iter->id); in irlap_seq_show()
/linux-4.1.27/drivers/gpu/drm/vmwgfx/
Dvmwgfx_gmr.c39 struct vmw_piter *iter, in vmw_gmr2_bind() argument
86 *cmd = vmw_piter_dma_addr(iter) >> PAGE_SHIFT; in vmw_gmr2_bind()
88 *((uint64_t *)cmd) = vmw_piter_dma_addr(iter) >> in vmw_gmr2_bind()
92 vmw_piter_next(iter); in vmw_gmr2_bind()
Dvmwgfx_buffer.c233 return __sg_page_iter_next(&viter->iter); in __vmw_piter_sg_next()
253 return sg_page_iter_page(&viter->iter); in __vmw_piter_sg_page()
278 return sg_page_iter_dma_address(&viter->iter); in __vmw_piter_sg_addr()
316 __sg_page_iter_start(&viter->iter, vsgt->sgt->sgl, in vmw_piter_start()
384 struct vmw_piter iter; in vmw_ttm_map_dma() local
440 for (vmw_piter_start(&iter, vsgt, 0); vmw_piter_next(&iter);) { in vmw_ttm_map_dma()
441 dma_addr_t cur = vmw_piter_dma_addr(&iter); in vmw_ttm_map_dma()
Dvmwgfx_mob.c104 struct vmw_piter iter; in vmw_setup_otable_base() local
110 vmw_piter_start(&iter, vsgt, offset >> PAGE_SHIFT); in vmw_setup_otable_base()
111 WARN_ON(!vmw_piter_next(&iter)); in vmw_setup_otable_base()
121 mob->pt_root_page = vmw_piter_dma_addr(&iter); in vmw_setup_otable_base()
124 mob->pt_root_page = vmw_piter_dma_addr(&iter); in vmw_setup_otable_base()
130 vmw_mob_pt_setup(mob, iter, otable->size >> PAGE_SHIFT); in vmw_setup_otable_base()
/linux-4.1.27/scripts/coccinelle/null/
Dderef_null.cocci57 iterator iter;
66 iter(subE,...) S4 // no use
118 iterator iter;
127 iter(subE,...) S4 // no use
178 iterator iter;
187 iter(subE,...) S4 // no use
251 iterator iter;
260 iter(subE,...) S4 // no use
/linux-4.1.27/net/ipv4/netfilter/
Darp_tables.c625 struct arpt_entry *iter; in translate_table() local
642 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table()
643 ret = check_entry_size_and_hooks(iter, newinfo, entry0, in translate_table()
651 if (strcmp(arpt_get_target(iter)->u.user.name, in translate_table()
689 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table()
690 ret = find_check_entry(iter, repl->name, repl->size); in translate_table()
697 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table()
700 cleanup_entry(iter); in translate_table()
717 struct arpt_entry *iter; in get_counters() local
725 xt_entry_foreach(iter, t->entries[cpu], t->size) { in get_counters()
[all …]
Dip_tables.c260 const struct ipt_entry *iter; in trace_packet() local
270 xt_entry_foreach(iter, root, private->size - private->hook_entry[hook]) in trace_packet()
271 if (get_chainname_rulenum(iter, e, hookname, in trace_packet()
795 struct ipt_entry *iter; in translate_table() local
811 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table()
812 ret = check_entry_size_and_hooks(iter, newinfo, entry0, in translate_table()
820 if (strcmp(ipt_get_target(iter)->u.user.name, in translate_table()
853 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table()
854 ret = find_check_entry(iter, net, repl->name, repl->size); in translate_table()
861 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table()
[all …]
/linux-4.1.27/net/ipv6/netfilter/
Dip6_tables.c289 const struct ip6t_entry *iter; in trace_packet() local
299 xt_entry_foreach(iter, root, private->size - private->hook_entry[hook]) in trace_packet()
300 if (get_chainname_rulenum(iter, e, hookname, in trace_packet()
808 struct ip6t_entry *iter; in translate_table() local
824 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table()
825 ret = check_entry_size_and_hooks(iter, newinfo, entry0, in translate_table()
833 if (strcmp(ip6t_get_target(iter)->u.user.name, in translate_table()
866 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table()
867 ret = find_check_entry(iter, net, repl->name, repl->size); in translate_table()
874 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table()
[all …]
/linux-4.1.27/kernel/sched/
Drt.c459 #define for_each_rt_rq(rt_rq, iter, rq) \ argument
460 for (iter = container_of(&task_groups, typeof(*iter), list); \
461 (iter = next_task_group(iter)) && \
462 (rt_rq = iter->rt_rq[cpu_of(rq)]);)
563 #define for_each_rt_rq(rt_rq, iter, rq) \ argument
564 for ((void) iter, rt_rq = &rq->rt; rt_rq; rt_rq = NULL)
637 struct rt_rq *iter = sched_rt_period_rt_rq(rt_b, i); in do_balance_runtime() local
640 if (iter == rt_rq) in do_balance_runtime()
643 raw_spin_lock(&iter->rt_runtime_lock); in do_balance_runtime()
649 if (iter->rt_runtime == RUNTIME_INF) in do_balance_runtime()
[all …]
/linux-4.1.27/drivers/scsi/qla2xxx/
Dqla_inline.h75 uint32_t iter = bsize >> 2; in host_to_fcp_swap() local
77 for (; iter ; iter--) in host_to_fcp_swap()
88 uint32_t iter = bsize >> 2; in host_to_adap() local
90 for (; iter ; iter--) in host_to_adap()
Dqla_attr.c270 uint32_t *iter; in qla2x00_sysfs_write_nvram() local
273 iter = (uint32_t *)buf; in qla2x00_sysfs_write_nvram()
276 chksum += le32_to_cpu(*iter++); in qla2x00_sysfs_write_nvram()
278 *iter = cpu_to_le32(chksum); in qla2x00_sysfs_write_nvram()
280 uint8_t *iter; in qla2x00_sysfs_write_nvram() local
283 iter = (uint8_t *)buf; in qla2x00_sysfs_write_nvram()
286 chksum += *iter++; in qla2x00_sysfs_write_nvram()
288 *iter = chksum; in qla2x00_sysfs_write_nvram()
639 uint16_t iter, addr, offset; in qla2x00_sysfs_read_sfp() local
659 for (iter = 0, offset = 0; iter < (SFP_DEV_SIZE * 2) / SFP_BLOCK_SIZE; in qla2x00_sysfs_read_sfp()
[all …]
Dqla_sup.c2665 uint32_t istart, iend, iter, vend; in qla2x00_get_fcode_version() local
2678 iter = istart; in qla2x00_get_fcode_version()
2679 while ((iter < iend) && !do_next) { in qla2x00_get_fcode_version()
2680 iter++; in qla2x00_get_fcode_version()
2681 if (qla2x00_read_flash_byte(ha, iter) == '/') { in qla2x00_get_fcode_version()
2682 if (qla2x00_read_flash_byte(ha, iter + 2) == in qla2x00_get_fcode_version()
2686 iter + 3) == '/') in qla2x00_get_fcode_version()
2695 while ((iter > istart) && !do_next) { in qla2x00_get_fcode_version()
2696 iter--; in qla2x00_get_fcode_version()
2697 if (qla2x00_read_flash_byte(ha, iter) == ' ') in qla2x00_get_fcode_version()
[all …]
/linux-4.1.27/drivers/scsi/
Dsd_dif.c127 struct bvec_iter iter; in sd_dif_prepare() local
136 bip_for_each_vec(iv, bip, iter) { in sd_dif_prepare()
179 struct bvec_iter iter; in sd_dif_complete() local
183 bip_for_each_vec(iv, bip, iter) { in sd_dif_complete()
/linux-4.1.27/drivers/infiniband/hw/ehca/
Dehca_cq.c75 struct hlist_node *iter; in ehca_cq_unassign_qp() local
80 hlist_for_each(iter, &cq->qp_hashtab[key]) { in ehca_cq_unassign_qp()
81 qp = hlist_entry(iter, struct ehca_qp, list_entries); in ehca_cq_unassign_qp()
83 hlist_del(iter); in ehca_cq_unassign_qp()
104 struct hlist_node *iter; in ehca_cq_get_qp() local
106 hlist_for_each(iter, &cq->qp_hashtab[key]) { in ehca_cq_get_qp()
107 qp = hlist_entry(iter, struct ehca_qp, list_entries); in ehca_cq_get_qp()
/linux-4.1.27/scripts/kconfig/
Dzconf.lex.c2355 struct file *iter; in zconf_nextfile() local
2371 for (iter = current_file->parent; iter; iter = iter->parent ) { in zconf_nextfile()
2372 if (!strcmp(current_file->name,iter->name) ) { in zconf_nextfile()
2377 iter = current_file->parent; in zconf_nextfile()
2378 while (iter && \ in zconf_nextfile()
2379 strcmp(iter->name,current_file->name)) { in zconf_nextfile()
2381 iter->name, iter->lineno-1); in zconf_nextfile()
2382 iter = iter->parent; in zconf_nextfile()
2384 if (iter) in zconf_nextfile()
2386 iter->name, iter->lineno+1); in zconf_nextfile()
Dgconf.c784 GtkTreeIter iter; in renderer_edited() local
789 if (!gtk_tree_model_get_iter(model2, &iter, path)) in renderer_edited()
792 gtk_tree_model_get(model2, &iter, COL_MENU, &menu, -1); in renderer_edited()
795 gtk_tree_model_get(model2, &iter, COL_VALUE, &old_def, -1); in renderer_edited()
886 GtkTreeIter iter; in on_treeview2_button_press_event() local
903 if (!gtk_tree_model_get_iter(model2, &iter, path)) in on_treeview2_button_press_event()
905 gtk_tree_model_get(model2, &iter, COL_MENU, &menu, -1); in on_treeview2_button_press_event()
943 GtkTreeIter iter; in on_treeview2_key_press_event() local
963 gtk_tree_model_get_iter(model2, &iter, path); in on_treeview2_key_press_event()
964 gtk_tree_model_get(model2, &iter, COL_MENU, &menu, -1); in on_treeview2_key_press_event()
[all …]
/linux-4.1.27/kernel/events/
Dhw_breakpoint.c120 struct perf_event *iter; in task_bp_pinned() local
123 list_for_each_entry(iter, &bp_task_head, hw.bp_list) { in task_bp_pinned()
124 if (iter->hw.target == tsk && in task_bp_pinned()
125 find_slot_idx(iter) == type && in task_bp_pinned()
126 (iter->cpu < 0 || cpu == iter->cpu)) in task_bp_pinned()
127 count += hw_breakpoint_weight(iter); in task_bp_pinned()
/linux-4.1.27/arch/x86/kernel/
Dftrace.c541 struct ftrace_rec_iter *iter; in ftrace_replace_code() local
547 for_ftrace_rec_iter(iter) { in ftrace_replace_code()
548 rec = ftrace_rec_iter_record(iter); in ftrace_replace_code()
560 for_ftrace_rec_iter(iter) { in ftrace_replace_code()
561 rec = ftrace_rec_iter_record(iter); in ftrace_replace_code()
572 for_ftrace_rec_iter(iter) { in ftrace_replace_code()
573 rec = ftrace_rec_iter_record(iter); in ftrace_replace_code()
587 for_ftrace_rec_iter(iter) { in ftrace_replace_code()
588 rec = ftrace_rec_iter_record(iter); in ftrace_replace_code()
/linux-4.1.27/kernel/locking/
Dlockdep_proc.c569 struct lock_stat_data *iter; in ls_start() local
574 iter = data->stats + (*pos - 1); in ls_start()
575 if (iter >= data->iter_end) in ls_start()
576 iter = NULL; in ls_start()
578 return iter; in ls_start()
619 struct lock_stat_data *iter = data->stats; in lock_stat_open() local
623 iter->class = class; in lock_stat_open()
624 iter->stats = lock_stats(class); in lock_stat_open()
625 iter++; in lock_stat_open()
627 data->iter_end = iter; in lock_stat_open()
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4vf/
Dt4vf_common.h228 #define for_each_port(adapter, iter) \ argument
229 for (iter = 0; iter < (adapter)->params.nports; iter++)
Dadapter.h348 #define for_each_ethrxq(sge, iter) \ argument
349 for (iter = 0; iter < (sge)->ethqsets; iter++)
/linux-4.1.27/drivers/gpu/drm/i915/
Di915_gem_gtt.h344 #define gen6_for_each_pde(pt, pd, start, length, temp, iter) \ argument
345 for (iter = gen6_pde_index(start); \
346 pt = (pd)->page_table[iter], length > 0 && iter < I915_PDES; \
347 iter++, \
/linux-4.1.27/drivers/net/ethernet/qlogic/qlge/
Dqlge_ethtool.c247 u64 *iter = &qdev->nic_stats.tx_pkts; in ql_update_stats() local
265 *iter = data; in ql_update_stats()
266 iter++; in ql_update_stats()
279 *iter = data; in ql_update_stats()
280 iter++; in ql_update_stats()
284 iter += QLGE_RCV_MAC_ERR_STATS; in ql_update_stats()
296 *iter = data; in ql_update_stats()
297 iter++; in ql_update_stats()
310 *iter = data; in ql_update_stats()
311 iter++; in ql_update_stats()
[all …]
/linux-4.1.27/fs/f2fs/
Dtrace.c129 struct radix_tree_iter iter; in gang_lookup_pids() local
136 radix_tree_for_each_slot(slot, &pids, &iter, first_index) { in gang_lookup_pids()
137 results[ret] = iter.index; in gang_lookup_pids()
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb/
Dcommon.h314 #define for_each_port(adapter, iter) \ argument
315 for (iter = 0; iter < (adapter)->params.nports; ++iter)
/linux-4.1.27/fs/ocfs2/dlm/
Ddlmmaster.c743 struct dlm_node_iter iter; in dlm_get_lock_resource() local
956 dlm_node_iter_init(mle->vote_map, &iter); in dlm_get_lock_resource()
957 while ((nodenum = dlm_node_iter_next(&iter)) >= 0) { in dlm_get_lock_resource()
1191 static void dlm_bitmap_diff_iter_init(struct dlm_bitmap_diff_iter *iter, in dlm_bitmap_diff_iter_init() argument
1198 iter->curnode = -1; in dlm_bitmap_diff_iter_init()
1199 iter->orig_bm = orig_bm; in dlm_bitmap_diff_iter_init()
1200 iter->cur_bm = cur_bm; in dlm_bitmap_diff_iter_init()
1203 p1 = *(iter->orig_bm + i); in dlm_bitmap_diff_iter_init()
1204 p2 = *(iter->cur_bm + i); in dlm_bitmap_diff_iter_init()
1205 iter->diff_bm[i] = (p1 & ~p2) | (p2 & ~p1); in dlm_bitmap_diff_iter_init()
[all …]
Ddlmcommon.h1113 struct dlm_node_iter *iter) in dlm_node_iter_init() argument
1115 memcpy(iter->node_map, map, sizeof(iter->node_map)); in dlm_node_iter_init()
1116 iter->curnode = -1; in dlm_node_iter_init()
1119 static inline int dlm_node_iter_next(struct dlm_node_iter *iter) in dlm_node_iter_next() argument
1122 bit = find_next_bit(iter->node_map, O2NM_MAX_NODES, iter->curnode+1); in dlm_node_iter_next()
1124 iter->curnode = O2NM_MAX_NODES; in dlm_node_iter_next()
1127 iter->curnode = bit; in dlm_node_iter_next()
/linux-4.1.27/tools/perf/
Dbuiltin-report.c89 static int hist_iter__report_callback(struct hist_entry_iter *iter, in hist_iter__report_callback() argument
95 struct hist_entry *he = iter->he; in hist_iter__report_callback()
96 struct perf_evsel *evsel = iter->evsel; in hist_iter__report_callback()
139 struct hist_entry_iter iter = { in process_sample_event() local
158 iter.ops = &hist_iter_branch; in process_sample_event()
160 iter.ops = &hist_iter_mem; in process_sample_event()
162 iter.ops = &hist_iter_cumulative; in process_sample_event()
164 iter.ops = &hist_iter_normal; in process_sample_event()
169 ret = hist_entry_iter__add(&iter, &al, evsel, sample, rep->max_stack, in process_sample_event()
/linux-4.1.27/fs/9p/
Dvfs_addr.c248 v9fs_direct_IO(struct kiocb *iocb, struct iov_iter *iter, loff_t pos) in v9fs_direct_IO() argument
253 if (iov_iter_rw(iter) == WRITE) { in v9fs_direct_IO()
254 n = p9_client_write(file->private_data, pos, iter, &err); in v9fs_direct_IO()
262 n = p9_client_read(file->private_data, pos, iter, &err); in v9fs_direct_IO()
/linux-4.1.27/arch/x86/kvm/
Dmmu.c1078 static u64 *rmap_get_first(unsigned long rmap, struct rmap_iterator *iter) in rmap_get_first() argument
1084 iter->desc = NULL; in rmap_get_first()
1088 iter->desc = (struct pte_list_desc *)(rmap & ~1ul); in rmap_get_first()
1089 iter->pos = 0; in rmap_get_first()
1090 return iter->desc->sptes[iter->pos]; in rmap_get_first()
1098 static u64 *rmap_get_next(struct rmap_iterator *iter) in rmap_get_next() argument
1100 if (iter->desc) { in rmap_get_next()
1101 if (iter->pos < PTE_LIST_EXT - 1) { in rmap_get_next()
1104 ++iter->pos; in rmap_get_next()
1105 sptep = iter->desc->sptes[iter->pos]; in rmap_get_next()
[all …]
Dmmu_audit.c193 struct rmap_iterator iter; in audit_write_protection() local
200 for (sptep = rmap_get_first(*rmapp, &iter); sptep; in audit_write_protection()
201 sptep = rmap_get_next(&iter)) { in audit_write_protection()
/linux-4.1.27/include/trace/
Dftrace.h307 ftrace_raw_output_##call(struct trace_iterator *iter, int flags, \
310 struct trace_seq *s = &iter->seq; \
311 struct trace_seq __maybe_unused *p = &iter->tmp_seq; \
315 field = (typeof(field))iter->ent; \
317 ret = ftrace_raw_output_prep(iter, trace_event); \
332 ftrace_raw_output_##call(struct trace_iterator *iter, int flags, \
337 struct trace_seq *p = &iter->tmp_seq; \
339 entry = iter->ent; \
349 return ftrace_output_call(iter, #call, print); \
/linux-4.1.27/net/core/
Ddev.c1461 struct list_head *iter; in dev_disable_lro() local
1469 netdev_for_each_lower_dev(dev, lower_dev, iter) in dev_disable_lro()
4829 struct list_head **iter) in netdev_upper_get_next_dev_rcu() argument
4835 upper = list_entry_rcu((*iter)->next, struct netdev_adjacent, list); in netdev_upper_get_next_dev_rcu()
4840 *iter = &upper->list; in netdev_upper_get_next_dev_rcu()
4855 struct list_head **iter) in netdev_all_upper_get_next_dev_rcu() argument
4861 upper = list_entry_rcu((*iter)->next, struct netdev_adjacent, list); in netdev_all_upper_get_next_dev_rcu()
4866 *iter = &upper->list; in netdev_all_upper_get_next_dev_rcu()
4884 struct list_head **iter) in netdev_lower_get_next_private() argument
4888 lower = list_entry(*iter, struct netdev_adjacent, list); in netdev_lower_get_next_private()
[all …]
/linux-4.1.27/arch/sh/mm/
Dpmb.c145 struct pmb_entry *pmbe, *iter; in pmb_mapping_exists() local
175 for (iter = pmbe->link; iter; iter = iter->link) in pmb_mapping_exists()
176 span += iter->size; in pmb_mapping_exists()
815 static int pmb_seq_show(struct seq_file *file, void *iter) in pmb_seq_show() argument
Dasids-debugfs.c27 static int asids_seq_show(struct seq_file *file, void *iter) in asids_seq_show() argument
/linux-4.1.27/fs/jfs/
Dinode.c333 static ssize_t jfs_direct_IO(struct kiocb *iocb, struct iov_iter *iter, in jfs_direct_IO() argument
339 size_t count = iov_iter_count(iter); in jfs_direct_IO()
342 ret = blockdev_direct_IO(iocb, inode, iter, offset, jfs_get_block); in jfs_direct_IO()
348 if (unlikely(iov_iter_rw(iter) == WRITE && ret < 0)) { in jfs_direct_IO()
/linux-4.1.27/security/selinux/ss/
Debitmap.c94 unsigned int iter; in ebitmap_netlbl_export() local
108 for (iter = 0; iter < EBITMAP_UNIT_NUMS; iter++) { in ebitmap_netlbl_export()
109 e_map = e_iter->maps[iter]; in ebitmap_netlbl_export()
/linux-4.1.27/fs/proc/
Dbase.c2816 static struct tgid_iter next_tgid(struct pid_namespace *ns, struct tgid_iter iter) in next_tgid() argument
2820 if (iter.task) in next_tgid()
2821 put_task_struct(iter.task); in next_tgid()
2824 iter.task = NULL; in next_tgid()
2825 pid = find_ge_pid(iter.tgid, ns); in next_tgid()
2827 iter.tgid = pid_nr_ns(pid, ns); in next_tgid()
2828 iter.task = pid_task(pid, PIDTYPE_PID); in next_tgid()
2841 if (!iter.task || !has_group_leader_pid(iter.task)) { in next_tgid()
2842 iter.tgid += 1; in next_tgid()
2845 get_task_struct(iter.task); in next_tgid()
[all …]
/linux-4.1.27/drivers/block/
Dps3disk.c96 struct req_iterator iter; in ps3disk_scatter_gather() local
102 rq_for_each_segment(bvec, req, iter) { in ps3disk_scatter_gather()
105 __func__, __LINE__, i, bio_sectors(iter.bio), in ps3disk_scatter_gather()
106 iter.bio->bi_iter.bi_sector); in ps3disk_scatter_gather()
133 struct req_iterator iter; in ps3disk_submit_request_sg() local
135 rq_for_each_segment(bv, req, iter) in ps3disk_submit_request_sg()
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/engine/pm/
Dbase.c123 args->v0.version, args->v0.iter); in nvkm_perfctr_query()
124 di = (args->v0.iter & 0xff000000) >> 24; in nvkm_perfctr_query()
125 si = (args->v0.iter & 0x00ffffff) - 1; in nvkm_perfctr_query()
151 args->v0.iter = (di << 24) | ++si; in nvkm_perfctr_query()
160 args->v0.iter = 0xffffffff; in nvkm_perfctr_query()
/linux-4.1.27/tools/perf/tests/
Dhists_cumulate.c89 struct hist_entry_iter iter = { in add_hist_entries() local
94 iter.ops = &hist_iter_cumulative; in add_hist_entries()
96 iter.ops = &hist_iter_normal; in add_hist_entries()
107 if (hist_entry_iter__add(&iter, &al, evsel, &sample, in add_hist_entries()
/linux-4.1.27/drivers/iommu/
Ddmar.c547 struct acpi_dmar_header *iter, *next; in dmar_walk_remapping_entries() local
550 for (iter = start; iter < end && ret == 0; iter = next) { in dmar_walk_remapping_entries()
551 next = (void *)iter + iter->length; in dmar_walk_remapping_entries()
552 if (iter->length == 0) { in dmar_walk_remapping_entries()
564 dmar_table_print_dmar_entry(iter); in dmar_walk_remapping_entries()
566 if (iter->type >= ACPI_DMAR_TYPE_RESERVED) { in dmar_walk_remapping_entries()
569 iter->type); in dmar_walk_remapping_entries()
570 } else if (cb->cb[iter->type]) { in dmar_walk_remapping_entries()
571 ret = cb->cb[iter->type](iter, cb->arg[iter->type]); in dmar_walk_remapping_entries()
574 iter->type); in dmar_walk_remapping_entries()
/linux-4.1.27/fs/btrfs/tests/
Dbtrfs-tests.c130 struct radix_tree_iter iter; in btrfs_free_dummy_fs_info() local
135 radix_tree_for_each_slot(slot, &fs_info->buffer_radix, &iter, 0) { in btrfs_free_dummy_fs_info()
/linux-4.1.27/scripts/coccinelle/free/
Dkfree.cocci72 iterator iter;
80 iter(...,subE,...) S // no use
/linux-4.1.27/drivers/target/
Dtarget_core_file.c326 struct iov_iter iter; in fd_do_rw() local
346 iov_iter_bvec(&iter, ITER_BVEC, bvec, sgl_nents, len); in fd_do_rw()
348 ret = vfs_iter_write(fd, &iter, &pos); in fd_do_rw()
350 ret = vfs_iter_read(fd, &iter, &pos); in fd_do_rw()
435 struct iov_iter iter; in fd_execute_write_same() local
472 iov_iter_bvec(&iter, ITER_BVEC, bvec, nolb, len); in fd_execute_write_same()
473 ret = vfs_iter_write(fd_dev->fd_file, &iter, &pos); in fd_execute_write_same()
/linux-4.1.27/fs/ext4/
Dindirect.c645 ssize_t ext4_ind_direct_IO(struct kiocb *iocb, struct iov_iter *iter, in ext4_ind_direct_IO() argument
654 size_t count = iov_iter_count(iter); in ext4_ind_direct_IO()
657 if (iov_iter_rw(iter) == WRITE) { in ext4_ind_direct_IO()
679 if (iov_iter_rw(iter) == READ && ext4_should_dioread_nolock(inode)) { in ext4_ind_direct_IO()
693 ret = dax_do_io(iocb, inode, iter, offset, in ext4_ind_direct_IO()
697 inode->i_sb->s_bdev, iter, in ext4_ind_direct_IO()
704 ret = dax_do_io(iocb, inode, iter, offset, in ext4_ind_direct_IO()
707 ret = blockdev_direct_IO(iocb, inode, iter, offset, in ext4_ind_direct_IO()
710 if (unlikely(iov_iter_rw(iter) == WRITE && ret < 0)) { in ext4_ind_direct_IO()
/linux-4.1.27/drivers/net/
Dmacvtap.c800 struct iov_iter *iter) in macvtap_put_user() argument
810 if (iov_iter_count(iter) < vnet_hdr_len) in macvtap_put_user()
815 if (copy_to_iter(&vnet_hdr, sizeof(vnet_hdr), iter) != in macvtap_put_user()
819 iov_iter_advance(iter, vnet_hdr_len - sizeof(vnet_hdr)); in macvtap_put_user()
835 ret = skb_copy_datagram_iter(skb, 0, iter, vlan_offset); in macvtap_put_user()
836 if (ret || !iov_iter_count(iter)) in macvtap_put_user()
839 ret = copy_to_iter(&veth, sizeof(veth), iter); in macvtap_put_user()
840 if (ret != sizeof(veth) || !iov_iter_count(iter)) in macvtap_put_user()
844 ret = skb_copy_datagram_iter(skb, vlan_offset, iter, in macvtap_put_user()
/linux-4.1.27/drivers/iio/
Dindustrialio-trigger.c104 struct iio_trigger *trig = NULL, *iter; in iio_trigger_find_by_name() local
107 list_for_each_entry(iter, &iio_trigger_list, list) in iio_trigger_find_by_name()
108 if (sysfs_streq(iter->name, name)) { in iio_trigger_find_by_name()
109 trig = iter; in iio_trigger_find_by_name()
/linux-4.1.27/drivers/char/
Dmem.c634 static ssize_t read_iter_zero(struct kiocb *iocb, struct iov_iter *iter) in read_iter_zero() argument
638 while (iov_iter_count(iter)) { in read_iter_zero()
639 size_t chunk = iov_iter_count(iter), n; in read_iter_zero()
643 n = iov_iter_zero(chunk, iter); in read_iter_zero()
644 if (!n && iov_iter_count(iter)) in read_iter_zero()
/linux-4.1.27/drivers/phy/
Dphy-core.c465 struct class_dev_iter iter; in of_phy_simple_xlate() local
467 class_dev_iter_init(&iter, phy_class, NULL, NULL); in of_phy_simple_xlate()
468 while ((dev = class_dev_iter_next(&iter))) { in of_phy_simple_xlate()
473 class_dev_iter_exit(&iter); in of_phy_simple_xlate()
477 class_dev_iter_exit(&iter); in of_phy_simple_xlate()
/linux-4.1.27/arch/x86/kernel/cpu/
Dperf_event_intel_cqm.c830 struct perf_event *iter; in intel_cqm_setup_event() local
834 list_for_each_entry(iter, &cache_groups, hw.cqm_groups_entry) { in intel_cqm_setup_event()
835 rmid = iter->hw.cqm_rmid; in intel_cqm_setup_event()
837 if (__match_event(iter, event)) { in intel_cqm_setup_event()
840 *group = iter; in intel_cqm_setup_event()
848 if (__conflict_event(iter, event) && __rmid_valid(rmid)) in intel_cqm_setup_event()
/linux-4.1.27/drivers/cpufreq/
Dacpi-cpufreq.c956 struct freq_attr **iter; in acpi_cpufreq_init() local
960 for (iter = acpi_cpufreq_attr; *iter != NULL; iter++) in acpi_cpufreq_init()
964 if (iter[1] == NULL) in acpi_cpufreq_init()
965 *iter = &cpb; in acpi_cpufreq_init()
/linux-4.1.27/arch/m68k/emu/
Dnfblock.c66 struct bvec_iter iter; in nfhd_make_request() local
72 bio_for_each_segment(bvec, bio, iter) { in nfhd_make_request()
/linux-4.1.27/arch/xtensa/platforms/iss/
Dsimdisk.c107 struct bvec_iter iter; in simdisk_xfer_bio() local
110 bio_for_each_segment(bvec, bio, iter) { in simdisk_xfer_bio()
111 char *buffer = __bio_kmap_atomic(bio, iter); in simdisk_xfer_bio()

12