/linux-4.1.27/block/ |
D | t10-pi.c | 49 static int t10_pi_generate(struct blk_integrity_iter *iter, csum_fn *fn, in t10_pi_generate() argument 54 for (i = 0 ; i < iter->data_size ; i += iter->interval) { in t10_pi_generate() 55 struct t10_pi_tuple *pi = iter->prot_buf; in t10_pi_generate() 57 pi->guard_tag = fn(iter->data_buf, iter->interval); in t10_pi_generate() 61 pi->ref_tag = cpu_to_be32(lower_32_bits(iter->seed)); in t10_pi_generate() 65 iter->data_buf += iter->interval; in t10_pi_generate() 66 iter->prot_buf += sizeof(struct t10_pi_tuple); in t10_pi_generate() 67 iter->seed++; in t10_pi_generate() 73 static int t10_pi_verify(struct blk_integrity_iter *iter, csum_fn *fn, in t10_pi_verify() argument 78 for (i = 0 ; i < iter->data_size ; i += iter->interval) { in t10_pi_verify() [all …]
|
D | blk-map.c | 65 const struct iov_iter *iter, gfp_t gfp_mask) in blk_rq_map_user_iov() argument 72 if (!iter || !iter->count) in blk_rq_map_user_iov() 75 iov_for_each(iov, i, *iter) { in blk_rq_map_user_iov() 88 if (unaligned || (q->dma_pad_mask & iter->count) || map_data) in blk_rq_map_user_iov() 89 bio = bio_copy_user_iov(q, map_data, iter, gfp_mask); in blk_rq_map_user_iov() 91 bio = bio_map_user_iov(q, iter, gfp_mask); in blk_rq_map_user_iov() 99 if (bio->bi_iter.bi_size != iter->count) { in blk_rq_map_user_iov()
|
D | bio.c | 506 struct bvec_iter iter; in zero_fill_bio() local 508 bio_for_each_segment(bv, bio, iter) { in zero_fill_bio() 618 struct bvec_iter iter; in bio_clone_bioset() local 661 bio_for_each_segment(bv, bio_src, iter) in bio_clone_bioset() 1025 struct iov_iter iter; member 1047 static int bio_copy_from_iter(struct bio *bio, struct iov_iter iter) in bio_copy_from_iter() argument 1058 &iter); in bio_copy_from_iter() 1060 if (!iov_iter_count(&iter)) in bio_copy_from_iter() 1078 static int bio_copy_to_iter(struct bio *bio, struct iov_iter iter) in bio_copy_to_iter() argument 1089 &iter); in bio_copy_to_iter() [all …]
|
D | bio-integrity.c | 218 struct blk_integrity_iter iter; in bio_integrity_process() local 226 iter.disk_name = bio->bi_bdev->bd_disk->disk_name; in bio_integrity_process() 227 iter.interval = bi->interval; in bio_integrity_process() 228 iter.seed = bip_get_seed(bip); in bio_integrity_process() 229 iter.prot_buf = prot_buf; in bio_integrity_process() 234 iter.data_buf = kaddr + bv.bv_offset; in bio_integrity_process() 235 iter.data_size = bv.bv_len; in bio_integrity_process() 237 ret = proc_fn(&iter); in bio_integrity_process()
|
D | genhd.c | 740 struct class_dev_iter iter; in printk_all_partitions() local 743 class_dev_iter_init(&iter, &block_class, NULL, &disk_type); in printk_all_partitions() 744 while ((dev = class_dev_iter_next(&iter))) { in printk_all_partitions() 785 class_dev_iter_exit(&iter); in printk_all_partitions() 793 struct class_dev_iter *iter; in disk_seqf_start() local 796 iter = kmalloc(sizeof(*iter), GFP_KERNEL); in disk_seqf_start() 797 if (!iter) in disk_seqf_start() 800 seqf->private = iter; in disk_seqf_start() 801 class_dev_iter_init(iter, &block_class, NULL, &disk_type); in disk_seqf_start() 803 dev = class_dev_iter_next(iter); in disk_seqf_start() [all …]
|
D | bounce.c | 105 struct bvec_iter iter; in copy_to_high_bio_irq() local 107 bio_for_each_segment(tovec, to, iter) { in copy_to_high_bio_irq() 205 struct bvec_iter iter; in __blk_queue_bounce() local 210 bio_for_each_segment(from, *bio_orig, iter) in __blk_queue_bounce()
|
D | blk-merge.c | 20 struct bvec_iter iter; in __blk_recalc_rq_segments() local 41 bio_for_each_segment(bv, bio, iter) { in __blk_recalc_rq_segments() 127 struct bvec_iter iter; in blk_phys_contig_segment() local 139 bio_for_each_segment(end_bv, bio, iter) in blk_phys_contig_segment() 140 if (end_bv.bv_len == iter.bi_size) in blk_phys_contig_segment() 206 struct bvec_iter iter; in __blk_bios_map_sg() local 236 bio_for_each_segment(bvec, bio, iter) in __blk_bios_map_sg()
|
D | blk-integrity.c | 49 struct bvec_iter iter; in blk_rq_count_integrity_sg() local 52 bio_for_each_integrity_vec(iv, bio, iter) { in blk_rq_count_integrity_sg() 95 struct bvec_iter iter; in blk_rq_map_integrity_sg() local 98 bio_for_each_integrity_vec(iv, bio, iter) { in blk_rq_map_integrity_sg()
|
/linux-4.1.27/net/netlabel/ |
D | netlabel_addrlist.c | 63 struct netlbl_af4list *iter; in netlbl_af4list_search() local 65 list_for_each_entry_rcu(iter, head, list) in netlbl_af4list_search() 66 if (iter->valid && (addr & iter->mask) == iter->addr) in netlbl_af4list_search() 67 return iter; in netlbl_af4list_search() 88 struct netlbl_af4list *iter; in netlbl_af4list_search_exact() local 90 list_for_each_entry_rcu(iter, head, list) in netlbl_af4list_search_exact() 91 if (iter->valid && iter->addr == addr && iter->mask == mask) in netlbl_af4list_search_exact() 92 return iter; in netlbl_af4list_search_exact() 113 struct netlbl_af6list *iter; in netlbl_af6list_search() local 115 list_for_each_entry_rcu(iter, head, list) in netlbl_af6list_search() [all …]
|
D | netlabel_addrlist.h | 96 #define netlbl_af4list_foreach(iter, head) \ argument 97 for (iter = __af4list_valid((head)->next, head); \ 98 &iter->list != (head); \ 99 iter = __af4list_valid(iter->list.next, head)) 101 #define netlbl_af4list_foreach_rcu(iter, head) \ argument 102 for (iter = __af4list_valid_rcu((head)->next, head); \ 103 &iter->list != (head); \ 104 iter = __af4list_valid_rcu(iter->list.next, head)) 106 #define netlbl_af4list_foreach_safe(iter, tmp, head) \ argument 107 for (iter = __af4list_valid((head)->next, head), \ [all …]
|
D | netlabel_kapi.c | 431 struct netlbl_lsm_catmap *iter = *catmap; in _netlbl_catmap_getnode() local 434 if (iter == NULL) in _netlbl_catmap_getnode() 436 if (offset < iter->startbit) in _netlbl_catmap_getnode() 438 while (iter && offset >= (iter->startbit + NETLBL_CATMAP_SIZE)) { in _netlbl_catmap_getnode() 439 prev = iter; in _netlbl_catmap_getnode() 440 iter = iter->next; in _netlbl_catmap_getnode() 442 if (iter == NULL || offset < iter->startbit) in _netlbl_catmap_getnode() 445 return iter; in _netlbl_catmap_getnode() 449 return iter; in _netlbl_catmap_getnode() 454 iter = netlbl_catmap_alloc(gfp_flags); in _netlbl_catmap_getnode() [all …]
|
D | netlabel_cipso_v4.c | 105 u32 iter = 0; in netlbl_cipsov4_add_common() local 116 if (iter >= CIPSO_V4_TAG_MAXCNT) in netlbl_cipsov4_add_common() 118 doi_def->tags[iter++] = nla_get_u8(nla); in netlbl_cipsov4_add_common() 120 while (iter < CIPSO_V4_TAG_MAXCNT) in netlbl_cipsov4_add_common() 121 doi_def->tags[iter++] = CIPSO_V4_TAG_INVALID; in netlbl_cipsov4_add_common() 150 u32 iter; in netlbl_cipsov4_add_std() local 220 for (iter = 0; iter < doi_def->map.std->lvl.local_size; iter++) in netlbl_cipsov4_add_std() 221 doi_def->map.std->lvl.local[iter] = CIPSO_V4_INV_LVL; in netlbl_cipsov4_add_std() 222 for (iter = 0; iter < doi_def->map.std->lvl.cipso_size; iter++) in netlbl_cipsov4_add_std() 223 doi_def->map.std->lvl.cipso[iter] = CIPSO_V4_INV_LVL; in netlbl_cipsov4_add_std() [all …]
|
D | netlabel_domainhash.c | 117 u32 iter; in netlbl_domhsh_hash() local 124 for (iter = 0, val = 0, len = strlen(key); iter < len; iter++) in netlbl_domhsh_hash() 125 val = (val << 4 | (val >> (8 * sizeof(u32) - 4))) ^ key[iter]; in netlbl_domhsh_hash() 144 struct netlbl_dom_map *iter; in netlbl_domhsh_search() local 149 list_for_each_entry_rcu(iter, bkt_list, list) in netlbl_domhsh_search() 150 if (iter->valid && strcmp(iter->domain, domain) == 0) in netlbl_domhsh_search() 151 return iter; in netlbl_domhsh_search() 327 u32 iter; in netlbl_domhsh_init() local 344 for (iter = 0; iter < hsh_tbl->size; iter++) in netlbl_domhsh_init() 345 INIT_LIST_HEAD(&hsh_tbl->tbl[iter]); in netlbl_domhsh_init()
|
D | netlabel_domainhash.h | 56 #define netlbl_domhsh_addr4_entry(iter) \ argument 57 container_of(iter, struct netlbl_domaddr4_map, list) 63 #define netlbl_domhsh_addr6_entry(iter) \ argument 64 container_of(iter, struct netlbl_domaddr6_map, list)
|
D | netlabel_unlabeled.c | 80 #define netlbl_unlhsh_addr4_entry(iter) \ argument 81 container_of(iter, struct netlbl_unlhsh_addr4, list) 88 #define netlbl_unlhsh_addr6_entry(iter) \ argument 89 container_of(iter, struct netlbl_unlhsh_addr6, list) 226 struct netlbl_unlhsh_iface *iter; in netlbl_unlhsh_search_iface() local 230 list_for_each_entry_rcu(iter, bkt_list, list) in netlbl_unlhsh_search_iface() 231 if (iter->valid && iter->ifindex == ifindex) in netlbl_unlhsh_search_iface() 232 return iter; in netlbl_unlhsh_search_iface() 1420 u32 iter; in netlbl_unlabel_init() local 1437 for (iter = 0; iter < hsh_tbl->size; iter++) in netlbl_unlabel_init() [all …]
|
/linux-4.1.27/kernel/gcov/ |
D | gcc_3_4.c | 346 static struct gcov_fn_info *get_func(struct gcov_iterator *iter) in get_func() argument 348 return get_fn_info(iter->info, iter->function); in get_func() 351 static struct type_info *get_type(struct gcov_iterator *iter) in get_type() argument 353 return &iter->type_info[iter->type]; in get_type() 364 struct gcov_iterator *iter; in gcov_iter_new() local 366 iter = kzalloc(sizeof(struct gcov_iterator) + in gcov_iter_new() 369 if (iter) in gcov_iter_new() 370 iter->info = info; in gcov_iter_new() 372 return iter; in gcov_iter_new() 379 void gcov_iter_free(struct gcov_iterator *iter) in gcov_iter_free() argument [all …]
|
D | gcc_4_7.c | 476 struct gcov_iterator *iter; in gcov_iter_new() local 478 iter = kzalloc(sizeof(struct gcov_iterator), GFP_KERNEL); in gcov_iter_new() 479 if (!iter) in gcov_iter_new() 482 iter->info = info; in gcov_iter_new() 484 iter->size = convert_to_gcda(NULL, info); in gcov_iter_new() 485 iter->buffer = vmalloc(iter->size); in gcov_iter_new() 486 if (!iter->buffer) in gcov_iter_new() 489 convert_to_gcda(iter->buffer, info); in gcov_iter_new() 491 return iter; in gcov_iter_new() 494 kfree(iter); in gcov_iter_new() [all …]
|
D | gcov.h | 63 void gcov_iter_free(struct gcov_iterator *iter); 64 void gcov_iter_start(struct gcov_iterator *iter); 65 int gcov_iter_next(struct gcov_iterator *iter); 66 int gcov_iter_write(struct gcov_iterator *iter, struct seq_file *seq); 67 struct gcov_info *gcov_iter_get_info(struct gcov_iterator *iter);
|
D | fs.c | 109 struct gcov_iterator *iter = data; in gcov_seq_next() local 111 if (gcov_iter_next(iter)) in gcov_seq_next() 115 return iter; in gcov_seq_next() 121 struct gcov_iterator *iter = data; in gcov_seq_show() local 123 if (gcov_iter_write(iter, seq)) in gcov_seq_show() 181 struct gcov_iterator *iter; in gcov_seq_open() local 195 iter = gcov_iter_new(info); in gcov_seq_open() 196 if (!iter) in gcov_seq_open() 202 seq->private = iter; in gcov_seq_open() 208 gcov_iter_free(iter); in gcov_seq_open() [all …]
|
/linux-4.1.27/drivers/gpu/drm/ |
D | drm_vma_manager.c | 162 struct rb_node *iter; in drm_vma_offset_lookup_locked() local 165 iter = mgr->vm_addr_space_rb.rb_node; in drm_vma_offset_lookup_locked() 168 while (likely(iter)) { in drm_vma_offset_lookup_locked() 169 node = rb_entry(iter, struct drm_vma_offset_node, vm_rb); in drm_vma_offset_lookup_locked() 172 iter = iter->rb_right; in drm_vma_offset_lookup_locked() 177 iter = iter->rb_left; in drm_vma_offset_lookup_locked() 196 struct rb_node **iter = &mgr->vm_addr_space_rb.rb_node; in _drm_vma_offset_add_rb() local 200 while (likely(*iter)) { in _drm_vma_offset_add_rb() 201 parent = *iter; in _drm_vma_offset_add_rb() 202 iter_node = rb_entry(*iter, struct drm_vma_offset_node, vm_rb); in _drm_vma_offset_add_rb() [all …]
|
/linux-4.1.27/include/linux/ |
D | radix-tree.h | 341 radix_tree_iter_init(struct radix_tree_iter *iter, unsigned long start) in radix_tree_iter_init() argument 351 iter->index = 0; in radix_tree_iter_init() 352 iter->next_index = start; in radix_tree_iter_init() 370 struct radix_tree_iter *iter, unsigned flags); 382 void **radix_tree_iter_retry(struct radix_tree_iter *iter) in radix_tree_iter_retry() argument 384 iter->next_index = iter->index; in radix_tree_iter_retry() 395 radix_tree_chunk_size(struct radix_tree_iter *iter) in radix_tree_chunk_size() argument 397 return iter->next_index - iter->index; in radix_tree_chunk_size() 412 radix_tree_next_slot(void **slot, struct radix_tree_iter *iter, unsigned flags) in radix_tree_next_slot() argument 415 iter->tags >>= 1; in radix_tree_next_slot() [all …]
|
D | bio.h | 64 #define __bvec_iter_bvec(bvec, iter) (&(bvec)[(iter).bi_idx]) argument 66 #define bvec_iter_page(bvec, iter) \ argument 67 (__bvec_iter_bvec((bvec), (iter))->bv_page) 69 #define bvec_iter_len(bvec, iter) \ argument 70 min((iter).bi_size, \ 71 __bvec_iter_bvec((bvec), (iter))->bv_len - (iter).bi_bvec_done) 73 #define bvec_iter_offset(bvec, iter) \ argument 74 (__bvec_iter_bvec((bvec), (iter))->bv_offset + (iter).bi_bvec_done) 76 #define bvec_iter_bvec(bvec, iter) \ argument 78 .bv_page = bvec_iter_page((bvec), (iter)), \ [all …]
|
D | uio.h | 57 static inline struct iovec iov_iter_iovec(const struct iov_iter *iter) in iov_iter_iovec() argument 60 .iov_base = iter->iov->iov_base + iter->iov_offset, in iov_iter_iovec() 61 .iov_len = min(iter->count, in iov_iter_iovec() 62 iter->iov->iov_len - iter->iov_offset), in iov_iter_iovec() 66 #define iov_for_each(iov, iter, start) \ argument 68 for (iter = (start); \ 69 (iter).count && \ 70 ((iov = iov_iter_iovec(&(iter))), 1); \ 71 iov_iter_advance(&(iter), (iov).iov_len))
|
D | ring_buffer.h | 130 void ring_buffer_read_start(struct ring_buffer_iter *iter); 131 void ring_buffer_read_finish(struct ring_buffer_iter *iter); 134 ring_buffer_iter_peek(struct ring_buffer_iter *iter, u64 *ts); 136 ring_buffer_read(struct ring_buffer_iter *iter, u64 *ts); 137 void ring_buffer_iter_reset(struct ring_buffer_iter *iter); 138 int ring_buffer_iter_empty(struct ring_buffer_iter *iter);
|
D | ftrace.h | 417 struct ftrace_rec_iter *ftrace_rec_iter_next(struct ftrace_rec_iter *iter); 418 struct dyn_ftrace *ftrace_rec_iter_record(struct ftrace_rec_iter *iter); 420 #define for_ftrace_rec_iter(iter) \ argument 421 for (iter = ftrace_rec_iter_start(); \ 422 iter; \ 423 iter = ftrace_rec_iter_next(iter))
|
D | netdevice.h | 3484 struct list_head **iter); 3486 struct list_head **iter); 3489 #define netdev_for_each_upper_dev_rcu(dev, updev, iter) \ argument 3490 for (iter = &(dev)->adj_list.upper, \ 3491 updev = netdev_upper_get_next_dev_rcu(dev, &(iter)); \ 3493 updev = netdev_upper_get_next_dev_rcu(dev, &(iter))) 3496 #define netdev_for_each_all_upper_dev_rcu(dev, updev, iter) \ argument 3497 for (iter = &(dev)->all_adj_list.upper, \ 3498 updev = netdev_all_upper_get_next_dev_rcu(dev, &(iter)); \ 3500 updev = netdev_all_upper_get_next_dev_rcu(dev, &(iter))) [all …]
|
D | pnfs_osd_xdr.h | 294 struct pnfs_osd_xdr_decode_layout_iter *iter, struct xdr_stream *xdr); 297 struct pnfs_osd_xdr_decode_layout_iter *iter, struct xdr_stream *xdr,
|
D | ftrace_event.h | 55 int ftrace_raw_output_prep(struct trace_iterator *iter, 118 typedef enum print_line_t (*trace_print_func)(struct trace_iterator *iter, 186 int ftrace_output_call(struct trace_iterator *iter, char *name, char *fmt, ...); 222 int ftrace_output_event(struct trace_iterator *iter, struct ftrace_event_call *event,
|
D | rhashtable.h | 347 int rhashtable_walk_init(struct rhashtable *ht, struct rhashtable_iter *iter); 348 void rhashtable_walk_exit(struct rhashtable_iter *iter); 349 int rhashtable_walk_start(struct rhashtable_iter *iter) __acquires(RCU); 350 void *rhashtable_walk_next(struct rhashtable_iter *iter); 351 void rhashtable_walk_stop(struct rhashtable_iter *iter) __releases(RCU);
|
/linux-4.1.27/kernel/trace/ |
D | trace_kdb.c | 22 static struct trace_iterator iter; in ftrace_dump_buf() local 27 trace_init_global_iter(&iter); in ftrace_dump_buf() 28 iter.buffer_iter = buffer_iter; in ftrace_dump_buf() 31 atomic_inc(&per_cpu_ptr(iter.trace_buffer->data, cpu)->disabled); in ftrace_dump_buf() 42 memset(&iter.seq, 0, in ftrace_dump_buf() 45 iter.iter_flags |= TRACE_FILE_LAT_FMT; in ftrace_dump_buf() 46 iter.pos = -1; in ftrace_dump_buf() 50 iter.buffer_iter[cpu] = in ftrace_dump_buf() 51 ring_buffer_read_prepare(iter.trace_buffer->buffer, cpu); in ftrace_dump_buf() 52 ring_buffer_read_start(iter.buffer_iter[cpu]); in ftrace_dump_buf() [all …]
|
D | trace_output.c | 23 enum print_line_t trace_print_bputs_msg_only(struct trace_iterator *iter) in trace_print_bputs_msg_only() argument 25 struct trace_seq *s = &iter->seq; in trace_print_bputs_msg_only() 26 struct trace_entry *entry = iter->ent; in trace_print_bputs_msg_only() 36 enum print_line_t trace_print_bprintk_msg_only(struct trace_iterator *iter) in trace_print_bprintk_msg_only() argument 38 struct trace_seq *s = &iter->seq; in trace_print_bprintk_msg_only() 39 struct trace_entry *entry = iter->ent; in trace_print_bprintk_msg_only() 49 enum print_line_t trace_print_printk_msg_only(struct trace_iterator *iter) in trace_print_printk_msg_only() argument 51 struct trace_seq *s = &iter->seq; in trace_print_printk_msg_only() 52 struct trace_entry *entry = iter->ent; in trace_print_printk_msg_only() 225 int ftrace_raw_output_prep(struct trace_iterator *iter, in ftrace_raw_output_prep() argument [all …]
|
D | trace.c | 1130 static int wait_on_pipe(struct trace_iterator *iter, bool full) in wait_on_pipe() argument 1133 if (trace_buffer_iter(iter, iter->cpu_file)) in wait_on_pipe() 1136 return ring_buffer_wait(iter->trace_buffer->buffer, iter->cpu_file, in wait_on_pipe() 2278 static void trace_iterator_increment(struct trace_iterator *iter) in trace_iterator_increment() argument 2280 struct ring_buffer_iter *buf_iter = trace_buffer_iter(iter, iter->cpu); in trace_iterator_increment() 2282 iter->idx++; in trace_iterator_increment() 2288 peek_next_entry(struct trace_iterator *iter, int cpu, u64 *ts, in peek_next_entry() argument 2292 struct ring_buffer_iter *buf_iter = trace_buffer_iter(iter, cpu); in peek_next_entry() 2297 event = ring_buffer_peek(iter->trace_buffer->buffer, cpu, ts, in peek_next_entry() 2301 iter->ent_size = ring_buffer_event_length(event); in peek_next_entry() [all …]
|
D | trace_functions_graph.c | 571 get_return_for_leaf(struct trace_iterator *iter, in get_return_for_leaf() argument 574 struct fgraph_data *data = iter->private; in get_return_for_leaf() 588 ring_iter = trace_buffer_iter(iter, iter->cpu); in get_return_for_leaf() 598 ring_buffer_consume(iter->trace_buffer->buffer, iter->cpu, in get_return_for_leaf() 600 event = ring_buffer_peek(iter->trace_buffer->buffer, iter->cpu, in get_return_for_leaf() 653 print_graph_irq(struct trace_iterator *iter, unsigned long addr, in print_graph_irq() argument 656 struct trace_seq *s = &iter->seq; in print_graph_irq() 657 struct trace_entry *ent = iter->ent; in print_graph_irq() 666 print_graph_abs_time(iter->ts, s); in print_graph_irq() 761 print_graph_entry_leaf(struct trace_iterator *iter, in print_graph_entry_leaf() argument [all …]
|
D | trace_mmiotrace.c | 102 static void mmio_pipe_open(struct trace_iterator *iter) in mmio_pipe_open() argument 105 struct trace_seq *s = &iter->seq; in mmio_pipe_open() 114 iter->private = hiter; in mmio_pipe_open() 118 static void mmio_close(struct trace_iterator *iter) in mmio_close() argument 120 struct header_iter *hiter = iter->private; in mmio_close() 122 iter->private = NULL; in mmio_close() 125 static unsigned long count_overruns(struct trace_iterator *iter) in count_overruns() argument 128 unsigned long over = ring_buffer_overruns(iter->trace_buffer->buffer); in count_overruns() 136 static ssize_t mmio_read(struct trace_iterator *iter, struct file *filp, in mmio_read() argument 140 struct header_iter *hiter = iter->private; in mmio_read() [all …]
|
D | trace_output.h | 8 trace_print_bputs_msg_only(struct trace_iterator *iter); 10 trace_print_bprintk_msg_only(struct trace_iterator *iter); 12 trace_print_printk_msg_only(struct trace_iterator *iter); 22 extern int trace_print_context(struct trace_iterator *iter); 23 extern int trace_print_lat_context(struct trace_iterator *iter); 29 extern enum print_line_t trace_nop_print(struct trace_iterator *iter,
|
D | ftrace.c | 2357 struct ftrace_rec_iter *iter = &ftrace_rec_iter; in ftrace_rec_iter_start() local 2359 iter->pg = ftrace_pages_start; in ftrace_rec_iter_start() 2360 iter->index = 0; in ftrace_rec_iter_start() 2363 while (iter->pg && !iter->pg->index) in ftrace_rec_iter_start() 2364 iter->pg = iter->pg->next; in ftrace_rec_iter_start() 2366 if (!iter->pg) in ftrace_rec_iter_start() 2369 return iter; in ftrace_rec_iter_start() 2378 struct ftrace_rec_iter *ftrace_rec_iter_next(struct ftrace_rec_iter *iter) in ftrace_rec_iter_next() argument 2380 iter->index++; in ftrace_rec_iter_next() 2382 if (iter->index >= iter->pg->index) { in ftrace_rec_iter_next() [all …]
|
D | trace_printk.c | 49 const char **iter; in hold_module_trace_bprintk_format() local 57 for (iter = start; iter < end; iter++) { in hold_module_trace_bprintk_format() 58 struct trace_bprintk_fmt *tb_fmt = lookup_format(*iter); in hold_module_trace_bprintk_format() 60 *iter = tb_fmt->fmt; in hold_module_trace_bprintk_format() 67 fmt = kmalloc(strlen(*iter) + 1, GFP_KERNEL); in hold_module_trace_bprintk_format() 70 strcpy(fmt, *iter); in hold_module_trace_bprintk_format() 75 *iter = fmt; in hold_module_trace_bprintk_format()
|
D | ring_buffer.c | 1884 rb_iter_head_event(struct ring_buffer_iter *iter) in rb_iter_head_event() argument 1886 return __rb_page_index(iter->head_page, iter->head); in rb_iter_head_event() 1987 static void rb_inc_iter(struct ring_buffer_iter *iter) in rb_inc_iter() argument 1989 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in rb_inc_iter() 1997 if (iter->head_page == cpu_buffer->reader_page) in rb_inc_iter() 1998 iter->head_page = rb_set_head_page(cpu_buffer); in rb_inc_iter() 2000 rb_inc_page(cpu_buffer, &iter->head_page); in rb_inc_iter() 2002 iter->read_stamp = iter->head_page->page->time_stamp; in rb_inc_iter() 2003 iter->head = 0; in rb_inc_iter() 3402 static void rb_iter_reset(struct ring_buffer_iter *iter) in rb_iter_reset() argument [all …]
|
D | trace.h | 363 void (*open)(struct trace_iterator *iter); 364 void (*pipe_open)(struct trace_iterator *iter); 365 void (*close)(struct trace_iterator *iter); 366 void (*pipe_close)(struct trace_iterator *iter); 367 ssize_t (*read)(struct trace_iterator *iter, 370 ssize_t (*splice_read)(struct trace_iterator *iter, 381 enum print_line_t (*print_line)(struct trace_iterator *iter); 525 trace_buffer_iter(struct trace_iterator *iter, int cpu) in trace_buffer_iter() argument 527 if (iter->buffer_iter && iter->buffer_iter[cpu]) in trace_buffer_iter() 528 return iter->buffer_iter[cpu]; in trace_buffer_iter() [all …]
|
D | blktrace.c | 1145 typedef void (blk_log_action_t) (struct trace_iterator *iter, const char *act); 1147 static void blk_log_action_classic(struct trace_iterator *iter, const char *act) in blk_log_action_classic() argument 1150 unsigned long long ts = iter->ts; in blk_log_action_classic() 1153 const struct blk_io_trace *t = te_blk_io_trace(iter->ent); in blk_log_action_classic() 1157 trace_seq_printf(&iter->seq, in blk_log_action_classic() 1159 MAJOR(t->device), MINOR(t->device), iter->cpu, in blk_log_action_classic() 1160 secs, nsec_rem, iter->ent->pid, act, rwbs); in blk_log_action_classic() 1163 static void blk_log_action(struct trace_iterator *iter, const char *act) in blk_log_action() argument 1166 const struct blk_io_trace *t = te_blk_io_trace(iter->ent); in blk_log_action() 1169 trace_seq_printf(&iter->seq, "%3d,%-3d %2s %3s ", in blk_log_action() [all …]
|
D | trace_irqsoff.c | 210 static void irqsoff_trace_open(struct trace_iterator *iter) in irqsoff_trace_open() argument 213 graph_trace_open(iter); in irqsoff_trace_open() 217 static void irqsoff_trace_close(struct trace_iterator *iter) in irqsoff_trace_close() argument 219 if (iter->private) in irqsoff_trace_close() 220 graph_trace_close(iter); in irqsoff_trace_close() 228 static enum print_line_t irqsoff_print_line(struct trace_iterator *iter) in irqsoff_print_line() argument 235 return print_graph_function_flags(iter, GRAPH_TRACER_FLAGS); in irqsoff_print_line() 273 static enum print_line_t irqsoff_print_line(struct trace_iterator *iter) in irqsoff_print_line() argument 279 static void irqsoff_trace_open(struct trace_iterator *iter) { } in irqsoff_trace_open() argument 280 static void irqsoff_trace_close(struct trace_iterator *iter) { } in irqsoff_trace_close() argument
|
D | trace_sched_wakeup.c | 260 static void wakeup_trace_open(struct trace_iterator *iter) in wakeup_trace_open() argument 263 graph_trace_open(iter); in wakeup_trace_open() 266 static void wakeup_trace_close(struct trace_iterator *iter) in wakeup_trace_close() argument 268 if (iter->private) in wakeup_trace_close() 269 graph_trace_close(iter); in wakeup_trace_close() 276 static enum print_line_t wakeup_print_line(struct trace_iterator *iter) in wakeup_print_line() argument 283 return print_graph_function_flags(iter, GRAPH_TRACER_FLAGS); in wakeup_print_line() 320 static enum print_line_t wakeup_print_line(struct trace_iterator *iter) in wakeup_print_line() argument 326 static void wakeup_trace_open(struct trace_iterator *iter) { } in wakeup_trace_open() argument 327 static void wakeup_trace_close(struct trace_iterator *iter) { } in wakeup_trace_close() argument
|
D | trace_branch.c | 149 static enum print_line_t trace_branch_print(struct trace_iterator *iter, in trace_branch_print() argument 154 trace_assign_type(field, iter->ent); in trace_branch_print() 156 trace_seq_printf(&iter->seq, "[%s] %s:%s:%d\n", in trace_branch_print() 162 return trace_handle_return(&iter->seq); in trace_branch_print()
|
D | trace_syscalls.c | 110 print_syscall_enter(struct trace_iterator *iter, int flags, in print_syscall_enter() argument 113 struct trace_seq *s = &iter->seq; in print_syscall_enter() 114 struct trace_entry *ent = iter->ent; in print_syscall_enter() 156 print_syscall_exit(struct trace_iterator *iter, int flags, in print_syscall_exit() argument 159 struct trace_seq *s = &iter->seq; in print_syscall_exit() 160 struct trace_entry *ent = iter->ent; in print_syscall_exit()
|
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/ |
D | icm.h | 88 struct mlx4_icm_iter *iter) in mlx4_icm_first() argument 90 iter->icm = icm; in mlx4_icm_first() 91 iter->chunk = list_empty(&icm->chunk_list) ? in mlx4_icm_first() 94 iter->page_idx = 0; in mlx4_icm_first() 97 static inline int mlx4_icm_last(struct mlx4_icm_iter *iter) in mlx4_icm_last() argument 99 return !iter->chunk; in mlx4_icm_last() 102 static inline void mlx4_icm_next(struct mlx4_icm_iter *iter) in mlx4_icm_next() argument 104 if (++iter->page_idx >= iter->chunk->nsg) { in mlx4_icm_next() 105 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mlx4_icm_next() 106 iter->chunk = NULL; in mlx4_icm_next() [all …]
|
/linux-4.1.27/kernel/ |
D | kallsyms.c | 458 static int get_ksymbol_mod(struct kallsym_iter *iter) in get_ksymbol_mod() argument 460 if (module_get_kallsym(iter->pos - kallsyms_num_syms, &iter->value, in get_ksymbol_mod() 461 &iter->type, iter->name, iter->module_name, in get_ksymbol_mod() 462 &iter->exported) < 0) in get_ksymbol_mod() 468 static unsigned long get_ksymbol_core(struct kallsym_iter *iter) in get_ksymbol_core() argument 470 unsigned off = iter->nameoff; in get_ksymbol_core() 472 iter->module_name[0] = '\0'; in get_ksymbol_core() 473 iter->value = kallsyms_addresses[iter->pos]; in get_ksymbol_core() 475 iter->type = kallsyms_get_symbol_type(off); in get_ksymbol_core() 477 off = kallsyms_expand_symbol(off, iter->name, ARRAY_SIZE(iter->name)); in get_ksymbol_core() [all …]
|
D | jump_label.c | 140 struct jump_entry *iter; in __jump_label_text_reserved() local 142 iter = iter_start; in __jump_label_text_reserved() 143 while (iter < iter_stop) { in __jump_label_text_reserved() 144 if (addr_conflict(iter, start, end)) in __jump_label_text_reserved() 146 iter++; in __jump_label_text_reserved() 197 struct jump_entry *iter; in jump_label_init() local 202 for (iter = iter_start; iter < iter_stop; iter++) { in jump_label_init() 205 iterk = (struct static_key *)(unsigned long)iter->key; in jump_label_init() 206 arch_jump_label_transform_static(iter, jump_label_type(iterk)); in jump_label_init() 214 *((unsigned long *)&key->entries) += (unsigned long)iter; in jump_label_init() [all …]
|
D | tracepoint.c | 353 struct tracepoint * const *iter; in tp_module_going_check_quiescent() local 357 for (iter = begin; iter < end; iter++) in tp_module_going_check_quiescent() 358 WARN_ON_ONCE((*iter)->funcs); in tp_module_going_check_quiescent() 467 struct tracepoint * const *iter; in for_each_tracepoint_range() local 471 for (iter = begin; iter < end; iter++) in for_each_tracepoint_range() 472 fct(*iter, priv); in for_each_tracepoint_range()
|
/linux-4.1.27/net/rds/ |
D | info.c | 104 void rds_info_iter_unmap(struct rds_info_iterator *iter) in rds_info_iter_unmap() argument 106 if (iter->addr) { in rds_info_iter_unmap() 107 kunmap_atomic(iter->addr); in rds_info_iter_unmap() 108 iter->addr = NULL; in rds_info_iter_unmap() 115 void rds_info_copy(struct rds_info_iterator *iter, void *data, in rds_info_copy() argument 121 if (!iter->addr) in rds_info_copy() 122 iter->addr = kmap_atomic(*iter->pages); in rds_info_copy() 124 this = min(bytes, PAGE_SIZE - iter->offset); in rds_info_copy() 127 "bytes %lu\n", *iter->pages, iter->addr, in rds_info_copy() 128 iter->offset, this, data, bytes); in rds_info_copy() [all …]
|
D | stats.c | 81 void rds_stats_info_copy(struct rds_info_iterator *iter, in rds_stats_info_copy() argument 93 rds_info_copy(iter, &ctr, sizeof(ctr)); in rds_stats_info_copy() 109 struct rds_info_iterator *iter, in rds_stats_info() argument 133 rds_stats_info_copy(iter, (uint64_t *)&stats, rds_stat_names, in rds_stats_info() 139 lens->nr = rds_trans_stats_info_copy(iter, avail) + in rds_stats_info()
|
D | info.h | 18 struct rds_info_iterator *iter, 25 void rds_info_copy(struct rds_info_iterator *iter, void *data, 27 void rds_info_iter_unmap(struct rds_info_iterator *iter);
|
D | connection.c | 393 struct rds_info_iterator *iter, in rds_conn_message_info() argument 423 rds_inc_info_copy(&rm->m_inc, iter, in rds_conn_message_info() 438 struct rds_info_iterator *iter, in rds_conn_message_info_send() argument 441 rds_conn_message_info(sock, len, iter, lens, 1); in rds_conn_message_info_send() 446 struct rds_info_iterator *iter, in rds_conn_message_info_retrans() argument 449 rds_conn_message_info(sock, len, iter, lens, 0); in rds_conn_message_info_retrans() 453 struct rds_info_iterator *iter, in rds_for_each_conn_info() argument 480 rds_info_copy(iter, buffer, item_len); in rds_for_each_conn_info() 516 struct rds_info_iterator *iter, in rds_conn_info() argument 519 rds_for_each_conn_info(sock, len, iter, lens, in rds_conn_info()
|
D | transport.c | 110 unsigned int rds_trans_stats_info_copy(struct rds_info_iterator *iter, in rds_trans_stats_info_copy() argument 119 rds_info_iter_unmap(iter); in rds_trans_stats_info_copy() 128 part = trans->stats_info_copy(iter, avail); in rds_trans_stats_info_copy()
|
D | tcp_stats.c | 51 unsigned int rds_tcp_stats_info_copy(struct rds_info_iterator *iter, in rds_tcp_stats_info_copy() argument 70 rds_stats_info_copy(iter, (uint64_t *)&stats, rds_tcp_stat_names, in rds_tcp_stats_info_copy()
|
D | iw_stats.c | 72 unsigned int rds_iw_stats_info_copy(struct rds_info_iterator *iter, in rds_iw_stats_info_copy() argument 91 rds_stats_info_copy(iter, (uint64_t *)&stats, rds_iw_stat_names, in rds_iw_stats_info_copy()
|
D | ib_stats.c | 74 unsigned int rds_ib_stats_info_copy(struct rds_info_iterator *iter, in rds_ib_stats_info_copy() argument 93 rds_stats_info_copy(iter, (uint64_t *)&stats, rds_ib_stat_names, in rds_ib_stats_info_copy()
|
D | af_rds.c | 467 struct rds_info_iterator *iter, in rds_sock_inc_info() argument 485 rds_inc_info_copy(inc, iter, inc->i_saddr, in rds_sock_inc_info() 499 struct rds_info_iterator *iter, in rds_sock_info() argument 521 rds_info_copy(iter, &sinfo, sizeof(sinfo)); in rds_sock_info()
|
D | rds.h | 444 unsigned int (*stats_info_copy)(struct rds_info_iterator *iter, 624 struct rds_info_iterator *iter, 711 struct rds_info_iterator *iter, 770 void rds_stats_info_copy(struct rds_info_iterator *iter, 804 unsigned int rds_trans_stats_info_copy(struct rds_info_iterator *iter,
|
/linux-4.1.27/drivers/infiniband/ulp/ipoib/ |
D | ipoib_fs.c | 60 struct ipoib_mcast_iter *iter; in ipoib_mcg_seq_start() local 63 iter = ipoib_mcast_iter_init(file->private); in ipoib_mcg_seq_start() 64 if (!iter) in ipoib_mcg_seq_start() 68 if (ipoib_mcast_iter_next(iter)) { in ipoib_mcg_seq_start() 69 kfree(iter); in ipoib_mcg_seq_start() 74 return iter; in ipoib_mcg_seq_start() 80 struct ipoib_mcast_iter *iter = iter_ptr; in ipoib_mcg_seq_next() local 84 if (ipoib_mcast_iter_next(iter)) { in ipoib_mcg_seq_next() 85 kfree(iter); in ipoib_mcg_seq_next() 89 return iter; in ipoib_mcg_seq_next() [all …]
|
D | ipoib_multicast.c | 935 struct ipoib_mcast_iter *iter; in ipoib_mcast_iter_init() local 937 iter = kmalloc(sizeof *iter, GFP_KERNEL); in ipoib_mcast_iter_init() 938 if (!iter) in ipoib_mcast_iter_init() 941 iter->dev = dev; in ipoib_mcast_iter_init() 942 memset(iter->mgid.raw, 0, 16); in ipoib_mcast_iter_init() 944 if (ipoib_mcast_iter_next(iter)) { in ipoib_mcast_iter_init() 945 kfree(iter); in ipoib_mcast_iter_init() 949 return iter; in ipoib_mcast_iter_init() 952 int ipoib_mcast_iter_next(struct ipoib_mcast_iter *iter) in ipoib_mcast_iter_next() argument 954 struct ipoib_dev_priv *priv = netdev_priv(iter->dev); in ipoib_mcast_iter_next() [all …]
|
D | ipoib_main.c | 332 struct ipoib_path_iter *iter; in ipoib_path_iter_init() local 334 iter = kmalloc(sizeof *iter, GFP_KERNEL); in ipoib_path_iter_init() 335 if (!iter) in ipoib_path_iter_init() 338 iter->dev = dev; in ipoib_path_iter_init() 339 memset(iter->path.pathrec.dgid.raw, 0, 16); in ipoib_path_iter_init() 341 if (ipoib_path_iter_next(iter)) { in ipoib_path_iter_init() 342 kfree(iter); in ipoib_path_iter_init() 346 return iter; in ipoib_path_iter_init() 349 int ipoib_path_iter_next(struct ipoib_path_iter *iter) in ipoib_path_iter_next() argument 351 struct ipoib_dev_priv *priv = netdev_priv(iter->dev); in ipoib_path_iter_next() [all …]
|
/linux-4.1.27/drivers/infiniband/hw/mthca/ |
D | mthca_memfree.h | 100 struct mthca_icm_iter *iter) in mthca_icm_first() argument 102 iter->icm = icm; in mthca_icm_first() 103 iter->chunk = list_empty(&icm->chunk_list) ? in mthca_icm_first() 106 iter->page_idx = 0; in mthca_icm_first() 109 static inline int mthca_icm_last(struct mthca_icm_iter *iter) in mthca_icm_last() argument 111 return !iter->chunk; in mthca_icm_last() 114 static inline void mthca_icm_next(struct mthca_icm_iter *iter) in mthca_icm_next() argument 116 if (++iter->page_idx >= iter->chunk->nsg) { in mthca_icm_next() 117 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mthca_icm_next() 118 iter->chunk = NULL; in mthca_icm_next() [all …]
|
/linux-4.1.27/drivers/s390/cio/ |
D | blacklist.c | 287 struct ccwdev_iter *iter = s->private; in cio_ignore_proc_seq_start() local 291 memset(iter, 0, sizeof(*iter)); in cio_ignore_proc_seq_start() 292 iter->ssid = *offset / (__MAX_SUBCHANNEL + 1); in cio_ignore_proc_seq_start() 293 iter->devno = *offset % (__MAX_SUBCHANNEL + 1); in cio_ignore_proc_seq_start() 294 return iter; in cio_ignore_proc_seq_start() 305 struct ccwdev_iter *iter; in cio_ignore_proc_seq_next() local 309 iter = it; in cio_ignore_proc_seq_next() 310 if (iter->devno == __MAX_SUBCHANNEL) { in cio_ignore_proc_seq_next() 311 iter->devno = 0; in cio_ignore_proc_seq_next() 312 iter->ssid++; in cio_ignore_proc_seq_next() [all …]
|
/linux-4.1.27/tools/perf/util/ |
D | hist.c | 478 iter_next_nop_entry(struct hist_entry_iter *iter __maybe_unused, in iter_next_nop_entry() 485 iter_add_next_nop_entry(struct hist_entry_iter *iter __maybe_unused, in iter_add_next_nop_entry() 492 iter_prepare_mem_entry(struct hist_entry_iter *iter, struct addr_location *al) in iter_prepare_mem_entry() argument 494 struct perf_sample *sample = iter->sample; in iter_prepare_mem_entry() 501 iter->priv = mi; in iter_prepare_mem_entry() 506 iter_add_single_mem_entry(struct hist_entry_iter *iter, struct addr_location *al) in iter_add_single_mem_entry() argument 509 struct mem_info *mi = iter->priv; in iter_add_single_mem_entry() 510 struct hists *hists = evsel__hists(iter->evsel); in iter_add_single_mem_entry() 516 cost = iter->sample->weight; in iter_add_single_mem_entry() 527 he = __hists__add_entry(hists, al, iter->parent, NULL, mi, in iter_add_single_mem_entry() [all …]
|
D | comm.c | 50 struct comm_str *iter, *new; in comm_str__findnew() local 55 iter = rb_entry(parent, struct comm_str, rb_node); in comm_str__findnew() 57 cmp = strcmp(str, iter->str); in comm_str__findnew() 59 return iter; in comm_str__findnew()
|
D | ordered-events.c | 188 struct ordered_event *tmp, *iter; in __ordered_events__flush() local 201 list_for_each_entry_safe(iter, tmp, head, list) { in __ordered_events__flush() 205 if (iter->timestamp > limit) in __ordered_events__flush() 207 ret = oe->deliver(oe, iter); in __ordered_events__flush() 211 ordered_events__delete(oe, iter); in __ordered_events__flush() 212 oe->last_flush = iter->timestamp; in __ordered_events__flush()
|
/linux-4.1.27/drivers/dma/ppc4xx/ |
D | adma.c | 186 struct ppc440spe_adma_desc_slot *iter) in print_cb_list() argument 188 for (; iter; iter = iter->hw_next) in print_cb_list() 189 print_cb(chan, iter->hw_desc); in print_cb_list() 334 struct ppc440spe_adma_desc_slot *iter; in ppc440spe_desc_init_dma01pq() local 348 list_for_each_entry(iter, &desc->group_list, chain_node) { in ppc440spe_desc_init_dma01pq() 349 hw_desc = iter->hw_desc; in ppc440spe_desc_init_dma01pq() 350 memset(iter->hw_desc, 0, sizeof(struct dma_cdb)); in ppc440spe_desc_init_dma01pq() 352 if (likely(!list_is_last(&iter->chain_node, in ppc440spe_desc_init_dma01pq() 355 iter->hw_next = list_entry(iter->chain_node.next, in ppc440spe_desc_init_dma01pq() 357 clear_bit(PPC440SPE_DESC_INT, &iter->flags); in ppc440spe_desc_init_dma01pq() [all …]
|
/linux-4.1.27/lib/ |
D | rhashtable.c | 498 int rhashtable_walk_init(struct rhashtable *ht, struct rhashtable_iter *iter) in rhashtable_walk_init() argument 500 iter->ht = ht; in rhashtable_walk_init() 501 iter->p = NULL; in rhashtable_walk_init() 502 iter->slot = 0; in rhashtable_walk_init() 503 iter->skip = 0; in rhashtable_walk_init() 505 iter->walker = kmalloc(sizeof(*iter->walker), GFP_KERNEL); in rhashtable_walk_init() 506 if (!iter->walker) in rhashtable_walk_init() 510 iter->walker->tbl = in rhashtable_walk_init() 512 list_add(&iter->walker->list, &iter->walker->tbl->walkers); in rhashtable_walk_init() 525 void rhashtable_walk_exit(struct rhashtable_iter *iter) in rhashtable_walk_exit() argument [all …]
|
D | cordic.c | 59 unsigned iter; in cordic_calc_iq() local 80 for (iter = 0; iter < CORDIC_NUM_ITER; iter++) { in cordic_calc_iq() 82 valtmp = coord.i - (coord.q >> iter); in cordic_calc_iq() 83 coord.q += (coord.i >> iter); in cordic_calc_iq() 84 angle += arctan_table[iter]; in cordic_calc_iq() 86 valtmp = coord.i + (coord.q >> iter); in cordic_calc_iq() 87 coord.q -= (coord.i >> iter); in cordic_calc_iq() 88 angle -= arctan_table[iter]; in cordic_calc_iq()
|
D | plist.c | 76 struct plist_node *first, *iter, *prev = NULL; in plist_add() local 86 first = iter = plist_first(head); in plist_add() 89 if (node->prio < iter->prio) { in plist_add() 90 node_next = &iter->node_list; in plist_add() 94 prev = iter; in plist_add() 95 iter = list_entry(iter->prio_list.next, in plist_add() 97 } while (iter != first); in plist_add() 100 list_add_tail(&node->prio_list, &iter->prio_list); in plist_add() 148 struct plist_node *iter; in plist_requeue() local 158 iter = plist_next(node); in plist_requeue() [all …]
|
D | dynamic_debug.c | 684 static struct _ddebug *ddebug_iter_first(struct ddebug_iter *iter) in ddebug_iter_first() argument 687 iter->table = NULL; in ddebug_iter_first() 688 iter->idx = 0; in ddebug_iter_first() 691 iter->table = list_entry(ddebug_tables.next, in ddebug_iter_first() 693 iter->idx = 0; in ddebug_iter_first() 694 return &iter->table->ddebugs[iter->idx]; in ddebug_iter_first() 703 static struct _ddebug *ddebug_iter_next(struct ddebug_iter *iter) in ddebug_iter_next() argument 705 if (iter->table == NULL) in ddebug_iter_next() 707 if (++iter->idx == iter->table->num_ddebugs) { in ddebug_iter_next() 709 iter->idx = 0; in ddebug_iter_next() [all …]
|
D | radix-tree.c | 748 struct radix_tree_iter *iter, unsigned flags) in radix_tree_next_chunk() argument 766 index = iter->next_index; in radix_tree_next_chunk() 767 if (!index && iter->index) in radix_tree_next_chunk() 775 iter->index = 0; in radix_tree_next_chunk() 776 iter->next_index = 1; in radix_tree_next_chunk() 777 iter->tags = 1; in radix_tree_next_chunk() 831 iter->index = index; in radix_tree_next_chunk() 832 iter->next_index = (index | RADIX_TREE_MAP_MASK) + 1; in radix_tree_next_chunk() 840 iter->tags = node->tags[tag][tag_long] >> tag_bit; in radix_tree_next_chunk() 845 iter->tags |= node->tags[tag][tag_long + 1] << in radix_tree_next_chunk() [all …]
|
/linux-4.1.27/drivers/net/wireless/libertas/ |
D | firmware.c | 85 const struct lbs_fw_table *iter; in load_next_firmware_from_table() local 88 iter = priv->fw_table; in load_next_firmware_from_table() 90 iter = ++priv->fw_iter; in load_next_firmware_from_table() 98 if (!iter->helper) { in load_next_firmware_from_table() 104 if (iter->model != priv->fw_model) { in load_next_firmware_from_table() 105 iter++; in load_next_firmware_from_table() 109 priv->fw_iter = iter; in load_next_firmware_from_table() 110 do_load_firmware(priv, iter->helper, helper_firmware_cb); in load_next_firmware_from_table() 176 const struct lbs_fw_table *iter; in lbs_get_firmware() local 183 iter = fw_table; in lbs_get_firmware() [all …]
|
/linux-4.1.27/kernel/time/ |
D | timer_list.c | 293 struct timer_list_iter *iter = v; in timer_list_show() local 295 if (iter->cpu == -1 && !iter->second_pass) in timer_list_show() 296 timer_list_header(m, iter->now); in timer_list_show() 297 else if (!iter->second_pass) in timer_list_show() 298 print_cpu(m, iter->cpu, iter->now); in timer_list_show() 300 else if (iter->cpu == -1 && iter->second_pass) in timer_list_show() 303 print_tickdevice(m, tick_get_device(iter->cpu), iter->cpu); in timer_list_show() 326 static void *move_iter(struct timer_list_iter *iter, loff_t offset) in move_iter() argument 329 iter->cpu = cpumask_next(iter->cpu, cpu_online_mask); in move_iter() 330 if (iter->cpu >= nr_cpu_ids) { in move_iter() [all …]
|
/linux-4.1.27/drivers/staging/lustre/lustre/lov/ |
D | lov_pool.c | 175 struct pool_iterator *iter = (struct pool_iterator *)s->private; in pool_proc_next() local 178 LASSERTF(iter->magic == POOL_IT_MAGIC, "%08X", iter->magic); in pool_proc_next() 181 if (*pos >= pool_tgt_count(iter->pool)) in pool_proc_next() 185 prev_idx = iter->idx; in pool_proc_next() 186 down_read(&pool_tgt_rw_sem(iter->pool)); in pool_proc_next() 187 iter->idx++; in pool_proc_next() 188 if (iter->idx == pool_tgt_count(iter->pool)) { in pool_proc_next() 189 iter->idx = prev_idx; /* we stay on the last entry */ in pool_proc_next() 190 up_read(&pool_tgt_rw_sem(iter->pool)); in pool_proc_next() 193 up_read(&pool_tgt_rw_sem(iter->pool)); in pool_proc_next() [all …]
|
/linux-4.1.27/mm/ |
D | process_vm_access.c | 36 struct iov_iter *iter, in process_vm_rw_pages() argument 40 while (len && iov_iter_count(iter)) { in process_vm_rw_pages() 49 copied = copy_page_from_iter(page, offset, copy, iter); in process_vm_rw_pages() 52 copied = copy_page_to_iter(page, offset, copy, iter); in process_vm_rw_pages() 55 if (copied < copy && iov_iter_count(iter)) in process_vm_rw_pages() 79 struct iov_iter *iter, in process_vm_rw_single_vec() argument 97 while (!rc && nr_pages && iov_iter_count(iter)) { in process_vm_rw_single_vec() 112 start_offset, bytes, iter, in process_vm_rw_single_vec() 141 static ssize_t process_vm_rw_core(pid_t pid, struct iov_iter *iter, in process_vm_rw_core() argument 155 size_t total_len = iov_iter_count(iter); in process_vm_rw_core() [all …]
|
D | slab.h | 170 #define for_each_memcg_cache(iter, root) \ argument 171 list_for_each_entry(iter, &(root)->memcg_params.list, \ 174 #define for_each_memcg_cache_safe(iter, tmp, root) \ argument 175 list_for_each_entry_safe(iter, tmp, &(root)->memcg_params.list, \ 256 #define for_each_memcg_cache(iter, root) \ argument 257 for ((void)(iter), (void)(root); 0; ) 258 #define for_each_memcg_cache_safe(iter, tmp, root) \ argument 259 for ((void)(iter), (void)(tmp), (void)(root); 0; )
|
D | filemap.c | 1161 struct radix_tree_iter iter; in find_get_entries() local 1168 radix_tree_for_each_slot(slot, &mapping->page_tree, &iter, start) { in find_get_entries() 1193 indices[ret] = iter.index; in find_get_entries() 1221 struct radix_tree_iter iter; in find_get_pages() local 1230 radix_tree_for_each_slot(slot, &mapping->page_tree, &iter, start) { in find_get_pages() 1244 WARN_ON(iter.index); in find_get_pages() 1288 struct radix_tree_iter iter; in find_get_pages_contig() local 1297 radix_tree_for_each_contig(slot, &mapping->page_tree, &iter, index) { in find_get_pages_contig() 1336 if (page->mapping == NULL || page->index != iter.index) { in find_get_pages_contig() 1364 struct radix_tree_iter iter; in find_get_pages_tag() local [all …]
|
D | memcontrol.c | 148 struct reclaim_iter iter[DEF_PRIORITY + 1]; member 1013 struct reclaim_iter *uninitialized_var(iter); in mem_cgroup_iter() 1039 iter = &mz->iter[reclaim->priority]; in mem_cgroup_iter() 1041 if (prev && reclaim->generation != iter->generation) in mem_cgroup_iter() 1045 pos = READ_ONCE(iter->position); in mem_cgroup_iter() 1097 if (cmpxchg(&iter->position, pos, memcg) == pos) { in mem_cgroup_iter() 1112 iter->generation++; in mem_cgroup_iter() 1114 reclaim->generation = iter->generation; in mem_cgroup_iter() 1145 #define for_each_mem_cgroup_tree(iter, root) \ argument 1146 for (iter = mem_cgroup_iter(root, NULL, NULL); \ [all …]
|
/linux-4.1.27/net/ipv4/ |
D | cipso_ipv4.c | 253 u32 iter; in cipso_v4_cache_init() local 261 for (iter = 0; iter < CIPSO_V4_CACHE_BUCKETS; iter++) { in cipso_v4_cache_init() 262 spin_lock_init(&cipso_v4_cache[iter].lock); in cipso_v4_cache_init() 263 cipso_v4_cache[iter].size = 0; in cipso_v4_cache_init() 264 INIT_LIST_HEAD(&cipso_v4_cache[iter].list); in cipso_v4_cache_init() 281 u32 iter; in cipso_v4_cache_invalidate() local 283 for (iter = 0; iter < CIPSO_V4_CACHE_BUCKETS; iter++) { in cipso_v4_cache_invalidate() 284 spin_lock_bh(&cipso_v4_cache[iter].lock); in cipso_v4_cache_invalidate() 287 &cipso_v4_cache[iter].list, list) { in cipso_v4_cache_invalidate() 291 cipso_v4_cache[iter].size = 0; in cipso_v4_cache_invalidate() [all …]
|
D | fib_trie.c | 2010 static struct key_vector *fib_trie_get_next(struct fib_trie_iter *iter) in fib_trie_get_next() argument 2012 unsigned long cindex = iter->index; in fib_trie_get_next() 2013 struct key_vector *pn = iter->tnode; in fib_trie_get_next() 2017 iter->tnode, iter->index, iter->depth); in fib_trie_get_next() 2027 iter->tnode = pn; in fib_trie_get_next() 2028 iter->index = cindex; in fib_trie_get_next() 2031 iter->tnode = n; in fib_trie_get_next() 2032 iter->index = 0; in fib_trie_get_next() 2033 ++iter->depth; in fib_trie_get_next() 2043 --iter->depth; in fib_trie_get_next() [all …]
|
/linux-4.1.27/net/ipv6/ |
D | ip6_fib.c | 694 struct rt6_info *iter = NULL; in fib6_add_rt2node() local 707 for (iter = fn->leaf; iter; iter = iter->dst.rt6_next) { in fib6_add_rt2node() 712 if (iter->rt6i_metric == rt->rt6i_metric) { in fib6_add_rt2node() 720 if (rt_can_ecmp == rt6_qualify_for_ecmp(iter)) { in fib6_add_rt2node() 729 if (iter->dst.dev == rt->dst.dev && in fib6_add_rt2node() 730 iter->rt6i_idev == rt->rt6i_idev && in fib6_add_rt2node() 731 ipv6_addr_equal(&iter->rt6i_gateway, in fib6_add_rt2node() 735 if (!(iter->rt6i_flags & RTF_EXPIRES)) in fib6_add_rt2node() 738 rt6_clean_expires(iter); in fib6_add_rt2node() 740 rt6_set_expires(iter, rt->dst.expires); in fib6_add_rt2node() [all …]
|
D | ip6mr.c | 389 struct ipmr_vif_iter *iter, in ip6mr_vif_seq_idx() argument 392 struct mr6_table *mrt = iter->mrt; in ip6mr_vif_seq_idx() 394 for (iter->ct = 0; iter->ct < mrt->maxvif; ++iter->ct) { in ip6mr_vif_seq_idx() 395 if (!MIF_EXISTS(mrt, iter->ct)) in ip6mr_vif_seq_idx() 398 return &mrt->vif6_table[iter->ct]; in ip6mr_vif_seq_idx() 406 struct ipmr_vif_iter *iter = seq->private; in ip6mr_vif_seq_start() local 414 iter->mrt = mrt; in ip6mr_vif_seq_start() 423 struct ipmr_vif_iter *iter = seq->private; in ip6mr_vif_seq_next() local 425 struct mr6_table *mrt = iter->mrt; in ip6mr_vif_seq_next() 429 return ip6mr_vif_seq_idx(net, iter, 0); in ip6mr_vif_seq_next() [all …]
|
/linux-4.1.27/drivers/base/ |
D | class.c | 294 void class_dev_iter_init(struct class_dev_iter *iter, struct class *class, in class_dev_iter_init() argument 301 klist_iter_init_node(&class->p->klist_devices, &iter->ki, start_knode); in class_dev_iter_init() 302 iter->type = type; in class_dev_iter_init() 318 struct device *class_dev_iter_next(struct class_dev_iter *iter) in class_dev_iter_next() argument 324 knode = klist_next(&iter->ki); in class_dev_iter_next() 328 if (!iter->type || iter->type == dev->type) in class_dev_iter_next() 341 void class_dev_iter_exit(struct class_dev_iter *iter) in class_dev_iter_exit() argument 343 klist_iter_exit(&iter->ki); in class_dev_iter_exit() 368 struct class_dev_iter iter; in class_for_each_device() local 380 class_dev_iter_init(&iter, class, start, NULL); in class_for_each_device() [all …]
|
D | attribute_container.c | 183 #define klist_for_each_entry(pos, head, member, iter) \ argument 184 for (klist_iter_init(head, iter); (pos = ({ \ 185 struct klist_node *n = klist_next(iter); \ 187 ({ klist_iter_exit(iter) ; NULL; }); \ 217 struct klist_iter iter; in attribute_container_remove_device() local 225 klist_for_each_entry(ic, &cont->containers, node, &iter) { in attribute_container_remove_device() 261 struct klist_iter iter; in attribute_container_device_trigger() local 271 klist_for_each_entry(ic, &cont->containers, node, &iter) { in attribute_container_device_trigger() 429 struct klist_iter iter; in attribute_container_find_class_device() local 431 klist_for_each_entry(ic, &cont->containers, node, &iter) { in attribute_container_find_class_device() [all …]
|
D | bus.c | 1062 void subsys_dev_iter_init(struct subsys_dev_iter *iter, struct bus_type *subsys, in subsys_dev_iter_init() argument 1069 klist_iter_init_node(&subsys->p->klist_devices, &iter->ki, start_knode); in subsys_dev_iter_init() 1070 iter->type = type; in subsys_dev_iter_init() 1086 struct device *subsys_dev_iter_next(struct subsys_dev_iter *iter) in subsys_dev_iter_next() argument 1092 knode = klist_next(&iter->ki); in subsys_dev_iter_next() 1096 if (!iter->type || iter->type == dev->type) in subsys_dev_iter_next() 1109 void subsys_dev_iter_exit(struct subsys_dev_iter *iter) in subsys_dev_iter_exit() argument 1111 klist_iter_exit(&iter->ki); in subsys_dev_iter_exit() 1118 struct subsys_dev_iter iter; in subsys_interface_register() local 1131 subsys_dev_iter_init(&iter, subsys, NULL, NULL); in subsys_interface_register() [all …]
|
/linux-4.1.27/drivers/md/bcache/ |
D | bset.c | 55 struct btree_iter iter; in __bch_count_data() local 59 for_each_key(b, k, &iter) in __bch_count_data() 68 struct btree_iter iter; in __bch_check_keys() local 71 for_each_key(b, k, &iter) { in __bch_check_keys() 109 static void bch_btree_iter_next_check(struct btree_iter *iter) in bch_btree_iter_next_check() argument 111 struct bkey *k = iter->data->k, *next = bkey_next(k); in bch_btree_iter_next_check() 113 if (next < iter->data->end && in bch_btree_iter_next_check() 114 bkey_cmp(k, iter->b->ops->is_extents ? in bch_btree_iter_next_check() 116 bch_dump_bucket(iter->b); in bch_btree_iter_next_check() 123 static inline void bch_btree_iter_next_check(struct btree_iter *iter) {} in bch_btree_iter_next_check() argument [all …]
|
D | extents.c | 29 static void sort_key_next(struct btree_iter *iter, in sort_key_next() argument 35 *i = iter->data[--iter->used]; in sort_key_next() 227 struct btree_iter *iter, in bch_btree_ptr_insert_fixup() argument 264 static struct bkey *bch_extent_sort_fixup(struct btree_iter *iter, in bch_extent_sort_fixup() argument 267 while (iter->used > 1) { in bch_extent_sort_fixup() 268 struct btree_iter_set *top = iter->data, *i = top + 1; in bch_extent_sort_fixup() 270 if (iter->used > 2 && in bch_extent_sort_fixup() 278 sort_key_next(iter, i); in bch_extent_sort_fixup() 279 heap_sift(iter, i - top, bch_extent_sort_cmp); in bch_extent_sort_fixup() 285 sort_key_next(iter, i); in bch_extent_sort_fixup() [all …]
|
D | btree.h | 200 #define for_each_cached_btree(b, c, iter) \ argument 201 for (iter = 0; \ 202 iter < ARRAY_SIZE((c)->bucket_hash); \ 203 iter++) \ 204 hlist_for_each_entry_rcu((b), (c)->bucket_hash + iter, hash)
|
D | bset.h | 338 #define for_each_key_filter(b, k, iter, filter) \ argument 339 for (bch_btree_iter_init((b), (iter), NULL); \ 340 ((k) = bch_btree_iter_next_filter((iter), (b), filter));) 342 #define for_each_key(b, k, iter) \ argument 343 for (bch_btree_iter_init((b), (iter), NULL); \ 344 ((k) = bch_btree_iter_next(iter));)
|
D | alloc.c | 414 size_t iter; in bch_bucket_alloc() local 418 for (iter = 0; iter < prio_buckets(ca) * 2; iter++) in bch_bucket_alloc() 419 BUG_ON(ca->prio_buckets[iter] == (uint64_t) r); in bch_bucket_alloc() 422 fifo_for_each(i, &ca->free[j], iter) in bch_bucket_alloc() 424 fifo_for_each(i, &ca->free_inc, iter) in bch_bucket_alloc()
|
D | btree.c | 202 struct btree_iter *iter; in bch_btree_node_read_done() local 204 iter = mempool_alloc(b->c->fill_iter, GFP_NOIO); in bch_btree_node_read_done() 205 iter->size = b->c->sb.bucket_size / b->c->sb.block_size; in bch_btree_node_read_done() 206 iter->used = 0; in bch_btree_node_read_done() 209 iter->b = &b->keys; in bch_btree_node_read_done() 247 bch_btree_iter_push(iter, i->start, bset_bkey_last(i)); in bch_btree_node_read_done() 259 bch_btree_sort_and_fix_extents(&b->keys, iter, &b->c->sort); in bch_btree_node_read_done() 271 mempool_free(iter, b->c->fill_iter); in bch_btree_node_read_done() 1250 struct btree_iter iter; in btree_gc_mark_node() local 1255 for_each_key_filter(&b->keys, k, &iter, bch_ptr_invalid) { in btree_gc_mark_node() [all …]
|
D | bcache.h | 816 #define for_each_cache(ca, cs, iter) \ argument 817 for (iter = 0; ca = cs->cache[iter], iter < (cs)->sb.nr_in_set; iter++)
|
D | util.h | 126 #define fifo_for_each(c, fifo, iter) \ argument 127 for (iter = (fifo)->front; \ 128 c = (fifo)->data[iter], iter != (fifo)->back; \ 129 iter = (iter + 1) & (fifo)->mask)
|
D | debug.c | 110 struct bvec_iter iter; in bch_data_verify() local 122 bio_for_each_segment(bv, bio, iter) { in bch_data_verify() 124 void *p2 = page_address(check->bi_io_vec[iter.bi_idx].bv_page); in bch_data_verify()
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
D | qib_debugfs.c | 193 struct qib_qp_iter *iter; in DEBUGFS_FILE() local 197 iter = qib_qp_iter_init(s->private); in DEBUGFS_FILE() 198 if (!iter) in DEBUGFS_FILE() 202 if (qib_qp_iter_next(iter)) { in DEBUGFS_FILE() 203 kfree(iter); in DEBUGFS_FILE() 208 return iter; in DEBUGFS_FILE() 214 struct qib_qp_iter *iter = iter_ptr; in _qp_stats_seq_next() local 218 if (qib_qp_iter_next(iter)) { in _qp_stats_seq_next() 219 kfree(iter); in _qp_stats_seq_next() 223 return iter; in _qp_stats_seq_next() [all …]
|
D | qib_qp.c | 1323 struct qib_qp_iter *iter; in qib_qp_iter_init() local 1325 iter = kzalloc(sizeof(*iter), GFP_KERNEL); in qib_qp_iter_init() 1326 if (!iter) in qib_qp_iter_init() 1329 iter->dev = dev; in qib_qp_iter_init() 1330 if (qib_qp_iter_next(iter)) { in qib_qp_iter_init() 1331 kfree(iter); in qib_qp_iter_init() 1335 return iter; in qib_qp_iter_init() 1338 int qib_qp_iter_next(struct qib_qp_iter *iter) in qib_qp_iter_next() argument 1340 struct qib_ibdev *dev = iter->dev; in qib_qp_iter_next() 1341 int n = iter->n; in qib_qp_iter_next() [all …]
|
/linux-4.1.27/drivers/gpu/drm/radeon/ |
D | drm_buffer.h | 130 int iter = buffer->iterator + offset * 4; in drm_buffer_pointer_to_dword() local 131 return &buffer->data[iter / PAGE_SIZE][iter & (PAGE_SIZE - 1)]; in drm_buffer_pointer_to_dword() 144 int iter = buffer->iterator + offset; in drm_buffer_pointer_to_byte() local 145 return &buffer->data[iter / PAGE_SIZE][iter & (PAGE_SIZE - 1)]; in drm_buffer_pointer_to_byte()
|
/linux-4.1.27/arch/sparc/prom/ |
D | bootstr_32.c | 18 int iter; in prom_getbootargs() local 30 for (iter = 1; iter < 8; iter++) { in prom_getbootargs() 31 arg = (*(romvec->pv_v0bootargs))->argv[iter]; in prom_getbootargs()
|
/linux-4.1.27/net/sunrpc/ |
D | debugfs.c | 47 struct rpc_clnt_iter *iter = f->private; in tasks_start() local 49 struct rpc_clnt *clnt = iter->clnt; in tasks_start() 52 iter->pos = pos + 1; in tasks_start() 63 struct rpc_clnt_iter *iter = f->private; in tasks_next() local 64 struct rpc_clnt *clnt = iter->clnt; in tasks_next() 68 ++iter->pos; in tasks_next() 81 struct rpc_clnt_iter *iter = f->private; in tasks_stop() local 82 struct rpc_clnt *clnt = iter->clnt; in tasks_stop() 101 struct rpc_clnt_iter *iter = seq->private; in tasks_open() local 103 iter->clnt = inode->i_private; in tasks_open() [all …]
|
/linux-4.1.27/fs/nfs/objlayout/ |
D | pnfs_osd_xdr_cli.c | 160 struct pnfs_osd_xdr_decode_layout_iter *iter, struct xdr_stream *xdr) in pnfs_osd_xdr_decode_layout_map() argument 164 memset(iter, 0, sizeof(*iter)); in pnfs_osd_xdr_decode_layout_map() 176 iter->total_comps = layout->olo_num_comps; in pnfs_osd_xdr_decode_layout_map() 181 struct pnfs_osd_xdr_decode_layout_iter *iter, struct xdr_stream *xdr, in pnfs_osd_xdr_decode_layout_comp() argument 184 BUG_ON(iter->decoded_comps > iter->total_comps); in pnfs_osd_xdr_decode_layout_comp() 185 if (iter->decoded_comps == iter->total_comps) in pnfs_osd_xdr_decode_layout_comp() 192 iter->decoded_comps, iter->total_comps); in pnfs_osd_xdr_decode_layout_comp() 204 iter->decoded_comps++; in pnfs_osd_xdr_decode_layout_comp()
|
/linux-4.1.27/ipc/ |
D | util.c | 778 struct ipc_proc_iter *iter = s->private; in sysvipc_proc_next() local 779 struct ipc_proc_iface *iface = iter->iface; in sysvipc_proc_next() 786 return sysvipc_find_ipc(&iter->ns->ids[iface->ids], *pos, pos); in sysvipc_proc_next() 795 struct ipc_proc_iter *iter = s->private; in sysvipc_proc_start() local 796 struct ipc_proc_iface *iface = iter->iface; in sysvipc_proc_start() 799 ids = &iter->ns->ids[iface->ids]; in sysvipc_proc_start() 822 struct ipc_proc_iter *iter = s->private; in sysvipc_proc_stop() local 823 struct ipc_proc_iface *iface = iter->iface; in sysvipc_proc_stop() 830 ids = &iter->ns->ids[iface->ids]; in sysvipc_proc_stop() 837 struct ipc_proc_iter *iter = s->private; in sysvipc_proc_show() local [all …]
|
/linux-4.1.27/drivers/dma/ |
D | mv_xor.c | 235 struct mv_xor_desc_slot *iter, *_iter; in mv_xor_clean_completed_slots() local 238 list_for_each_entry_safe(iter, _iter, &mv_chan->completed_slots, in mv_xor_clean_completed_slots() 241 if (async_tx_test_ack(&iter->async_tx)) { in mv_xor_clean_completed_slots() 242 list_del(&iter->completed_node); in mv_xor_clean_completed_slots() 243 mv_xor_free_slots(mv_chan, iter); in mv_xor_clean_completed_slots() 272 struct mv_xor_desc_slot *iter, *_iter; in mv_xor_slot_cleanup() local 287 list_for_each_entry_safe(iter, _iter, &mv_chan->chain, in mv_xor_slot_cleanup() 291 hw_desc = iter->hw_desc; in mv_xor_slot_cleanup() 293 cookie = mv_xor_run_tx_complete_actions(iter, mv_chan, in mv_xor_slot_cleanup() 297 mv_xor_clean_slot(iter, mv_chan); in mv_xor_slot_cleanup() [all …]
|
D | iop-adma.c | 116 struct iop_adma_desc_slot *iter, *_iter, *grp_start = NULL; in __iop_adma_slot_cleanup() local 126 list_for_each_entry_safe(iter, _iter, &iop_chan->chain, in __iop_adma_slot_cleanup() 130 iter->async_tx.cookie, iter->idx, busy, in __iop_adma_slot_cleanup() 131 iter->async_tx.phys, iop_desc_get_next_desc(iter), in __iop_adma_slot_cleanup() 132 async_tx_test_ack(&iter->async_tx)); in __iop_adma_slot_cleanup() 147 if (iter->async_tx.phys == current_desc) { in __iop_adma_slot_cleanup() 149 if (busy || iop_desc_get_next_desc(iter)) in __iop_adma_slot_cleanup() 155 slot_cnt = iter->slot_cnt; in __iop_adma_slot_cleanup() 156 slots_per_op = iter->slots_per_op; in __iop_adma_slot_cleanup() 166 grp_start = iter; in __iop_adma_slot_cleanup() [all …]
|
D | fsl-edma.c | 541 u16 soff, doff, iter; in fsl_edma_prep_dma_cyclic() local 554 iter = period_len / nbytes; in fsl_edma_prep_dma_cyclic() 576 fsl_chan->fsc.attr, soff, nbytes, 0, iter, in fsl_edma_prep_dma_cyclic() 577 iter, doff, last_sg, true, false, true); in fsl_edma_prep_dma_cyclic() 593 u16 soff, doff, iter; in fsl_edma_prep_slave_sg() local 621 iter = sg_dma_len(sg) / nbytes; in fsl_edma_prep_slave_sg() 626 nbytes, 0, iter, iter, doff, last_sg, in fsl_edma_prep_slave_sg() 632 nbytes, 0, iter, iter, doff, last_sg, in fsl_edma_prep_slave_sg()
|
/linux-4.1.27/arch/unicore32/include/asm/ |
D | memblock.h | 34 #define for_each_bank(iter, mi) \ argument 35 for (iter = 0; iter < (mi)->nr_banks; iter++)
|
/linux-4.1.27/net/sctp/ |
D | tsnmap.c | 158 struct sctp_tsnmap_iter *iter) in sctp_tsnmap_iter_init() argument 161 iter->start = map->cumulative_tsn_ack_point + 1; in sctp_tsnmap_iter_init() 168 struct sctp_tsnmap_iter *iter, in sctp_tsnmap_next_gap_ack() argument 175 if (TSN_lte(map->max_tsn_seen, iter->start)) in sctp_tsnmap_next_gap_ack() 178 offset = iter->start - map->base_tsn; in sctp_tsnmap_next_gap_ack() 197 iter->start = map->cumulative_tsn_ack_point + *end + 1; in sctp_tsnmap_next_gap_ack() 335 struct sctp_tsnmap_iter iter; in sctp_tsnmap_num_gabs() local 341 sctp_tsnmap_iter_init(map, &iter); in sctp_tsnmap_num_gabs() 342 while (sctp_tsnmap_next_gap_ack(map, &iter, in sctp_tsnmap_num_gabs()
|
/linux-4.1.27/security/selinux/ |
D | netport.c | 239 int iter; in sel_netport_init() local 244 for (iter = 0; iter < SEL_NETPORT_HASH_SIZE; iter++) { in sel_netport_init() 245 INIT_LIST_HEAD(&sel_netport_hash[iter].list); in sel_netport_init() 246 sel_netport_hash[iter].size = 0; in sel_netport_init()
|
D | netnode.c | 305 int iter; in sel_netnode_init() local 310 for (iter = 0; iter < SEL_NETNODE_HASH_SIZE; iter++) { in sel_netnode_init() 311 INIT_LIST_HEAD(&sel_netnode_hash[iter].list); in sel_netnode_init() 312 sel_netnode_hash[iter].size = 0; in sel_netnode_init()
|
D | xfrm.c | 456 struct dst_entry *iter; in selinux_xfrm_postroute_last() local 458 for (iter = dst; iter != NULL; iter = iter->child) { in selinux_xfrm_postroute_last() 459 struct xfrm_state *x = iter->xfrm; in selinux_xfrm_postroute_last()
|
/linux-4.1.27/net/netfilter/ |
D | nft_hash.c | 186 struct nft_set_iter *iter) in nft_hash_walk() argument 196 iter->err = err; in nft_hash_walk() 202 iter->err = err; in nft_hash_walk() 210 iter->err = err; in nft_hash_walk() 217 if (iter->count < iter->skip) in nft_hash_walk() 226 iter->err = iter->fn(ctx, set, iter, &elem); in nft_hash_walk() 227 if (iter->err < 0) in nft_hash_walk() 231 iter->count++; in nft_hash_walk()
|
D | nft_rbtree.c | 175 struct nft_set_iter *iter) in nft_rbtree_walk() argument 187 if (iter->count < iter->skip) in nft_rbtree_walk() 194 iter->err = iter->fn(ctx, set, iter, &elem); in nft_rbtree_walk() 195 if (iter->err < 0) { in nft_rbtree_walk() 200 iter->count++; in nft_rbtree_walk()
|
D | nf_tables_api.c | 2842 const struct nft_set_iter *iter, in nf_tables_bind_check_setelem() argument 2859 struct nft_set_iter iter; in nf_tables_bind_set() local 2874 iter.skip = 0; in nf_tables_bind_set() 2875 iter.count = 0; in nf_tables_bind_set() 2876 iter.err = 0; in nf_tables_bind_set() 2877 iter.fn = nf_tables_bind_check_setelem; in nf_tables_bind_set() 2879 set->ops->walk(ctx, set, &iter); in nf_tables_bind_set() 2880 if (iter.err < 0) { in nf_tables_bind_set() 2885 return iter.err; in nf_tables_bind_set() 3047 struct nft_set_iter iter; member [all …]
|
/linux-4.1.27/drivers/scsi/qla4xxx/ |
D | ql4_attr.c | 129 struct sysfs_entry *iter; in qla4_8xxx_alloc_sysfs_attr() local 132 for (iter = bin_file_entries; iter->name; iter++) { in qla4_8xxx_alloc_sysfs_attr() 134 iter->attr); in qla4_8xxx_alloc_sysfs_attr() 138 iter->name, ret); in qla4_8xxx_alloc_sysfs_attr() 145 struct sysfs_entry *iter; in qla4_8xxx_free_sysfs_attr() local 147 for (iter = bin_file_entries; iter->name; iter++) in qla4_8xxx_free_sysfs_attr() 149 iter->attr); in qla4_8xxx_free_sysfs_attr()
|
/linux-4.1.27/fs/nfs/ |
D | direct.c | 256 ssize_t nfs_direct_IO(struct kiocb *iocb, struct iov_iter *iter, loff_t pos) in nfs_direct_IO() argument 264 VM_BUG_ON(iov_iter_count(iter) != PAGE_SIZE); in nfs_direct_IO() 266 if (iov_iter_rw(iter) == READ) in nfs_direct_IO() 267 return nfs_file_direct_read(iocb, iter, pos); in nfs_direct_IO() 268 return nfs_file_direct_write(iocb, iter); in nfs_direct_IO() 472 struct iov_iter *iter, in nfs_direct_read_schedule_iovec() argument 487 while (iov_iter_count(iter)) { in nfs_direct_read_schedule_iovec() 493 result = iov_iter_get_pages_alloc(iter, &pagevec, in nfs_direct_read_schedule_iovec() 499 iov_iter_advance(iter, bytes); in nfs_direct_read_schedule_iovec() 567 ssize_t nfs_file_direct_read(struct kiocb *iocb, struct iov_iter *iter, in nfs_file_direct_read() argument [all …]
|
/linux-4.1.27/tools/perf/ui/gtk/ |
D | hists.c | 101 GtkTreeIter iter, new_parent; in perf_gtk__add_callchain() local 117 gtk_tree_store_append(store, &iter, &new_parent); in perf_gtk__add_callchain() 120 gtk_tree_store_set(store, &iter, 0, buf, -1); in perf_gtk__add_callchain() 123 gtk_tree_store_set(store, &iter, col, buf, -1); in perf_gtk__add_callchain() 130 new_parent = iter; in perf_gtk__add_callchain() 141 perf_gtk__add_callchain(&node->rb_root, store, &iter, col, in perf_gtk__add_callchain() 225 GtkTreeIter iter; in perf_gtk__show_hists() local 236 gtk_tree_store_append(store, &iter, NULL); in perf_gtk__show_hists() 249 gtk_tree_store_set(store, &iter, col_idx++, s, -1); in perf_gtk__show_hists() 257 perf_gtk__add_callchain(&h->sorted_chain, store, &iter, in perf_gtk__show_hists()
|
D | annotate.c | 121 GtkTreeIter iter; in perf_gtk__annotate_symbol() local 124 gtk_list_store_append(store, &iter); in perf_gtk__annotate_symbol() 140 gtk_list_store_set(store, &iter, ANN_COL__PERCENT, s, -1); in perf_gtk__annotate_symbol() 142 gtk_list_store_set(store, &iter, ANN_COL__OFFSET, s, -1); in perf_gtk__annotate_symbol() 144 gtk_list_store_set(store, &iter, ANN_COL__LINE, s, -1); in perf_gtk__annotate_symbol()
|
/linux-4.1.27/include/net/ |
D | bonding.h | 82 #define bond_for_each_slave(bond, pos, iter) \ argument 83 netdev_for_each_lower_private((bond)->dev, pos, iter) 86 #define bond_for_each_slave_rcu(bond, pos, iter) \ argument 87 netdev_for_each_lower_private_rcu((bond)->dev, pos, iter) 354 struct list_head *iter; in bond_slave_state_change() local 357 bond_for_each_slave(bond, tmp, iter) { in bond_slave_state_change() 367 struct list_head *iter; in bond_slave_state_notify() local 370 bond_for_each_slave(bond, tmp, iter) { in bond_slave_state_notify() 585 struct list_head *iter; in bond_slave_has_mac() local 588 bond_for_each_slave(bond, tmp, iter) in bond_slave_has_mac() [all …]
|
D | netlabel.h | 284 struct netlbl_lsm_catmap *iter; in netlbl_catmap_free() local 287 iter = catmap; in netlbl_catmap_free() 289 kfree(iter); in netlbl_catmap_free()
|
/linux-4.1.27/drivers/s390/block/ |
D | scm_blk_cluster.c | 79 struct scm_request *iter; in scm_reserve_cluster() local 86 list_for_each_entry(iter, &bdev->cluster_list, cluster.list) { in scm_reserve_cluster() 87 if (iter == scmrq) { in scm_reserve_cluster() 95 for (pos = 0; pos < iter->aob->request.msb_count; pos++) { in scm_reserve_cluster() 96 if (clusters_intersect(req, iter->request[pos]) && in scm_reserve_cluster() 98 rq_data_dir(iter->request[pos]) == WRITE)) { in scm_reserve_cluster() 136 struct req_iterator iter; in scm_prepare_cluster_request() local 183 rq_for_each_segment(bv, req, iter) { in scm_prepare_cluster_request()
|
D | scm_blk.c | 52 struct list_head *iter, *safe; in scm_free_rqs() local 56 list_for_each_safe(iter, safe, &inactive_requests) { in scm_free_rqs() 57 scmrq = list_entry(iter, struct scm_request, list); in scm_free_rqs() 188 struct req_iterator iter; in scm_request_prepare() local 203 rq_for_each_segment(bv, req, iter) { in scm_request_prepare()
|
D | dasd_fba.c | 262 struct req_iterator iter; in dasd_fba_build_cp() local 285 rq_for_each_segment(bv, req, iter) { in dasd_fba_build_cp() 326 rq_for_each_segment(bv, req, iter) { in dasd_fba_build_cp() 384 struct req_iterator iter; in dasd_fba_free_cp() local 399 rq_for_each_segment(bv, req, iter) { in dasd_fba_free_cp()
|
/linux-4.1.27/net/switchdev/ |
D | switchdev.c | 52 struct list_head *iter; in netdev_switch_port_stp_update() local 58 netdev_for_each_lower_dev(dev, lower_dev, iter) { in netdev_switch_port_stp_update() 195 struct list_head *iter; in ndo_dflt_netdev_switch_port_bridge_setlink() local 201 netdev_for_each_lower_dev(dev, lower_dev, iter) { in ndo_dflt_netdev_switch_port_bridge_setlink() 225 struct list_head *iter; in ndo_dflt_netdev_switch_port_bridge_dellink() local 231 netdev_for_each_lower_dev(dev, lower_dev, iter) { in ndo_dflt_netdev_switch_port_bridge_dellink() 246 struct list_head *iter; in netdev_switch_get_lowest_dev() local 256 netdev_for_each_lower_dev(dev, lower_dev, iter) { in netdev_switch_get_lowest_dev()
|
/linux-4.1.27/drivers/staging/lustre/lustre/llite/ |
D | rw26.c | 362 static ssize_t ll_direct_IO_26(struct kiocb *iocb, struct iov_iter *iter, in ll_direct_IO_26() argument 370 ssize_t count = iov_iter_count(iter); in ll_direct_IO_26() 390 if (iov_iter_alignment(iter) & ~CFS_PAGE_MASK) in ll_direct_IO_26() 402 if (iov_iter_rw(iter) == READ) in ll_direct_IO_26() 406 while (iov_iter_count(iter)) { in ll_direct_IO_26() 410 count = min_t(size_t, iov_iter_count(iter), size); in ll_direct_IO_26() 411 if (iov_iter_rw(iter) == READ) { in ll_direct_IO_26() 418 result = iov_iter_get_pages_alloc(iter, &pages, count, &offs); in ll_direct_IO_26() 421 result = ll_direct_IO_26_seg(env, io, iov_iter_rw(iter), in ll_direct_IO_26() 425 ll_free_user_pages(pages, n, iov_iter_rw(iter) == READ); in ll_direct_IO_26() [all …]
|
/linux-4.1.27/drivers/net/wireless/ath/carl9170/ |
D | debug.c | 301 struct carl9170_sta_tid *iter; in carl9170_debugfs_ampdu_state_read() local 307 list_for_each_entry_rcu(iter, &ar->tx_ampdu_list, list) { in carl9170_debugfs_ampdu_state_read() 309 spin_lock_bh(&iter->lock); in carl9170_debugfs_ampdu_state_read() 312 cnt, iter->tid, iter->bsn, iter->snx, iter->hsn, in carl9170_debugfs_ampdu_state_read() 313 iter->max, iter->state, iter->counter); in carl9170_debugfs_ampdu_state_read() 316 CARL9170_BAW_BITS, iter->bitmap); in carl9170_debugfs_ampdu_state_read() 325 offset = BM_STR_OFF(SEQ_DIFF(iter->snx, iter->bsn)); in carl9170_debugfs_ampdu_state_read() 328 offset = BM_STR_OFF(((int)iter->hsn - (int)iter->bsn) % in carl9170_debugfs_ampdu_state_read() 333 " currently queued:%d\n", skb_queue_len(&iter->queue)); in carl9170_debugfs_ampdu_state_read() 336 skb_queue_walk(&iter->queue, skb) { in carl9170_debugfs_ampdu_state_read() [all …]
|
D | fw.c | 36 const struct carl9170fw_desc_head *iter; in carl9170_fw_find_desc() local 38 carl9170fw_for_each_hdr(iter, ar->fw.desc) { in carl9170_fw_find_desc() 39 if (carl9170fw_desc_cmp(iter, descid, len, in carl9170_fw_find_desc() 41 return (void *)iter; in carl9170_fw_find_desc() 45 if (carl9170fw_desc_cmp(iter, descid, len, in carl9170_fw_find_desc() 47 return (void *)iter; in carl9170_fw_find_desc()
|
D | tx.c | 605 struct carl9170_sta_tid *iter; in carl9170_tx_ampdu_timeout() local 612 list_for_each_entry_rcu(iter, &ar->tx_ampdu_list, list) { in carl9170_tx_ampdu_timeout() 613 if (iter->state < CARL9170_TID_STATE_IDLE) in carl9170_tx_ampdu_timeout() 616 spin_lock_bh(&iter->lock); in carl9170_tx_ampdu_timeout() 617 skb = skb_peek(&iter->queue); in carl9170_tx_ampdu_timeout() 627 sta = iter->sta; in carl9170_tx_ampdu_timeout() 631 ieee80211_stop_tx_ba_session(sta, iter->tid); in carl9170_tx_ampdu_timeout() 633 spin_unlock_bh(&iter->lock); in carl9170_tx_ampdu_timeout() 1397 struct sk_buff *iter; in carl9170_tx_ampdu_queue() local 1432 skb_queue_reverse_walk(&agg->queue, iter) { in carl9170_tx_ampdu_queue() [all …]
|
/linux-4.1.27/net/nfc/ |
D | nfc.h | 113 static inline void nfc_device_iter_init(struct class_dev_iter *iter) in nfc_device_iter_init() argument 115 class_dev_iter_init(iter, &nfc_class, NULL, NULL); in nfc_device_iter_init() 118 static inline struct nfc_dev *nfc_device_iter_next(struct class_dev_iter *iter) in nfc_device_iter_next() argument 120 struct device *d = class_dev_iter_next(iter); in nfc_device_iter_next() 127 static inline void nfc_device_iter_exit(struct class_dev_iter *iter) in nfc_device_iter_exit() argument 129 class_dev_iter_exit(iter); in nfc_device_iter_exit()
|
D | netlink.c | 580 struct class_dev_iter *iter = (struct class_dev_iter *) cb->args[0]; in nfc_genl_dump_devices() local 584 if (!iter) { in nfc_genl_dump_devices() 586 iter = kmalloc(sizeof(struct class_dev_iter), GFP_KERNEL); in nfc_genl_dump_devices() 587 if (!iter) in nfc_genl_dump_devices() 589 cb->args[0] = (long) iter; in nfc_genl_dump_devices() 597 nfc_device_iter_init(iter); in nfc_genl_dump_devices() 598 dev = nfc_device_iter_next(iter); in nfc_genl_dump_devices() 609 dev = nfc_device_iter_next(iter); in nfc_genl_dump_devices() 621 struct class_dev_iter *iter = (struct class_dev_iter *) cb->args[0]; in nfc_genl_dump_devices_done() local 623 nfc_device_iter_exit(iter); in nfc_genl_dump_devices_done() [all …]
|
/linux-4.1.27/fs/ |
D | dax.c | 101 static ssize_t dax_io(struct inode *inode, struct iov_iter *iter, in dax_io() argument 112 if (iov_iter_rw(iter) != WRITE) in dax_io() 127 iov_iter_rw(iter) == WRITE); in dax_io() 140 hole = iov_iter_rw(iter) != WRITE && !buffer_written(bh); in dax_io() 157 if (iov_iter_rw(iter) == WRITE) in dax_io() 158 len = copy_from_iter(addr, max - pos, iter); in dax_io() 160 len = copy_to_iter(addr, max - pos, iter); in dax_io() 162 len = iov_iter_zero(max - pos, iter); in dax_io() 192 struct iov_iter *iter, loff_t pos, get_block_t get_block, in dax_do_io() argument 197 loff_t end = pos + iov_iter_count(iter); in dax_do_io() [all …]
|
D | read_write.c | 332 ssize_t vfs_iter_read(struct file *file, struct iov_iter *iter, loff_t *ppos) in vfs_iter_read() argument 343 iter->type |= READ; in vfs_iter_read() 344 ret = file->f_op->read_iter(&kiocb, iter); in vfs_iter_read() 352 ssize_t vfs_iter_write(struct file *file, struct iov_iter *iter, loff_t *ppos) in vfs_iter_write() argument 363 iter->type |= WRITE; in vfs_iter_write() 364 ret = file->f_op->write_iter(&kiocb, iter); in vfs_iter_write() 415 struct iov_iter iter; in new_sync_read() local 420 iov_iter_init(&iter, READ, &iov, 1, len); in new_sync_read() 422 ret = filp->f_op->read_iter(&kiocb, &iter); in new_sync_read() 471 struct iov_iter iter; in new_sync_write() local [all …]
|
D | direct-io.c | 98 struct iov_iter *iter; member 160 ret = iov_iter_get_pages(sdio->iter, dio->pages, LONG_MAX, DIO_PAGES, in dio_refill_pages() 182 iov_iter_advance(sdio->iter, ret); in dio_refill_pages() 1099 struct block_device *bdev, struct iov_iter *iter, in do_blockdev_direct_IO() argument 1107 size_t count = iov_iter_count(iter); in do_blockdev_direct_IO() 1113 unsigned long align = offset | iov_iter_alignment(iter); in do_blockdev_direct_IO() 1129 if (iov_iter_rw(iter) == READ && !iov_iter_count(iter)) in do_blockdev_direct_IO() 1145 if (iov_iter_rw(iter) == READ) { in do_blockdev_direct_IO() 1164 if (iov_iter_rw(iter) == READ && offset >= dio->i_size) { in do_blockdev_direct_IO() 1181 iov_iter_rw(iter) == WRITE && end > i_size_read(inode)) in do_blockdev_direct_IO() [all …]
|
/linux-4.1.27/drivers/hid/ |
D | hid-wiimote-core.c | 619 const __u8 *mods, *iter; in wiimote_modules_load() local 625 for (iter = mods; *iter != WIIMOD_NULL; ++iter) { in wiimote_modules_load() 626 if (wiimod_table[*iter]->flags & WIIMOD_FLAG_INPUT) { in wiimote_modules_load() 646 for (iter = mods; *iter != WIIMOD_NULL; ++iter) { in wiimote_modules_load() 647 ops = wiimod_table[*iter]; in wiimote_modules_load() 668 for ( ; iter-- != mods; ) { in wiimote_modules_load() 669 ops = wiimod_table[*iter]; in wiimote_modules_load() 682 const __u8 *mods, *iter; in wiimote_modules_unload() local 693 for (iter = mods; *iter != WIIMOD_NULL; ++iter) in wiimote_modules_unload() 701 for ( ; iter-- != mods; ) { in wiimote_modules_unload() [all …]
|
/linux-4.1.27/drivers/block/aoe/ |
D | aoecmd.c | 199 memset(&f->iter, 0, sizeof(f->iter)); in aoe_freetframe() 297 skb_fillup(struct sk_buff *skb, struct bio *bio, struct bvec_iter iter) in skb_fillup() argument 302 __bio_for_each_segment(bv, bio, iter, iter) in skb_fillup() 343 ah->scnt = f->iter.bi_size >> 9; in ata_rw_frameinit() 344 put_lba(ah, f->iter.bi_sector); in ata_rw_frameinit() 353 skb_fillup(skb, f->buf->bio, f->iter); in ata_rw_frameinit() 355 skb->len += f->iter.bi_size; in ata_rw_frameinit() 356 skb->data_len = f->iter.bi_size; in ata_rw_frameinit() 357 skb->truesize += f->iter.bi_size; in ata_rw_frameinit() 385 f->iter = buf->iter; in aoecmd_ata_rw() [all …]
|
D | aoe.h | 104 struct bvec_iter iter; member 123 struct bvec_iter iter; member
|
/linux-4.1.27/arch/powerpc/kernel/ |
D | cacheinfo.c | 166 struct cache *iter; in release_cache_debugcheck() local 168 list_for_each_entry(iter, &cache_list, list) in release_cache_debugcheck() 169 WARN_ONCE(iter->next_local == cache, in release_cache_debugcheck() 171 iter->ofnode->full_name, in release_cache_debugcheck() 172 cache_type_string(iter), in release_cache_debugcheck() 304 struct cache *iter; in cache_find_first_sibling() local 310 list_for_each_entry(iter, &cache_list, list) in cache_find_first_sibling() 311 if (iter->ofnode == cache->ofnode && iter->next_local == cache) in cache_find_first_sibling() 312 return iter; in cache_find_first_sibling() 321 struct cache *iter; in cache_lookup_by_node() local [all …]
|
D | ftrace.c | 443 struct ftrace_rec_iter *iter; in ftrace_replace_code() local 447 for (iter = ftrace_rec_iter_start(); iter; in ftrace_replace_code() 448 iter = ftrace_rec_iter_next(iter)) { in ftrace_replace_code() 449 rec = ftrace_rec_iter_record(iter); in ftrace_replace_code()
|
/linux-4.1.27/net/appletalk/ |
D | aarp.c | 921 static struct aarp_entry *iter_next(struct aarp_iter_state *iter, loff_t *pos) in iter_next() argument 923 int ct = iter->bucket; in iter_next() 924 struct aarp_entry **table = iter->table; in iter_next() 932 iter->table = table; in iter_next() 933 iter->bucket = ct; in iter_next() 956 struct aarp_iter_state *iter = seq->private; in aarp_seq_start() local 959 iter->table = resolved; in aarp_seq_start() 960 iter->bucket = 0; in aarp_seq_start() 962 return *pos ? iter_next(iter, pos) : SEQ_START_TOKEN; in aarp_seq_start() 968 struct aarp_iter_state *iter = seq->private; in aarp_seq_next() local [all …]
|
/linux-4.1.27/arch/arm/include/asm/hardware/ |
D | iop3xx-adma.h | 576 struct iop3xx_desc_aau *hw_desc, *prev_hw_desc, *iter; in iop_desc_init_zero_sum() local 587 iter = iop_hw_desc_slot_idx(hw_desc, i); in iop_desc_init_zero_sum() 588 u_desc_ctrl.value = iop3xx_desc_init_xor(iter, src_cnt, flags); in iop_desc_init_zero_sum() 592 iter->desc_ctrl = u_desc_ctrl.value; in iop_desc_init_zero_sum() 695 struct iop3xx_desc_aau *hw_desc = desc->hw_desc, *iter; in iop_desc_set_zero_sum_byte_count() local 702 iter = iop_hw_desc_slot_idx(hw_desc, i); in iop_desc_set_zero_sum_byte_count() 703 iter->byte_count = IOP_ADMA_ZERO_SUM_MAX_BYTE_COUNT; in iop_desc_set_zero_sum_byte_count() 708 iter = iop_hw_desc_slot_idx(hw_desc, i); in iop_desc_set_zero_sum_byte_count() 709 iter->byte_count = len; in iop_desc_set_zero_sum_byte_count() 744 struct iop3xx_desc_aau *hw_desc = desc->hw_desc, *iter; in iop_desc_set_zero_sum_src_addr() local [all …]
|
/linux-4.1.27/drivers/net/bonding/ |
D | bond_main.c | 290 struct list_head *iter; in bond_vlan_rx_add_vid() local 293 bond_for_each_slave(bond, slave, iter) { in bond_vlan_rx_add_vid() 303 bond_for_each_slave(bond, rollback_slave, iter) { in bond_vlan_rx_add_vid() 322 struct list_head *iter; in bond_vlan_rx_kill_vid() local 325 bond_for_each_slave(bond, slave, iter) in bond_vlan_rx_kill_vid() 344 struct list_head *iter; in bond_set_carrier() local 353 bond_for_each_slave(bond, slave, iter) { in bond_set_carrier() 496 struct list_head *iter; in bond_set_promiscuity() local 507 bond_for_each_slave(bond, slave, iter) { in bond_set_promiscuity() 519 struct list_head *iter; in bond_set_allmulti() local [all …]
|
D | bond_procfs.c | 13 struct list_head *iter; in bond_info_seq_start() local 22 bond_for_each_slave_rcu(bond, slave, iter) in bond_info_seq_start() 32 struct list_head *iter; in bond_info_seq_next() local 40 bond_for_each_slave_rcu(bond, slave, iter) { in bond_info_seq_next()
|
D | bond_3ad.c | 701 struct list_head *iter; in __get_active_agg() local 704 bond_for_each_slave_rcu(bond, slave, iter) in __get_active_agg() 1311 struct list_head *iter; in ad_port_selection_logic() local 1375 bond_for_each_slave(bond, slave, iter) { in ad_port_selection_logic() 1583 struct list_head *iter; in ad_agg_selection_logic() local 1592 bond_for_each_slave_rcu(bond, slave, iter) { in ad_agg_selection_logic() 1637 bond_for_each_slave_rcu(bond, slave, iter) { in ad_agg_selection_logic() 1997 struct list_head *iter; in bond_3ad_unbind_slave() local 2030 bond_for_each_slave(bond, slave_iter, iter) { in bond_3ad_unbind_slave() 2107 bond_for_each_slave(bond, slave_iter, iter) { in bond_3ad_unbind_slave() [all …]
|
D | bond_alb.c | 198 struct list_head *iter; in tlb_get_least_loaded_slave() local 205 bond_for_each_slave_rcu(bond, slave, iter) { in tlb_get_least_loaded_slave() 338 struct list_head *iter; in __rlb_next_rx_slave() local 341 bond_for_each_slave_rcu(bond, slave, iter) { in __rlb_next_rx_slave() 961 struct list_head *iter; in alb_send_learning_packets() local 971 netdev_for_each_all_upper_dev_rcu(bond->dev, upper, iter) { in alb_send_learning_packets() 1143 struct list_head *iter; in alb_handle_addr_collision_on_attach() local 1167 bond_for_each_slave(bond, tmp_slave1, iter) { in alb_handle_addr_collision_on_attach() 1217 struct list_head *iter; in alb_set_mac_address() local 1225 bond_for_each_slave(bond, slave, iter) { in alb_set_mac_address() [all …]
|
D | bond_options.c | 911 struct list_head *iter; in _bond_options_arp_ip_target_set() local 915 bond_for_each_slave(bond, slave, iter) in _bond_options_arp_ip_target_set() 959 struct list_head *iter; in bond_option_arp_ip_target_rem() local 982 bond_for_each_slave(bond, slave, iter) { in bond_option_arp_ip_target_rem() 1060 struct list_head *iter; in bond_option_primary_set() local 1077 bond_for_each_slave(bond, slave, iter) { in bond_option_primary_set() 1160 struct list_head *iter; in bond_option_all_slaves_active_set() local 1166 bond_for_each_slave(bond, slave, iter) { in bond_option_all_slaves_active_set() 1241 struct list_head *iter; in bond_option_queue_id_set() local 1270 bond_for_each_slave(bond, slave, iter) { in bond_option_queue_id_set()
|
/linux-4.1.27/arch/arm/mach-iop13xx/include/mach/ |
D | adma.h | 372 struct iop13xx_adma_desc_hw *hw_desc = desc->hw_desc, *iter; in iop_desc_set_zero_sum_byte_count() local 379 iter = iop_hw_desc_slot_idx(hw_desc, i); in iop_desc_set_zero_sum_byte_count() 380 iter->byte_count = IOP_ADMA_ZERO_SUM_MAX_BYTE_COUNT; in iop_desc_set_zero_sum_byte_count() 386 iter = iop_hw_desc_slot_idx(hw_desc, i); in iop_desc_set_zero_sum_byte_count() 387 iter->byte_count = len; in iop_desc_set_zero_sum_byte_count() 425 struct iop13xx_adma_desc_hw *hw_desc = desc->hw_desc, *iter; in iop_desc_set_xor_src_addr() local 429 iter = iop_hw_desc_slot_idx(hw_desc, i); in iop_desc_set_xor_src_addr() 430 iter->src[src_idx].src_addr = addr; in iop_desc_set_xor_src_addr() 431 iter->src[src_idx].upper_src_addr = 0; in iop_desc_set_xor_src_addr() 445 struct iop13xx_adma_desc_hw *hw_desc = desc->hw_desc, *iter; in iop_desc_set_pq_src_addr() local [all …]
|
/linux-4.1.27/net/irda/ |
D | irlmp.c | 1834 static void *irlmp_seq_hb_idx(struct irlmp_iter_state *iter, loff_t *off) in irlmp_seq_hb_idx() argument 1838 spin_lock_irq(&iter->hashbin->hb_spinlock); in irlmp_seq_hb_idx() 1839 for (element = hashbin_get_first(iter->hashbin); in irlmp_seq_hb_idx() 1841 element = hashbin_get_next(iter->hashbin)) { in irlmp_seq_hb_idx() 1847 spin_unlock_irq(&iter->hashbin->hb_spinlock); in irlmp_seq_hb_idx() 1848 iter->hashbin = NULL; in irlmp_seq_hb_idx() 1855 struct irlmp_iter_state *iter = seq->private; in irlmp_seq_start() local 1859 iter->hashbin = NULL; in irlmp_seq_start() 1863 iter->hashbin = irlmp->unconnected_lsaps; in irlmp_seq_start() 1864 v = irlmp_seq_hb_idx(iter, &off); in irlmp_seq_start() [all …]
|
D | irlap.c | 1075 struct irlap_iter_state *iter = seq->private; in irlap_seq_start() local 1080 iter->id = 0; in irlap_seq_start() 1084 if (iter->id == *pos) in irlap_seq_start() 1086 ++iter->id; in irlap_seq_start() 1094 struct irlap_iter_state *iter = seq->private; in irlap_seq_next() local 1097 ++iter->id; in irlap_seq_next() 1108 const struct irlap_iter_state *iter = seq->private; in irlap_seq_show() local 1113 seq_printf(seq, "irlap%d ", iter->id); in irlap_seq_show()
|
/linux-4.1.27/drivers/gpu/drm/vmwgfx/ |
D | vmwgfx_gmr.c | 39 struct vmw_piter *iter, in vmw_gmr2_bind() argument 86 *cmd = vmw_piter_dma_addr(iter) >> PAGE_SHIFT; in vmw_gmr2_bind() 88 *((uint64_t *)cmd) = vmw_piter_dma_addr(iter) >> in vmw_gmr2_bind() 92 vmw_piter_next(iter); in vmw_gmr2_bind()
|
D | vmwgfx_buffer.c | 233 return __sg_page_iter_next(&viter->iter); in __vmw_piter_sg_next() 253 return sg_page_iter_page(&viter->iter); in __vmw_piter_sg_page() 278 return sg_page_iter_dma_address(&viter->iter); in __vmw_piter_sg_addr() 316 __sg_page_iter_start(&viter->iter, vsgt->sgt->sgl, in vmw_piter_start() 384 struct vmw_piter iter; in vmw_ttm_map_dma() local 440 for (vmw_piter_start(&iter, vsgt, 0); vmw_piter_next(&iter);) { in vmw_ttm_map_dma() 441 dma_addr_t cur = vmw_piter_dma_addr(&iter); in vmw_ttm_map_dma()
|
D | vmwgfx_mob.c | 104 struct vmw_piter iter; in vmw_setup_otable_base() local 110 vmw_piter_start(&iter, vsgt, offset >> PAGE_SHIFT); in vmw_setup_otable_base() 111 WARN_ON(!vmw_piter_next(&iter)); in vmw_setup_otable_base() 121 mob->pt_root_page = vmw_piter_dma_addr(&iter); in vmw_setup_otable_base() 124 mob->pt_root_page = vmw_piter_dma_addr(&iter); in vmw_setup_otable_base() 130 vmw_mob_pt_setup(mob, iter, otable->size >> PAGE_SHIFT); in vmw_setup_otable_base()
|
/linux-4.1.27/scripts/coccinelle/null/ |
D | deref_null.cocci | 57 iterator iter; 66 iter(subE,...) S4 // no use 118 iterator iter; 127 iter(subE,...) S4 // no use 178 iterator iter; 187 iter(subE,...) S4 // no use 251 iterator iter; 260 iter(subE,...) S4 // no use
|
/linux-4.1.27/net/ipv4/netfilter/ |
D | arp_tables.c | 625 struct arpt_entry *iter; in translate_table() local 642 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table() 643 ret = check_entry_size_and_hooks(iter, newinfo, entry0, in translate_table() 651 if (strcmp(arpt_get_target(iter)->u.user.name, in translate_table() 689 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table() 690 ret = find_check_entry(iter, repl->name, repl->size); in translate_table() 697 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table() 700 cleanup_entry(iter); in translate_table() 717 struct arpt_entry *iter; in get_counters() local 725 xt_entry_foreach(iter, t->entries[cpu], t->size) { in get_counters() [all …]
|
D | ip_tables.c | 260 const struct ipt_entry *iter; in trace_packet() local 270 xt_entry_foreach(iter, root, private->size - private->hook_entry[hook]) in trace_packet() 271 if (get_chainname_rulenum(iter, e, hookname, in trace_packet() 795 struct ipt_entry *iter; in translate_table() local 811 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table() 812 ret = check_entry_size_and_hooks(iter, newinfo, entry0, in translate_table() 820 if (strcmp(ipt_get_target(iter)->u.user.name, in translate_table() 853 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table() 854 ret = find_check_entry(iter, net, repl->name, repl->size); in translate_table() 861 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table() [all …]
|
/linux-4.1.27/net/ipv6/netfilter/ |
D | ip6_tables.c | 289 const struct ip6t_entry *iter; in trace_packet() local 299 xt_entry_foreach(iter, root, private->size - private->hook_entry[hook]) in trace_packet() 300 if (get_chainname_rulenum(iter, e, hookname, in trace_packet() 808 struct ip6t_entry *iter; in translate_table() local 824 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table() 825 ret = check_entry_size_and_hooks(iter, newinfo, entry0, in translate_table() 833 if (strcmp(ip6t_get_target(iter)->u.user.name, in translate_table() 866 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table() 867 ret = find_check_entry(iter, net, repl->name, repl->size); in translate_table() 874 xt_entry_foreach(iter, entry0, newinfo->size) { in translate_table() [all …]
|
/linux-4.1.27/kernel/sched/ |
D | rt.c | 459 #define for_each_rt_rq(rt_rq, iter, rq) \ argument 460 for (iter = container_of(&task_groups, typeof(*iter), list); \ 461 (iter = next_task_group(iter)) && \ 462 (rt_rq = iter->rt_rq[cpu_of(rq)]);) 563 #define for_each_rt_rq(rt_rq, iter, rq) \ argument 564 for ((void) iter, rt_rq = &rq->rt; rt_rq; rt_rq = NULL) 637 struct rt_rq *iter = sched_rt_period_rt_rq(rt_b, i); in do_balance_runtime() local 640 if (iter == rt_rq) in do_balance_runtime() 643 raw_spin_lock(&iter->rt_runtime_lock); in do_balance_runtime() 649 if (iter->rt_runtime == RUNTIME_INF) in do_balance_runtime() [all …]
|
/linux-4.1.27/drivers/scsi/qla2xxx/ |
D | qla_inline.h | 75 uint32_t iter = bsize >> 2; in host_to_fcp_swap() local 77 for (; iter ; iter--) in host_to_fcp_swap() 88 uint32_t iter = bsize >> 2; in host_to_adap() local 90 for (; iter ; iter--) in host_to_adap()
|
D | qla_attr.c | 270 uint32_t *iter; in qla2x00_sysfs_write_nvram() local 273 iter = (uint32_t *)buf; in qla2x00_sysfs_write_nvram() 276 chksum += le32_to_cpu(*iter++); in qla2x00_sysfs_write_nvram() 278 *iter = cpu_to_le32(chksum); in qla2x00_sysfs_write_nvram() 280 uint8_t *iter; in qla2x00_sysfs_write_nvram() local 283 iter = (uint8_t *)buf; in qla2x00_sysfs_write_nvram() 286 chksum += *iter++; in qla2x00_sysfs_write_nvram() 288 *iter = chksum; in qla2x00_sysfs_write_nvram() 639 uint16_t iter, addr, offset; in qla2x00_sysfs_read_sfp() local 659 for (iter = 0, offset = 0; iter < (SFP_DEV_SIZE * 2) / SFP_BLOCK_SIZE; in qla2x00_sysfs_read_sfp() [all …]
|
D | qla_sup.c | 2665 uint32_t istart, iend, iter, vend; in qla2x00_get_fcode_version() local 2678 iter = istart; in qla2x00_get_fcode_version() 2679 while ((iter < iend) && !do_next) { in qla2x00_get_fcode_version() 2680 iter++; in qla2x00_get_fcode_version() 2681 if (qla2x00_read_flash_byte(ha, iter) == '/') { in qla2x00_get_fcode_version() 2682 if (qla2x00_read_flash_byte(ha, iter + 2) == in qla2x00_get_fcode_version() 2686 iter + 3) == '/') in qla2x00_get_fcode_version() 2695 while ((iter > istart) && !do_next) { in qla2x00_get_fcode_version() 2696 iter--; in qla2x00_get_fcode_version() 2697 if (qla2x00_read_flash_byte(ha, iter) == ' ') in qla2x00_get_fcode_version() [all …]
|
/linux-4.1.27/drivers/scsi/ |
D | sd_dif.c | 127 struct bvec_iter iter; in sd_dif_prepare() local 136 bip_for_each_vec(iv, bip, iter) { in sd_dif_prepare() 179 struct bvec_iter iter; in sd_dif_complete() local 183 bip_for_each_vec(iv, bip, iter) { in sd_dif_complete()
|
/linux-4.1.27/drivers/infiniband/hw/ehca/ |
D | ehca_cq.c | 75 struct hlist_node *iter; in ehca_cq_unassign_qp() local 80 hlist_for_each(iter, &cq->qp_hashtab[key]) { in ehca_cq_unassign_qp() 81 qp = hlist_entry(iter, struct ehca_qp, list_entries); in ehca_cq_unassign_qp() 83 hlist_del(iter); in ehca_cq_unassign_qp() 104 struct hlist_node *iter; in ehca_cq_get_qp() local 106 hlist_for_each(iter, &cq->qp_hashtab[key]) { in ehca_cq_get_qp() 107 qp = hlist_entry(iter, struct ehca_qp, list_entries); in ehca_cq_get_qp()
|
/linux-4.1.27/scripts/kconfig/ |
D | zconf.lex.c | 2355 struct file *iter; in zconf_nextfile() local 2371 for (iter = current_file->parent; iter; iter = iter->parent ) { in zconf_nextfile() 2372 if (!strcmp(current_file->name,iter->name) ) { in zconf_nextfile() 2377 iter = current_file->parent; in zconf_nextfile() 2378 while (iter && \ in zconf_nextfile() 2379 strcmp(iter->name,current_file->name)) { in zconf_nextfile() 2381 iter->name, iter->lineno-1); in zconf_nextfile() 2382 iter = iter->parent; in zconf_nextfile() 2384 if (iter) in zconf_nextfile() 2386 iter->name, iter->lineno+1); in zconf_nextfile()
|
D | gconf.c | 784 GtkTreeIter iter; in renderer_edited() local 789 if (!gtk_tree_model_get_iter(model2, &iter, path)) in renderer_edited() 792 gtk_tree_model_get(model2, &iter, COL_MENU, &menu, -1); in renderer_edited() 795 gtk_tree_model_get(model2, &iter, COL_VALUE, &old_def, -1); in renderer_edited() 886 GtkTreeIter iter; in on_treeview2_button_press_event() local 903 if (!gtk_tree_model_get_iter(model2, &iter, path)) in on_treeview2_button_press_event() 905 gtk_tree_model_get(model2, &iter, COL_MENU, &menu, -1); in on_treeview2_button_press_event() 943 GtkTreeIter iter; in on_treeview2_key_press_event() local 963 gtk_tree_model_get_iter(model2, &iter, path); in on_treeview2_key_press_event() 964 gtk_tree_model_get(model2, &iter, COL_MENU, &menu, -1); in on_treeview2_key_press_event() [all …]
|
/linux-4.1.27/kernel/events/ |
D | hw_breakpoint.c | 120 struct perf_event *iter; in task_bp_pinned() local 123 list_for_each_entry(iter, &bp_task_head, hw.bp_list) { in task_bp_pinned() 124 if (iter->hw.target == tsk && in task_bp_pinned() 125 find_slot_idx(iter) == type && in task_bp_pinned() 126 (iter->cpu < 0 || cpu == iter->cpu)) in task_bp_pinned() 127 count += hw_breakpoint_weight(iter); in task_bp_pinned()
|
/linux-4.1.27/arch/x86/kernel/ |
D | ftrace.c | 541 struct ftrace_rec_iter *iter; in ftrace_replace_code() local 547 for_ftrace_rec_iter(iter) { in ftrace_replace_code() 548 rec = ftrace_rec_iter_record(iter); in ftrace_replace_code() 560 for_ftrace_rec_iter(iter) { in ftrace_replace_code() 561 rec = ftrace_rec_iter_record(iter); in ftrace_replace_code() 572 for_ftrace_rec_iter(iter) { in ftrace_replace_code() 573 rec = ftrace_rec_iter_record(iter); in ftrace_replace_code() 587 for_ftrace_rec_iter(iter) { in ftrace_replace_code() 588 rec = ftrace_rec_iter_record(iter); in ftrace_replace_code()
|
/linux-4.1.27/kernel/locking/ |
D | lockdep_proc.c | 569 struct lock_stat_data *iter; in ls_start() local 574 iter = data->stats + (*pos - 1); in ls_start() 575 if (iter >= data->iter_end) in ls_start() 576 iter = NULL; in ls_start() 578 return iter; in ls_start() 619 struct lock_stat_data *iter = data->stats; in lock_stat_open() local 623 iter->class = class; in lock_stat_open() 624 iter->stats = lock_stats(class); in lock_stat_open() 625 iter++; in lock_stat_open() 627 data->iter_end = iter; in lock_stat_open()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | t4vf_common.h | 228 #define for_each_port(adapter, iter) \ argument 229 for (iter = 0; iter < (adapter)->params.nports; iter++)
|
D | adapter.h | 348 #define for_each_ethrxq(sge, iter) \ argument 349 for (iter = 0; iter < (sge)->ethqsets; iter++)
|
/linux-4.1.27/drivers/gpu/drm/i915/ |
D | i915_gem_gtt.h | 344 #define gen6_for_each_pde(pt, pd, start, length, temp, iter) \ argument 345 for (iter = gen6_pde_index(start); \ 346 pt = (pd)->page_table[iter], length > 0 && iter < I915_PDES; \ 347 iter++, \
|
/linux-4.1.27/drivers/net/ethernet/qlogic/qlge/ |
D | qlge_ethtool.c | 247 u64 *iter = &qdev->nic_stats.tx_pkts; in ql_update_stats() local 265 *iter = data; in ql_update_stats() 266 iter++; in ql_update_stats() 279 *iter = data; in ql_update_stats() 280 iter++; in ql_update_stats() 284 iter += QLGE_RCV_MAC_ERR_STATS; in ql_update_stats() 296 *iter = data; in ql_update_stats() 297 iter++; in ql_update_stats() 310 *iter = data; in ql_update_stats() 311 iter++; in ql_update_stats() [all …]
|
/linux-4.1.27/fs/f2fs/ |
D | trace.c | 129 struct radix_tree_iter iter; in gang_lookup_pids() local 136 radix_tree_for_each_slot(slot, &pids, &iter, first_index) { in gang_lookup_pids() 137 results[ret] = iter.index; in gang_lookup_pids()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb/ |
D | common.h | 314 #define for_each_port(adapter, iter) \ argument 315 for (iter = 0; iter < (adapter)->params.nports; ++iter)
|
/linux-4.1.27/fs/ocfs2/dlm/ |
D | dlmmaster.c | 743 struct dlm_node_iter iter; in dlm_get_lock_resource() local 956 dlm_node_iter_init(mle->vote_map, &iter); in dlm_get_lock_resource() 957 while ((nodenum = dlm_node_iter_next(&iter)) >= 0) { in dlm_get_lock_resource() 1191 static void dlm_bitmap_diff_iter_init(struct dlm_bitmap_diff_iter *iter, in dlm_bitmap_diff_iter_init() argument 1198 iter->curnode = -1; in dlm_bitmap_diff_iter_init() 1199 iter->orig_bm = orig_bm; in dlm_bitmap_diff_iter_init() 1200 iter->cur_bm = cur_bm; in dlm_bitmap_diff_iter_init() 1203 p1 = *(iter->orig_bm + i); in dlm_bitmap_diff_iter_init() 1204 p2 = *(iter->cur_bm + i); in dlm_bitmap_diff_iter_init() 1205 iter->diff_bm[i] = (p1 & ~p2) | (p2 & ~p1); in dlm_bitmap_diff_iter_init() [all …]
|
D | dlmcommon.h | 1113 struct dlm_node_iter *iter) in dlm_node_iter_init() argument 1115 memcpy(iter->node_map, map, sizeof(iter->node_map)); in dlm_node_iter_init() 1116 iter->curnode = -1; in dlm_node_iter_init() 1119 static inline int dlm_node_iter_next(struct dlm_node_iter *iter) in dlm_node_iter_next() argument 1122 bit = find_next_bit(iter->node_map, O2NM_MAX_NODES, iter->curnode+1); in dlm_node_iter_next() 1124 iter->curnode = O2NM_MAX_NODES; in dlm_node_iter_next() 1127 iter->curnode = bit; in dlm_node_iter_next()
|
/linux-4.1.27/tools/perf/ |
D | builtin-report.c | 89 static int hist_iter__report_callback(struct hist_entry_iter *iter, in hist_iter__report_callback() argument 95 struct hist_entry *he = iter->he; in hist_iter__report_callback() 96 struct perf_evsel *evsel = iter->evsel; in hist_iter__report_callback() 139 struct hist_entry_iter iter = { in process_sample_event() local 158 iter.ops = &hist_iter_branch; in process_sample_event() 160 iter.ops = &hist_iter_mem; in process_sample_event() 162 iter.ops = &hist_iter_cumulative; in process_sample_event() 164 iter.ops = &hist_iter_normal; in process_sample_event() 169 ret = hist_entry_iter__add(&iter, &al, evsel, sample, rep->max_stack, in process_sample_event()
|
/linux-4.1.27/fs/9p/ |
D | vfs_addr.c | 248 v9fs_direct_IO(struct kiocb *iocb, struct iov_iter *iter, loff_t pos) in v9fs_direct_IO() argument 253 if (iov_iter_rw(iter) == WRITE) { in v9fs_direct_IO() 254 n = p9_client_write(file->private_data, pos, iter, &err); in v9fs_direct_IO() 262 n = p9_client_read(file->private_data, pos, iter, &err); in v9fs_direct_IO()
|
/linux-4.1.27/arch/x86/kvm/ |
D | mmu.c | 1078 static u64 *rmap_get_first(unsigned long rmap, struct rmap_iterator *iter) in rmap_get_first() argument 1084 iter->desc = NULL; in rmap_get_first() 1088 iter->desc = (struct pte_list_desc *)(rmap & ~1ul); in rmap_get_first() 1089 iter->pos = 0; in rmap_get_first() 1090 return iter->desc->sptes[iter->pos]; in rmap_get_first() 1098 static u64 *rmap_get_next(struct rmap_iterator *iter) in rmap_get_next() argument 1100 if (iter->desc) { in rmap_get_next() 1101 if (iter->pos < PTE_LIST_EXT - 1) { in rmap_get_next() 1104 ++iter->pos; in rmap_get_next() 1105 sptep = iter->desc->sptes[iter->pos]; in rmap_get_next() [all …]
|
D | mmu_audit.c | 193 struct rmap_iterator iter; in audit_write_protection() local 200 for (sptep = rmap_get_first(*rmapp, &iter); sptep; in audit_write_protection() 201 sptep = rmap_get_next(&iter)) { in audit_write_protection()
|
/linux-4.1.27/include/trace/ |
D | ftrace.h | 307 ftrace_raw_output_##call(struct trace_iterator *iter, int flags, \ 310 struct trace_seq *s = &iter->seq; \ 311 struct trace_seq __maybe_unused *p = &iter->tmp_seq; \ 315 field = (typeof(field))iter->ent; \ 317 ret = ftrace_raw_output_prep(iter, trace_event); \ 332 ftrace_raw_output_##call(struct trace_iterator *iter, int flags, \ 337 struct trace_seq *p = &iter->tmp_seq; \ 339 entry = iter->ent; \ 349 return ftrace_output_call(iter, #call, print); \
|
/linux-4.1.27/net/core/ |
D | dev.c | 1461 struct list_head *iter; in dev_disable_lro() local 1469 netdev_for_each_lower_dev(dev, lower_dev, iter) in dev_disable_lro() 4829 struct list_head **iter) in netdev_upper_get_next_dev_rcu() argument 4835 upper = list_entry_rcu((*iter)->next, struct netdev_adjacent, list); in netdev_upper_get_next_dev_rcu() 4840 *iter = &upper->list; in netdev_upper_get_next_dev_rcu() 4855 struct list_head **iter) in netdev_all_upper_get_next_dev_rcu() argument 4861 upper = list_entry_rcu((*iter)->next, struct netdev_adjacent, list); in netdev_all_upper_get_next_dev_rcu() 4866 *iter = &upper->list; in netdev_all_upper_get_next_dev_rcu() 4884 struct list_head **iter) in netdev_lower_get_next_private() argument 4888 lower = list_entry(*iter, struct netdev_adjacent, list); in netdev_lower_get_next_private() [all …]
|
/linux-4.1.27/arch/sh/mm/ |
D | pmb.c | 145 struct pmb_entry *pmbe, *iter; in pmb_mapping_exists() local 175 for (iter = pmbe->link; iter; iter = iter->link) in pmb_mapping_exists() 176 span += iter->size; in pmb_mapping_exists() 815 static int pmb_seq_show(struct seq_file *file, void *iter) in pmb_seq_show() argument
|
D | asids-debugfs.c | 27 static int asids_seq_show(struct seq_file *file, void *iter) in asids_seq_show() argument
|
/linux-4.1.27/fs/jfs/ |
D | inode.c | 333 static ssize_t jfs_direct_IO(struct kiocb *iocb, struct iov_iter *iter, in jfs_direct_IO() argument 339 size_t count = iov_iter_count(iter); in jfs_direct_IO() 342 ret = blockdev_direct_IO(iocb, inode, iter, offset, jfs_get_block); in jfs_direct_IO() 348 if (unlikely(iov_iter_rw(iter) == WRITE && ret < 0)) { in jfs_direct_IO()
|
/linux-4.1.27/security/selinux/ss/ |
D | ebitmap.c | 94 unsigned int iter; in ebitmap_netlbl_export() local 108 for (iter = 0; iter < EBITMAP_UNIT_NUMS; iter++) { in ebitmap_netlbl_export() 109 e_map = e_iter->maps[iter]; in ebitmap_netlbl_export()
|
/linux-4.1.27/fs/proc/ |
D | base.c | 2816 static struct tgid_iter next_tgid(struct pid_namespace *ns, struct tgid_iter iter) in next_tgid() argument 2820 if (iter.task) in next_tgid() 2821 put_task_struct(iter.task); in next_tgid() 2824 iter.task = NULL; in next_tgid() 2825 pid = find_ge_pid(iter.tgid, ns); in next_tgid() 2827 iter.tgid = pid_nr_ns(pid, ns); in next_tgid() 2828 iter.task = pid_task(pid, PIDTYPE_PID); in next_tgid() 2841 if (!iter.task || !has_group_leader_pid(iter.task)) { in next_tgid() 2842 iter.tgid += 1; in next_tgid() 2845 get_task_struct(iter.task); in next_tgid() [all …]
|
/linux-4.1.27/drivers/block/ |
D | ps3disk.c | 96 struct req_iterator iter; in ps3disk_scatter_gather() local 102 rq_for_each_segment(bvec, req, iter) { in ps3disk_scatter_gather() 105 __func__, __LINE__, i, bio_sectors(iter.bio), in ps3disk_scatter_gather() 106 iter.bio->bi_iter.bi_sector); in ps3disk_scatter_gather() 133 struct req_iterator iter; in ps3disk_submit_request_sg() local 135 rq_for_each_segment(bv, req, iter) in ps3disk_submit_request_sg()
|
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/engine/pm/ |
D | base.c | 123 args->v0.version, args->v0.iter); in nvkm_perfctr_query() 124 di = (args->v0.iter & 0xff000000) >> 24; in nvkm_perfctr_query() 125 si = (args->v0.iter & 0x00ffffff) - 1; in nvkm_perfctr_query() 151 args->v0.iter = (di << 24) | ++si; in nvkm_perfctr_query() 160 args->v0.iter = 0xffffffff; in nvkm_perfctr_query()
|
/linux-4.1.27/tools/perf/tests/ |
D | hists_cumulate.c | 89 struct hist_entry_iter iter = { in add_hist_entries() local 94 iter.ops = &hist_iter_cumulative; in add_hist_entries() 96 iter.ops = &hist_iter_normal; in add_hist_entries() 107 if (hist_entry_iter__add(&iter, &al, evsel, &sample, in add_hist_entries()
|
/linux-4.1.27/drivers/iommu/ |
D | dmar.c | 547 struct acpi_dmar_header *iter, *next; in dmar_walk_remapping_entries() local 550 for (iter = start; iter < end && ret == 0; iter = next) { in dmar_walk_remapping_entries() 551 next = (void *)iter + iter->length; in dmar_walk_remapping_entries() 552 if (iter->length == 0) { in dmar_walk_remapping_entries() 564 dmar_table_print_dmar_entry(iter); in dmar_walk_remapping_entries() 566 if (iter->type >= ACPI_DMAR_TYPE_RESERVED) { in dmar_walk_remapping_entries() 569 iter->type); in dmar_walk_remapping_entries() 570 } else if (cb->cb[iter->type]) { in dmar_walk_remapping_entries() 571 ret = cb->cb[iter->type](iter, cb->arg[iter->type]); in dmar_walk_remapping_entries() 574 iter->type); in dmar_walk_remapping_entries()
|
/linux-4.1.27/fs/btrfs/tests/ |
D | btrfs-tests.c | 130 struct radix_tree_iter iter; in btrfs_free_dummy_fs_info() local 135 radix_tree_for_each_slot(slot, &fs_info->buffer_radix, &iter, 0) { in btrfs_free_dummy_fs_info()
|
/linux-4.1.27/scripts/coccinelle/free/ |
D | kfree.cocci | 72 iterator iter; 80 iter(...,subE,...) S // no use
|
/linux-4.1.27/drivers/target/ |
D | target_core_file.c | 326 struct iov_iter iter; in fd_do_rw() local 346 iov_iter_bvec(&iter, ITER_BVEC, bvec, sgl_nents, len); in fd_do_rw() 348 ret = vfs_iter_write(fd, &iter, &pos); in fd_do_rw() 350 ret = vfs_iter_read(fd, &iter, &pos); in fd_do_rw() 435 struct iov_iter iter; in fd_execute_write_same() local 472 iov_iter_bvec(&iter, ITER_BVEC, bvec, nolb, len); in fd_execute_write_same() 473 ret = vfs_iter_write(fd_dev->fd_file, &iter, &pos); in fd_execute_write_same()
|
/linux-4.1.27/fs/ext4/ |
D | indirect.c | 645 ssize_t ext4_ind_direct_IO(struct kiocb *iocb, struct iov_iter *iter, in ext4_ind_direct_IO() argument 654 size_t count = iov_iter_count(iter); in ext4_ind_direct_IO() 657 if (iov_iter_rw(iter) == WRITE) { in ext4_ind_direct_IO() 679 if (iov_iter_rw(iter) == READ && ext4_should_dioread_nolock(inode)) { in ext4_ind_direct_IO() 693 ret = dax_do_io(iocb, inode, iter, offset, in ext4_ind_direct_IO() 697 inode->i_sb->s_bdev, iter, in ext4_ind_direct_IO() 704 ret = dax_do_io(iocb, inode, iter, offset, in ext4_ind_direct_IO() 707 ret = blockdev_direct_IO(iocb, inode, iter, offset, in ext4_ind_direct_IO() 710 if (unlikely(iov_iter_rw(iter) == WRITE && ret < 0)) { in ext4_ind_direct_IO()
|
/linux-4.1.27/drivers/net/ |
D | macvtap.c | 800 struct iov_iter *iter) in macvtap_put_user() argument 810 if (iov_iter_count(iter) < vnet_hdr_len) in macvtap_put_user() 815 if (copy_to_iter(&vnet_hdr, sizeof(vnet_hdr), iter) != in macvtap_put_user() 819 iov_iter_advance(iter, vnet_hdr_len - sizeof(vnet_hdr)); in macvtap_put_user() 835 ret = skb_copy_datagram_iter(skb, 0, iter, vlan_offset); in macvtap_put_user() 836 if (ret || !iov_iter_count(iter)) in macvtap_put_user() 839 ret = copy_to_iter(&veth, sizeof(veth), iter); in macvtap_put_user() 840 if (ret != sizeof(veth) || !iov_iter_count(iter)) in macvtap_put_user() 844 ret = skb_copy_datagram_iter(skb, vlan_offset, iter, in macvtap_put_user()
|
/linux-4.1.27/drivers/iio/ |
D | industrialio-trigger.c | 104 struct iio_trigger *trig = NULL, *iter; in iio_trigger_find_by_name() local 107 list_for_each_entry(iter, &iio_trigger_list, list) in iio_trigger_find_by_name() 108 if (sysfs_streq(iter->name, name)) { in iio_trigger_find_by_name() 109 trig = iter; in iio_trigger_find_by_name()
|
/linux-4.1.27/drivers/char/ |
D | mem.c | 634 static ssize_t read_iter_zero(struct kiocb *iocb, struct iov_iter *iter) in read_iter_zero() argument 638 while (iov_iter_count(iter)) { in read_iter_zero() 639 size_t chunk = iov_iter_count(iter), n; in read_iter_zero() 643 n = iov_iter_zero(chunk, iter); in read_iter_zero() 644 if (!n && iov_iter_count(iter)) in read_iter_zero()
|
/linux-4.1.27/drivers/phy/ |
D | phy-core.c | 465 struct class_dev_iter iter; in of_phy_simple_xlate() local 467 class_dev_iter_init(&iter, phy_class, NULL, NULL); in of_phy_simple_xlate() 468 while ((dev = class_dev_iter_next(&iter))) { in of_phy_simple_xlate() 473 class_dev_iter_exit(&iter); in of_phy_simple_xlate() 477 class_dev_iter_exit(&iter); in of_phy_simple_xlate()
|
/linux-4.1.27/arch/x86/kernel/cpu/ |
D | perf_event_intel_cqm.c | 830 struct perf_event *iter; in intel_cqm_setup_event() local 834 list_for_each_entry(iter, &cache_groups, hw.cqm_groups_entry) { in intel_cqm_setup_event() 835 rmid = iter->hw.cqm_rmid; in intel_cqm_setup_event() 837 if (__match_event(iter, event)) { in intel_cqm_setup_event() 840 *group = iter; in intel_cqm_setup_event() 848 if (__conflict_event(iter, event) && __rmid_valid(rmid)) in intel_cqm_setup_event()
|
/linux-4.1.27/drivers/cpufreq/ |
D | acpi-cpufreq.c | 956 struct freq_attr **iter; in acpi_cpufreq_init() local 960 for (iter = acpi_cpufreq_attr; *iter != NULL; iter++) in acpi_cpufreq_init() 964 if (iter[1] == NULL) in acpi_cpufreq_init() 965 *iter = &cpb; in acpi_cpufreq_init()
|
/linux-4.1.27/arch/m68k/emu/ |
D | nfblock.c | 66 struct bvec_iter iter; in nfhd_make_request() local 72 bio_for_each_segment(bvec, bio, iter) { in nfhd_make_request()
|
/linux-4.1.27/arch/xtensa/platforms/iss/ |
D | simdisk.c | 107 struct bvec_iter iter; in simdisk_xfer_bio() local 110 bio_for_each_segment(bvec, bio, iter) { in simdisk_xfer_bio() 111 char *buffer = __bio_kmap_atomic(bio, iter); in simdisk_xfer_bio()
|