Home
last modified time | relevance | path

Searched refs:rb_first (Results 1 – 140 of 140) sorted by relevance

/linux-4.4.14/net/ceph/
Ddebugfs.c68 for (n = rb_first(&map->pg_pools); n; n = rb_next(n)) { in osdmap_show()
87 for (n = rb_first(&map->pg_temp); n; n = rb_next(n)) { in osdmap_show()
98 for (n = rb_first(&map->primary_temp); n; n = rb_next(n)) { in osdmap_show()
125 for (rp = rb_first(&monc->generic_request_tree); rp; rp = rb_next(rp)) { in monc_show()
148 for (p = rb_first(&osdc->requests); p; p = rb_next(p)) { in osdc_show()
Dosdmap.c522 for (rbp = rb_first(&map->pg_pools); rbp; rbp = rb_next(rbp)) { in ceph_pg_poolid_by_name()
661 rb_entry(rb_first(&map->pg_temp), in ceph_osdmap_destroy()
668 rb_entry(rb_first(&map->primary_temp), in ceph_osdmap_destroy()
675 rb_entry(rb_first(&map->pg_pools), in ceph_osdmap_destroy()
Dauth_x.c658 while ((p = rb_first(&xi->ticket_handlers)) != NULL) { in ceph_x_destroy()
Dosd_client.c1093 struct ceph_osd *osd = rb_entry(rb_first(&osdc->osds), in remove_all_osds()
1952 for (p = rb_first(&osdc->osds); p; p = n) { in reset_changed_osds()
1983 for (p = rb_first(&osdc->requests); p; ) { in kick_requests()
Dmon_client.c731 for (p = rb_first(&monc->generic_request_tree); p; p = rb_next(p)) { in __resend_generic_request()
/linux-4.4.14/tools/perf/util/
Drblist.c107 struct rb_node *pos, *next = rb_first(&rblist->entries); in rblist__delete()
122 for (node = rb_first(&rblist->entries); node; node = rb_next(node)) { in rblist__entry()
Dcallchain.c250 n = rb_first(&node->rb_root_in); in __sort_chain_flat()
280 n = rb_first(&node->rb_root_in); in __sort_chain_graph_abs()
311 n = rb_first(&node->rb_root_in); in __sort_chain_graph_rel()
374 n = rb_first(&new->rb_root_in); in create_child()
660 n = rb_first(&src->rb_root_in); in merge_chain_branch()
813 n = rb_first(&node->rb_root_in); in free_callchain_node()
Dhist.c184 struct rb_node *next = rb_first(&hists->entries); in hists__output_recalc_col_len()
283 struct rb_node *next = rb_first(&hists->entries); in hists__decay_entries()
299 struct rb_node *next = rb_first(&hists->entries); in hists__delete_entries()
1063 next = rb_first(root); in hists__collapse_resort()
1179 next = rb_first(root); in hists__output_resort()
1240 for (nd = rb_first(&hists->entries); nd; nd = rb_next(nd)) { in hists__filter_by_dso()
1274 for (nd = rb_first(&hists->entries); nd; nd = rb_next(nd)) { in hists__filter_by_thread()
1306 for (nd = rb_first(&hists->entries); nd; nd = rb_next(nd)) { in hists__filter_by_symbol()
1337 for (nd = rb_first(&hists->entries); nd; nd = rb_next(nd)) { in hists__filter_by_socket()
1448 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in hists__match()
[all …]
Dbuild-id.c247 for (nd = rb_first(&session->machines.guests); nd; nd = rb_next(nd)) { in perf_session__write_buildid_table()
280 for (nd = rb_first(&session->machines.guests); nd; nd = rb_next(nd)) { in dsos__hit_all()
524 for (nd = rb_first(&session->machines.guests); nd; nd = rb_next(nd)) { in perf_session__cache_build_ids()
541 for (nd = rb_first(&session->machines.guests); nd; nd = rb_next(nd)) { in perf_session__read_build_ids()
Dmap.c262 struct rb_node *nd = rb_first(symbols); in map__fixup_start()
479 struct rb_node *next = rb_first(root); in __maps__purge()
564 for (nd = rb_first(&maps->entries); nd; nd = rb_next(nd)) { in maps__find_symbol_by_name()
617 for (nd = rb_first(&maps->entries); nd; nd = rb_next(nd)) { in maps__fprintf()
662 next = rb_first(root); in maps__fixup_overlappings()
821 struct rb_node *first = rb_first(&maps->entries); in maps__first()
Dsymbol.c154 nd = rb_first(symbols); in symbols__fixup_duplicate()
182 struct rb_node *nd, *prevnd = rb_first(symbols); in symbols__fixup_end()
294 struct rb_node *next = rb_first(symbols); in symbols__delete()
348 struct rb_node *n = rb_first(symbols); in symbols__first()
396 for (nd = rb_first(source); nd; nd = rb_next(nd)) { in symbols__sort_by_name()
506 for (nd = rb_first(&dso->symbol_names[type]); nd; nd = rb_next(nd)) { in dso__fprintf_symbols_by_name()
660 struct rb_node *next = rb_first(root); in dso__split_kallsyms_for_kcore()
713 struct rb_node *next = rb_first(root); in dso__split_kallsyms()
882 struct rb_node *next = rb_first(modules); in delete_modules()
964 from_node = rb_first(&from_modules); in compare_proc_modules()
[all …]
Dstrlist.h52 struct rb_node *rn = rb_first(&slist->rblist.entries); in strlist__first()
Dintlist.h47 struct rb_node *rn = rb_first(&ilist->rblist.entries); in intlist__first()
Dmachine.c113 nd = rb_first(&machine->threads); in machine__delete_threads()
192 for (nd = rb_first(&machines->guests); nd; nd = rb_next(nd)) { in machines__set_symbol_filter()
205 for (nd = rb_first(&machines->guests); nd; nd = rb_next(nd)) { in machines__set_comm_exec()
277 for (nd = rb_first(&machines->guests); nd; nd = rb_next(nd)) { in machines__process_guests()
304 for (node = rb_first(&machines->guests); node; node = rb_next(node)) { in machines__set_id_hdr_size()
598 for (nd = rb_first(&machines->guests); nd; nd = rb_next(nd)) { in machines__fprintf_dsos()
618 for (nd = rb_first(&machines->guests); nd; nd = rb_next(nd)) { in machines__fprintf_dsos_buildid()
651 for (nd = rb_first(&machine->threads); nd; nd = rb_next(nd)) { in machine__fprintf()
846 struct rb_node *next = rb_first(&machines->guests); in machines__destroy_kernel_maps()
1906 for (nd = rb_first(&machine->threads); nd; nd = rb_next(nd)) { in machine__for_each_thread()
[all …]
Dsymbol.h71 for (nd = rb_first(symbols); \
Ddso.c558 struct rb_node *next = rb_first(root); in dso_cache__free()
1314 for (nd = rb_first(&dso->symbols[type]); nd; nd = rb_next(nd)) { in dso__fprintf()
Dannotate.c1320 node = rb_first(src_root); in resort_source_line()
1427 node = rb_first(root); in print_summary()
/linux-4.4.14/fs/proc/
Dtask_nommu.c26 for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) { in task_mem()
87 for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) { in task_vsize()
105 for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) { in task_statm()
244 for (p = rb_first(&mm->mm_rb); p; p = rb_next(p)) in m_start()
Dnommu.c92 for (p = rb_first(&nommu_region_tree); p; p = rb_next(p)) in nommu_region_list_start()
Dgeneric.c43 return rb_entry_safe(rb_first(&dir->subdir), struct proc_dir_entry, in pde_subdir_first()
Dproc_sysctl.c367 ctl_node = first_usable_entry(rb_first(&dir->root)); in first_entry()
/linux-4.4.14/drivers/base/regmap/
Dregcache-rbtree.c153 for (node = rb_first(&rbtree_ctx->root); node != NULL; in rbtree_show()
240 next = rb_first(&rbtree_ctx->root); in regcache_rbtree_exit()
416 for (node = rb_first(&rbtree_ctx->root); node; in regcache_rbtree_write()
467 for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) { in regcache_rbtree_sync()
507 for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) { in regcache_rbtree_drop()
Dregmap-debugfs.c620 next = rb_first(&map->range_tree); in regmap_debugfs_init()
Dregmap.c423 next = rb_first(&map->range_tree); in regmap_range_exit()
/linux-4.4.14/tools/perf/tests/
Dhists_output.c108 node = rb_first(root_out); in del_hist_entries()
167 node = rb_first(root); in test1()
267 node = rb_first(root); in test2()
321 node = rb_first(root); in test3()
399 node = rb_first(root); in test4()
502 node = rb_first(root); in test5()
Dhists_common.c169 node = rb_first(root); in print_hists_in()
196 node = rb_first(root); in print_hists_out()
Dhists_cumulate.c142 node = rb_first(root_out); in del_hist_entries()
204 for (node = rb_first(root), i = 0; in do_test()
224 cnode = rb_entry(rb_first(root), struct callchain_node, rb_node); in do_test()
Dhists_link.c165 node = rb_first(root); in __validate_match()
217 node = rb_first(root); in __validate_link()
Dvmlinux-kallsyms.c111 for (nd = rb_first(&vmlinux_map->dso->symbols[type]); nd; nd = rb_next(nd)) { in test__vmlinux_matches_kallsyms()
/linux-4.4.14/drivers/iommu/
Diova.c383 node = rb_first(&iovad->rbroot); in put_iova_domain()
389 node = rb_first(&iovad->rbroot); in put_iova_domain()
461 for (node = rb_first(&iovad->rbroot); node; node = rb_next(node)) { in reserve_iova()
499 for (node = rb_first(&from->rbroot); node; node = rb_next(node)) { in copy_reserved_iova()
Darm-smmu.c1822 for (node = rb_first(&smmu->masters); node; node = rb_next(node)) { in arm_smmu_device_dt_probe()
1851 for (node = rb_first(&smmu->masters); node; node = rb_next(node)) { in arm_smmu_device_remove()
/linux-4.4.14/tools/perf/ui/gtk/
Dhists.c96 bool has_single_node = (rb_first(root) == rb_last(root)); in perf_gtk__add_callchain()
98 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in perf_gtk__add_callchain()
223 for (nd = rb_first(&hists->entries); nd; nd = rb_next(nd)) { in perf_gtk__show_hists()
/linux-4.4.14/tools/perf/ui/stdio/
Dhist.c95 node = rb_first(root); in __callchain__fprintf_graph()
176 node = rb_first(root); in callchain__fprintf_graph()
243 struct rb_node *rb_node = rb_first(tree); in callchain__fprintf_flat()
461 for (nd = rb_first(&hists->entries); nd; nd = rb_next(nd)) { in hists__fprintf()
/linux-4.4.14/tools/perf/ui/browsers/
Dhists.c59 for (nd = rb_first(&hists->entries); in hist_browser__get_folding()
161 for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) { in callchain_node__count_rows_rb_tree()
203 for (nd = rb_first(chain); nd; nd = rb_next(nd)) { in callchain__count_rows()
237 struct rb_node *nd = rb_first(&node->rb_root); in callchain_node__init_have_children_rb_tree()
239 for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) { in callchain_node__init_have_children_rb_tree()
276 struct rb_node *nd = rb_first(root); in callchain__init_have_children()
279 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in callchain__init_have_children()
334 for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) { in callchain_node__set_folding_rb_tree()
375 for (nd = rb_first(chain); nd; nd = rb_next(nd)) { in callchain__set_folding()
401 for (nd = rb_first(&hists->entries); in __hist_browser__set_folding()
[all …]
Dmap.c115 for (nd = rb_first(mb.b.entries); nd; nd = rb_next(nd)) { in map__browse()
Dannotate.c761 nd = rb_first(&browser->entries); in annotate_browser__run()
/linux-4.4.14/arch/powerpc/kernel/
Deeh_cache.c113 n = rb_first(&cache->rb_root); in eeh_addr_cache_print()
239 n = rb_first(&pci_io_addr_cache_root.rb_root); in __eeh_addr_cache_rmv_dev()
/linux-4.4.14/fs/ocfs2/
Dreservations.c96 node = rb_first(&resmap->m_reservations); in ocfs2_dump_resv()
151 node = rb_first(&resmap->m_reservations); in ocfs2_check_resmap()
378 node = rb_first(&resmap->m_reservations); in ocfs2_find_resv_lhs()
537 next = rb_first(root); in __ocfs2_resv_find_window()
/linux-4.4.14/tools/include/linux/
Drbtree.h68 extern struct rb_node *rb_first(const struct rb_root *);
/linux-4.4.14/kernel/trace/
Dtrace_stat.c189 node = rb_first(&session->stat_root); in stat_seq_start()
204 return rb_first(&session->stat_root); in stat_seq_next()
/linux-4.4.14/lib/
Drbtree_test.c143 for (rb = rb_first(&root); rb; rb = rb_next(rb)) { in check()
169 for (rb = rb_first(&root); rb; rb = rb_next(rb)) { in check_augmented()
Drbtree.c450 struct rb_node *rb_first(const struct rb_root *root) in rb_first() function
461 EXPORT_SYMBOL(rb_first);
/linux-4.4.14/include/linux/
Drbtree.h69 extern struct rb_node *rb_first(const struct rb_root *);
/linux-4.4.14/drivers/infiniband/hw/mlx4/
Dcm.c439 while (rb_first(sl_id_map)) { in mlx4_ib_cm_paravirt_clean()
441 rb_entry(rb_first(sl_id_map), in mlx4_ib_cm_paravirt_clean()
450 nd = rb_first(sl_id_map); in mlx4_ib_cm_paravirt_clean()
Dmcg.c1091 for (p = rb_first(&ctx->mcg_table); p; p = rb_next(p)) in _mlx4_ib_mcg_port_cleanup()
1105 while ((p = rb_first(&ctx->mcg_table)) != NULL) { in _mlx4_ib_mcg_port_cleanup()
1236 for (p = rb_first(&ctx->mcg_table); p; p = rb_next(p)) { in clean_vf_mcast()
/linux-4.4.14/fs/logfs/
Dgc.c241 cand = rb_entry(rb_first(&list->rb_tree), struct gc_candidate, rb_node); in get_best_cand()
352 return rb_entry(rb_first(&list->rb_tree), struct gc_candidate, rb_node); in first_in_list()
597 cand = rb_entry(rb_first(&super->s_free_list.rb_tree), in logfs_journal_wl_pass()
/linux-4.4.14/fs/ceph/
Dcaps.c347 for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) { in __ceph_get_cap_mds()
675 for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) { in __ceph_caps_issued()
706 for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) { in __ceph_caps_issued_other()
756 for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) { in __ceph_caps_issued_mask()
782 for (q = rb_first(&ci->i_caps); q != p; in __ceph_caps_issued_mask()
807 for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) { in __ceph_caps_revoking_other()
869 for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) { in __ceph_caps_mds_wanted()
1083 p = rb_first(&ci->i_caps); in ceph_queue_caps_release()
1470 struct rb_node *n = rb_first(&mdsc->cap_flush_tree); in __get_oldest_flush_tid()
1681 for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) { in ceph_check_caps()
[all …]
Ddebugfs.c60 for (rp = rb_first(&mdsc->request_tree); rp; rp = rb_next(rp)) { in mdsc_show()
Dxattr.c518 p = rb_first(&ci->i_xattrs.index); in __copy_xattr_names()
541 p = rb_first(&ci->i_xattrs.index); in __ceph_destroy_xattrs()
687 p = rb_first(&ci->i_xattrs.index); in __ceph_build_xattrs_blob()
Dmds_client.c813 cap = rb_entry(rb_first(&ci->i_caps), struct ceph_cap, ci_node); in __choose_mds()
1059 p = rb_first(&mdsc->request_tree); in cleanup_session_requests()
1160 struct rb_node *n = rb_first(&ci->i_cap_flush_tree); in remove_session_caps_cb()
1510 n = rb_first(&mdsc->cap_flush_tree); in check_caps_flush()
1740 return rb_entry(rb_first(&mdsc->request_tree), in __get_oldest_req()
2251 struct rb_node *p = rb_first(&mdsc->request_tree); in kick_requests()
2759 p = rb_first(&mdsc->request_tree); in replay_unsafe_requests()
2989 for (p = rb_first(&mdsc->snap_realms); p; p = rb_next(p)) { in send_mds_reconnect()
Dinode.c321 rb_node = rb_first(&ci->i_fragtree); in ceph_fill_fragtree()
335 rb_node = rb_first(&ci->i_fragtree); in ceph_fill_fragtree()
509 while ((n = rb_first(&ci->i_fragtree)) != NULL) { in ceph_destroy_inode()
Daddr.c1811 n = rb_first(&mdsc->pool_perm_tree); in ceph_pool_perm_destroy()
/linux-4.4.14/drivers/staging/android/ion/
Dion.c689 for (n = rb_first(&client->handles); n; n = rb_next(n)) { in ion_debug_client_show()
727 for (node = rb_first(root); node; node = rb_next(node)) { in ion_get_client_serial()
835 while ((n = rb_first(&client->handles))) { in ion_client_destroy()
1386 for (n = rb_first(&client->handles); n; n = rb_next(n)) { in ion_debug_heap_total()
1408 for (n = rb_first(&dev->clients); n; n = rb_next(n)) { in ion_debug_heap_show()
1429 for (n = rb_first(&dev->buffers); n; n = rb_next(n)) { in ion_debug_heap_show()
/linux-4.4.14/drivers/android/
Dbinder.c1054 for (n = rb_first(&proc->refs_by_desc); n != NULL; n = rb_next(n)) { in binder_get_ref_for_node()
2993 for (n = rb_first(&proc->threads); n != NULL; n = rb_next(n)) { in binder_deferred_flush()
3084 while ((n = rb_first(&proc->threads))) { in binder_deferred_release()
3094 while ((n = rb_first(&proc->nodes))) { in binder_deferred_release()
3104 while ((n = rb_first(&proc->refs_by_desc))) { in binder_deferred_release()
3116 while ((n = rb_first(&proc->allocated_buffers))) { in binder_deferred_release()
3368 for (n = rb_first(&proc->threads); n != NULL; n = rb_next(n)) in print_binder_proc()
3371 for (n = rb_first(&proc->nodes); n != NULL; n = rb_next(n)) { in print_binder_proc()
3378 for (n = rb_first(&proc->refs_by_desc); in print_binder_proc()
3384 for (n = rb_first(&proc->allocated_buffers); n != NULL; n = rb_next(n)) in print_binder_proc()
[all …]
/linux-4.4.14/fs/dlm/
Ddebug_fs.c457 for (node = rb_first(tree); node; node = rb_next(node)) { in table_seq_start()
489 node = rb_first(tree); in table_seq_start()
554 next = rb_first(tree); in table_seq_next()
Drecover.c906 for (n = rb_first(&ls->ls_rsbtbl[i].keep); n; n = rb_next(n)) { in dlm_create_root_list()
942 for (n = rb_first(&ls->ls_rsbtbl[i].toss); n; n = next) { in dlm_clear_toss()
Dlockspace.c817 while ((n = rb_first(&ls->ls_rsbtbl[i].keep))) { in release_lockspace()
823 while ((n = rb_first(&ls->ls_rsbtbl[i].toss))) { in release_lockspace()
Dlock.c1094 for (n = rb_first(&ls->ls_rsbtbl[i].keep); n; n = rb_next(n)) { in dlm_dump_rsb_hash()
1670 for (n = rb_first(&ls->ls_rsbtbl[b].toss); n; n = next) { in shrink_bucket()
5476 for (n = rb_first(&ls->ls_rsbtbl[bucket].keep); n; n = rb_next(n)) { in find_grant_rsb()
/linux-4.4.14/drivers/mtd/ubi/
Dwl.c317 e = rb_entry(rb_first(root), struct ubi_wl_entry, u.rb); in find_wl_entry()
358 first = rb_entry(rb_first(root), struct ubi_wl_entry, u.rb); in find_mean_wl_entry()
708 e1 = rb_entry(rb_first(&ubi->used), struct ubi_wl_entry, u.rb);
729 e1 = rb_entry(rb_first(&ubi->scrub), struct ubi_wl_entry, u.rb);
979 e1 = rb_entry(rb_first(&ubi->used), struct ubi_wl_entry, u.rb);
Dubi.h979 for (rb = rb_first(root), \
Dfastmap.c424 for (node = rb_first(&ai->volumes); node; node = rb_next(node)) { in unmap_peb()
427 for (node2 = rb_first(&av->root); node2; in unmap_peb()
Dattach.c638 while ((rb = rb_first(&av->root))) { in ubi_remove_av()
/linux-4.4.14/fs/jffs2/
Dnodelist.h329 struct rb_node *node = rb_first(root); in frag_first()
361 #define tn_first(list) rb_entry(rb_first(list), struct jffs2_tmp_dnode_info, rb)
/linux-4.4.14/kernel/power/
Dwakelock.c46 for (node = rb_first(&wakelocks_tree); node; node = rb_next(node)) { in pm_show_wakelocks()
/linux-4.4.14/net/sched/
Dsch_fq.c415 while ((p = rb_first(&q->delayed)) != NULL) { in fq_check_throttled()
535 while ((p = rb_first(root)) != NULL) { in fq_reset()
565 while ((op = rb_first(oroot)) != NULL) { in fq_rehash()
Dsch_netem.c367 while ((p = rb_first(&q->t_root))) { in tfifo_reset()
587 struct rb_node *p = rb_first(&q->t_root); in netem_drop()
626 p = rb_first(&q->t_root); in netem_dequeue()
Dsch_hfsc.c232 for (n = rb_first(&q->eligible); n != NULL; n = rb_next(n)) { in eltree_get_mindl()
248 n = rb_first(&q->eligible); in eltree_get_minel()
296 for (n = rb_first(&cl->vt_tree); n != NULL; n = rb_next(n)) { in vttree_firstfit()
661 struct rb_node *n = rb_first(&cl->cf_tree); in update_cfmin()
Dsch_htb.c703 struct rb_node *p = rb_first(wait_pq); in htb_do_events()
/linux-4.4.14/drivers/mtd/
Dmtdswap.c92 #define MTDSWAP_ECNT_MIN(rbroot) (rb_entry(rb_first(rbroot), struct swap_eb, \
256 p = rb_first(root); in mtdswap_rb_index()
625 eb = rb_entry(rb_first(clean_root), struct swap_eb, rb); in mtdswap_map_free_block()
910 eb = rb_entry(rb_first(rp), struct swap_eb, rb); in mtdswap_pick_gc_eblk()
1238 min[i] = rb_entry(rb_first(root), struct swap_eb, in mtdswap_show()
/linux-4.4.14/tools/perf/
Dbuiltin-kmem.c959 next = rb_first(root); in __print_slab_result()
1012 struct rb_node *next = rb_first(&page_alloc_sorted); in __print_page_alloc_result()
1062 struct rb_node *next = rb_first(&page_caller_sorted); in __print_page_caller_result()
1245 node = rb_first(root); in __sort_slab_result()
1292 node = rb_first(root); in __sort_page_result()
Dbuiltin-annotate.c125 struct rb_node *nd = rb_first(&hists->entries), *next; in hists__find_annotations()
Dbuiltin-diff.c437 next = rb_first(root); in hists__baseline_only()
459 next = rb_first(root); in hists__precompute()
Dbuiltin-sched.c1315 node = rb_first(root); in perf_sched__sort_lat()
1632 while ((node = rb_first(&sched->atom_root))) { in perf_sched__merge_lat()
1655 next = rb_first(&sched->sorted_atom_root); in perf_sched__lat()
Dbuiltin-lock.c766 node = rb_first(&thread_stats); in dump_threads()
Dbuiltin-top.c358 next = rb_first(&hists->entries); in perf_top__prompt_symbol()
Dbuiltin-kvm.c531 struct rb_node *node = rb_first(result); in pop_from_result()
/linux-4.4.14/fs/
Deventpoll.c756 for (rbp = rb_first(&ep->rbr); rbp; rbp = rb_next(rbp)) { in ep_free()
772 while ((rbp = rb_first(&ep->rbr)) != NULL) { in ep_free()
879 for (rbp = rb_first(&ep->rbr); rbp; rbp = rb_next(rbp)) { in ep_show_fdinfo()
1691 for (rbp = rb_first(&ep->rbr); rbp; rbp = rb_next(rbp)) { in ep_loop_check_proc()
/linux-4.4.14/fs/ext4/
Dblock_validity.c128 node = rb_first(&sbi->system_blks); in debug_print_tree()
Ddir.c544 info->curr_node = rb_first(&info->root); in ext4_dx_readdir()
566 info->curr_node = rb_first(&info->root); in ext4_dx_readdir()
Dextents_status.c182 node = rb_first(&tree->root); in ext4_es_print_tree()
Dmballoc.c3455 n = rb_first(&(grp->bb_free_root)); in ext4_mb_generate_from_freelist()
/linux-4.4.14/arch/x86/mm/
Dpat_rbtree.c235 node = rb_first(&memtype_rbroot); in rbt_memtype_copy_nth_element()
/linux-4.4.14/security/keys/
Dgc.c216 cursor = rb_first(&key_serial_tree); in key_garbage_collector()
Dproc.c290 struct rb_node *n = rb_first(r); in key_user_first()
/linux-4.4.14/tools/perf/ui/
Dbrowser.c132 nd = rb_first(root); in ui_browser__rb_tree_seek()
161 browser->top = rb_first(browser->entries); in ui_browser__rb_tree_refresh()
/linux-4.4.14/net/netfilter/
Dnft_rbtree.c184 for (node = rb_first(&priv->root); node != NULL; node = rb_next(node)) { in nft_rbtree_walk()
Dxt_connlimit.c406 while ((node = rb_first(r)) != NULL) { in destroy_tree()
/linux-4.4.14/fs/btrfs/
Dfree-space-cache.c646 for (n = rb_first(&ctl->free_space_offset); n; n = rb_next(n)) { in merge_space_tree()
912 struct rb_node *node = rb_first(&ctl->free_space_offset); in write_cache_extent_entries()
925 node = rb_first(&cluster->root); in write_cache_extent_entries()
947 node = rb_first(&cluster->root); in write_cache_extent_entries()
2056 node = rb_first(&cluster->root); in insert_into_bitmap()
2460 for (n = rb_first(&ctl->free_space_offset); n; n = rb_next(n)) { in btrfs_dump_free_space()
2519 node = rb_first(&cluster->root); in __btrfs_return_cluster_to_free_space()
2728 node = rb_first(&cluster->root); in btrfs_alloc_from_cluster()
3408 entry = rb_entry(rb_first(&ctl->free_space_offset), in btrfs_find_ino_for_alloc()
Dinode-map.c261 n = rb_first(rbroot); in btrfs_unpin_free_ino()
Dqgroup.c495 while ((n = rb_first(&fs_info->qgroup_tree))) { in btrfs_free_qgroup_config()
1441 node = rb_first(&delayed_refs->dirty_extent_root); in btrfs_qgroup_prepare_account_extents()
1751 while ((node = rb_first(&delayed_refs->dirty_extent_root))) { in btrfs_qgroup_account_extents()
2420 for (n = rb_first(&fs_info->qgroup_tree); n; n = rb_next(n)) { in qgroup_rescan_zero_tracking()
Drelocation.c2438 while ((rb_node = rb_first(blocks))) {
2971 rb_node = rb_first(blocks);
2979 rb_node = rb_first(blocks);
2990 rb_node = rb_first(blocks);
3004 if (ret != -EAGAIN || rb_node == rb_first(blocks))
Ddelayed-ref.c147 n = rb_first(root); in find_ref_head()
Ddelayed-inode.c513 p = rb_first(&delayed_node->ins_root); in __btrfs_first_delayed_insertion_item()
526 p = rb_first(&delayed_node->del_root); in __btrfs_first_delayed_deletion_item()
Dfile.c270 node = rb_first(&fs_info->defrag_inodes); in btrfs_cleanup_defrag_inodes()
278 node = rb_first(&fs_info->defrag_inodes); in btrfs_cleanup_defrag_inodes()
Dsend.c6161 n = rb_first(&sctx->pending_dir_moves); in btrfs_ioctl_send()
6178 n = rb_first(&sctx->waiting_dir_moves); in btrfs_ioctl_send()
6189 n = rb_first(&sctx->orphan_dirs); in btrfs_ioctl_send()
Dtransaction.c95 node = rb_first(&tree->state); in clear_btree_io_tree()
Dinode.c2655 node = rb_first(&new->root); in relink_file_extents()
5067 node = rb_first(&map_tree->map); in evict_inode_truncate_pages()
5104 node = rb_first(&io_tree->state); in evict_inode_truncate_pages()
Ddisk-io.c4194 while ((node = rb_first(&delayed_refs->href_root)) != NULL) { in btrfs_destroy_delayed_refs()
Dextent-tree.c2694 n = rb_first(root); in find_middle()
2936 node = rb_first(&delayed_refs->href_root); in btrfs_run_delayed_refs()
Dctree.c425 for (node = rb_first(tm_root); node; node = next) { in btrfs_put_tree_mod_seq()
/linux-4.4.14/Documentation/
Drbtree.txt171 struct rb_node *rb_first(struct rb_root *tree);
176 To start iterating, call rb_first() or rb_last() with a pointer to the root
190 for (node = rb_first(&mytree); node; node = rb_next(node))
/linux-4.4.14/net/rds/
Dcong.c397 while ((node = rb_first(&rds_cong_tree))) { in rds_cong_exit()
Drdma.c137 while ((node = rb_first(&rs->rs_rdma_keys))) { in rds_rdma_drop_keys()
/linux-4.4.14/net/rxrpc/
Dar-connevent.c38 for (p = rb_first(&conn->calls); p; p = rb_next(p)) { in rxrpc_abort_calls()
Dar-call.c790 for (p = rb_first(&rx->calls); p; p = rb_next(p)) { in rxrpc_release_calls_on_socket()
/linux-4.4.14/fs/afs/
Dcallback.c384 vnode = rb_entry(rb_first(&server->cb_promises),
/linux-4.4.14/drivers/vfio/
Dvfio_iommu_type1.c663 n = rb_first(&iommu->dma_list); in vfio_iommu_replay()
858 while ((node = rb_first(&iommu->dma_list))) in vfio_iommu_unmap_unpin_all()
/linux-4.4.14/tools/lib/
Drbtree.c413 struct rb_node *rb_first(const struct rb_root *root) in rb_first() function
/linux-4.4.14/fs/gfs2/
Drgrp.c547 n = rb_first(&sdp->sd_rindex_tree); in gfs2_rgrpd_get_first()
569 n = rb_first(&sdp->sd_rindex_tree); in gfs2_rgrpd_get_next()
712 while ((n = rb_first(&rgd->rd_rstree))) { in return_all_reservations()
725 while ((n = rb_first(&sdp->sd_rindex_tree))) { in gfs2_clear_rgrpd()
2239 for (n = rb_first(&rgd->rd_rstree); n; n = rb_next(&trs->rs_node)) { in gfs2_rgrp_dump()
/linux-4.4.14/fs/nfs/
Dnfs4state.c177 for (pos = rb_first(&server->state_owners); in nfs4_get_renew_cred_server_locked()
1559 for (pos = rb_first(&server->state_owners); in nfs4_reset_seqids()
1609 for (pos = rb_first(&server->state_owners); in nfs4_clear_reclaim_server()
1720 for (pos = rb_first(&server->state_owners); in nfs4_do_reclaim()
Ddir.c2181 while ((n = rb_first(root_node)) != NULL) { in __nfs_access_zap_cache()
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx5/core/
Dpagealloc.c499 p = rb_first(&dev->priv.page_root); in mlx5_reclaim_startup_pages()
/linux-4.4.14/fs/nfs/blocklayout/
Dextent_tree.c20 struct rb_node *node = rb_first(root); in ext_tree_first()
/linux-4.4.14/fs/ubifs/
Dlog.c307 p = rb_first(&c->buds); in remove_buds()
Ddebug.c622 for (rb = rb_first(&c->buds); rb; rb = rb_next(rb)) { in ubifs_dump_budg()
710 for (rb = rb_first((struct rb_root *)&c->buds); rb; rb = rb_next(rb)) { in ubifs_dump_lprop()
2151 struct rb_node *this = rb_first(&fsckd->inodes); in check_inodes()
Drecovery.c1476 struct rb_node *this = rb_first(&c->size_tree); in ubifs_recover_size()
/linux-4.4.14/drivers/md/bcache/
Dutil.h544 container_of_or_null(rb_first(root), type, member)
/linux-4.4.14/arch/blackfin/kernel/
Dtrace.c128 for (n = rb_first(&mm->mm_rb); n; n = rb_next(n)) { in decode_address()
/linux-4.4.14/net/802/
Dgarp.c388 for (node = rb_first(&app->gid); in garp_gid_event()
Dmrp.c574 for (node = rb_first(&app->mad); in mrp_mad_event()
/linux-4.4.14/mm/
Dksm.c2021 node = rb_first(root_stable_tree + nid); in ksm_check_stable_tree()
2031 node = rb_first(root_stable_tree + nid); in ksm_check_stable_tree()
Dnommu.c576 lastp = rb_first(&nommu_region_tree); in validate_nommu_regions()
1307 for (rb = rb_first(&nommu_region_tree); rb; rb = rb_next(rb)) { in do_mmap()
Dbacking-dev.c694 while ((rbn = rb_first(&bdi->cgwb_congested_tree))) { in cgwb_bdi_destroy()
Dmmap.c384 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in browse_rb()
427 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in validate_mm_rb()
Dmempolicy.c2501 next = rb_first(&p->root); in mpol_free_shared_policy()
/linux-4.4.14/fs/f2fs/
Dextent_cache.c136 node = rb_first(&et->root); in __free_extent_tree()
/linux-4.4.14/block/
Dcfq-iosched.c1172 root->left = rb_first(&root->rb); in cfq_rb_first()
1183 root->left = rb_first(&root->rb); in cfq_rb_first_group()
1224 rbnext = rb_first(&cfqq->sort_list); in cfq_find_next_rq()
3558 BUG_ON(rb_first(&cfqq->sort_list)); in cfq_put_queue()
Dblk-throttle.c417 parent_sq->first_pending = rb_first(&parent_sq->pending_tree); in throtl_rb_first()
/linux-4.4.14/drivers/infiniband/core/
Dmulticast.c763 for (node = rb_first(&port->table); node; node = rb_next(node)) { in mcast_groups_event()
/linux-4.4.14/fs/kernfs/
Ddir.c973 rbn = rb_first(&pos->dir.children); in kernfs_leftmost_descendant()
/linux-4.4.14/fs/ext2/
Dballoc.c207 n = rb_first(root); in __rsv_window_dump()
/linux-4.4.14/drivers/infiniband/ulp/ipoib/
Dipoib_multicast.c997 n = rb_first(&priv->multicast_tree); in ipoib_mcast_iter_next()
Dipoib_main.c586 n = rb_first(&priv->path_tree); in ipoib_path_iter_next()
/linux-4.4.14/fs/cifs/
Dconnect.c3910 while ((node = rb_first(root))) { in cifs_umount()
4194 struct rb_node *node = rb_first(root); in cifs_prune_tlinks()
4206 node = rb_first(root); in cifs_prune_tlinks()
/linux-4.4.14/drivers/block/
Dpktcdvd.c1217 n = rb_first(&pd->bio_queue); in pkt_handle_queue()
1235 n = rb_first(&pd->bio_queue); in pkt_handle_queue()
/linux-4.4.14/drivers/block/xen-blkback/
Dblkback.c184 for ((pos) = container_of(rb_first((rbtree)), typeof(*(pos)), node), \
/linux-4.4.14/drivers/md/
Ddm-crypt.c1236 io = crypt_io_from_node(rb_first(&write_tree)); in dmcrypt_write()
Ddm-thin.c1959 for (node = rb_first(&tc->sort_bio_list); node; node = rb_next(node)) { in __extract_sorted_bios()
/linux-4.4.14/fs/fuse/
Ddev.c2097 p = rb_first(&fc->polled_files); in end_polls()
/linux-4.4.14/drivers/staging/lustre/lustre/osc/
Dosc_cache.c169 return rb_extent(rb_first(&obj->oo_root)); in first_extent()